| /* |
| * Copyright 2018-present Open Networking Foundation |
| |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package model |
| |
| import ( |
| "bytes" |
| "compress/gzip" |
| "github.com/golang/protobuf/proto" |
| "github.com/opencord/voltha-go/common/log" |
| "github.com/opencord/voltha-go/db/kvstore" |
| "reflect" |
| "strings" |
| "sync" |
| ) |
| |
| // PersistedRevision holds information of revision meant to be saved in a persistent storage |
| type PersistedRevision struct { |
| Revision |
| Compress bool |
| |
| events chan *kvstore.Event |
| kvStore *Backend |
| mutex sync.RWMutex |
| isStored bool |
| isWatched bool |
| } |
| |
| type watchCache struct { |
| Cache sync.Map |
| } |
| |
| var watchCacheInstance *watchCache |
| var watchCacheOne sync.Once |
| |
| func Watches() *watchCache { |
| watchCacheOne.Do(func() { |
| watchCacheInstance = &watchCache{Cache: sync.Map{}} |
| }) |
| return watchCacheInstance |
| } |
| |
| // NewPersistedRevision creates a new instance of a PersistentRevision structure |
| func NewPersistedRevision(branch *Branch, data interface{}, children map[string][]Revision) Revision { |
| pr := &PersistedRevision{} |
| pr.kvStore = branch.Node.GetRoot().KvStore |
| pr.Revision = NewNonPersistedRevision(nil, branch, data, children) |
| return pr |
| } |
| |
| // Finalize is responsible of saving the revision in the persistent storage |
| func (pr *PersistedRevision) Finalize(skipOnExist bool) { |
| pr.store(skipOnExist) |
| } |
| |
| func (pr *PersistedRevision) store(skipOnExist bool) { |
| if pr.GetBranch().Txid != "" { |
| return |
| } |
| |
| log.Debugw("ready-to-store-revision", log.Fields{"hash": pr.GetHash(), "name": pr.GetName(), "data": pr.GetData()}) |
| |
| if blob, err := proto.Marshal(pr.GetConfig().Data.(proto.Message)); err != nil { |
| // TODO report error |
| } else { |
| if pr.Compress { |
| var b bytes.Buffer |
| w := gzip.NewWriter(&b) |
| w.Write(blob) |
| w.Close() |
| blob = b.Bytes() |
| } |
| |
| if err := pr.kvStore.Put(pr.GetName(), blob); err != nil { |
| log.Warnw("problem-storing-revision", log.Fields{"error": err, "hash": pr.GetHash(), "name": pr.GetName(), "data": pr.GetConfig().Data}) |
| } else { |
| log.Debugw("storing-revision", log.Fields{"hash": pr.GetHash(), "name": pr.GetName(), "data": pr.GetConfig().Data}) |
| pr.isStored = true |
| } |
| } |
| } |
| |
| func (pr *PersistedRevision) SetupWatch(key string) { |
| if key == "" { |
| log.Debugw("ignoring-watch", log.Fields{"key": key, "revision-hash": pr.GetHash()}) |
| return |
| } |
| |
| if _, exists := Watches().Cache.LoadOrStore(key+"-"+pr.GetHash(), struct{}{}); exists { |
| return |
| } |
| |
| if pr.events == nil { |
| pr.events = make(chan *kvstore.Event) |
| |
| log.Debugw("setting-watch-channel", log.Fields{"key": key, "revision-hash": pr.GetHash()}) |
| |
| pr.SetName(key) |
| pr.events = pr.kvStore.CreateWatch(key) |
| } |
| |
| if !pr.isWatched { |
| pr.isWatched = true |
| |
| log.Debugw("setting-watch-routine", log.Fields{"key": key, "revision-hash": pr.GetHash()}) |
| |
| // Start watching |
| go pr.startWatching() |
| } |
| } |
| |
| func (pr *PersistedRevision) startWatching() { |
| log.Debugw("starting-watch", log.Fields{"key": pr.GetHash(), "watch": pr.GetName()}) |
| |
| StopWatchLoop: |
| for { |
| latestRev := pr.GetBranch().GetLatest() |
| |
| select { |
| case event, ok := <-pr.events: |
| if !ok { |
| log.Errorw("event-channel-failure: stopping watch loop", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName()}) |
| break StopWatchLoop |
| } |
| log.Debugw("received-event", log.Fields{"type": event.EventType, "watch": latestRev.GetName()}) |
| |
| switch event.EventType { |
| case kvstore.DELETE: |
| log.Debugw("delete-from-memory", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName()}) |
| pr.Revision.Drop("", true) |
| break StopWatchLoop |
| |
| case kvstore.PUT: |
| log.Debugw("update-in-memory", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName()}) |
| |
| data := reflect.New(reflect.TypeOf(latestRev.GetData()).Elem()) |
| |
| if err := proto.Unmarshal(event.Value.([]byte), data.Interface().(proto.Message)); err != nil { |
| log.Errorw("failed-to-unmarshal-watch-data", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName(), "error": err}) |
| } else { |
| log.Debugw("un-marshaled-watch-data", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName(), "data": data.Interface()}) |
| |
| var pathLock string |
| var pac *proxyAccessControl |
| var blobs map[string]*kvstore.KVPair |
| |
| // The watch reported new persistence data. |
| // Construct an object that will be used to update the memory |
| blobs = make(map[string]*kvstore.KVPair) |
| key, _ := kvstore.ToString(event.Key) |
| blobs[key] = &kvstore.KVPair{ |
| Key: key, |
| Value: event.Value, |
| Session: "", |
| Lease: 0, |
| } |
| |
| if latestRev.GetNode().GetProxy() != nil { |
| // |
| // If a proxy exists for this revision, use it to lock access to the path |
| // and prevent simultaneous updates to the object in memory |
| // |
| pathLock, _ = latestRev.GetNode().GetProxy().parseForControlledPath(latestRev.GetNode().GetProxy().getFullPath()) |
| |
| //If the proxy already has a request in progress, then there is no need to process the watch |
| log.Debugw("checking-if-path-is-locked", log.Fields{"key": latestRev.GetHash(), "pathLock": pathLock}) |
| if PAC().IsReserved(pathLock) { |
| log.Debugw("operation-in-progress", log.Fields{ |
| "key": latestRev.GetHash(), |
| "path": latestRev.GetNode().GetProxy().getFullPath(), |
| "operation": latestRev.GetNode().GetProxy().Operation.String(), |
| }) |
| |
| //continue |
| |
| // Identify the operation type and determine if the watch event should be applied or not. |
| switch latestRev.GetNode().GetProxy().Operation { |
| case PROXY_REMOVE: |
| fallthrough |
| |
| case PROXY_ADD: |
| fallthrough |
| |
| case PROXY_UPDATE: |
| // We will need to reload once the operation completes. |
| // Therefore, the data of the current event is most likely out-dated |
| // and should be ignored |
| log.Debugw("ignore-watch-event", log.Fields{ |
| "key": latestRev.GetHash(), |
| "path": latestRev.GetNode().GetProxy().getFullPath(), |
| "operation": latestRev.GetNode().GetProxy().Operation.String(), |
| }) |
| |
| continue |
| |
| case PROXY_CREATE: |
| fallthrough |
| |
| case PROXY_LIST: |
| fallthrough |
| |
| case PROXY_GET: |
| fallthrough |
| |
| case PROXY_WATCH: |
| fallthrough |
| |
| default: |
| log.Debugw("process-watch-event", log.Fields{ |
| "key": latestRev.GetHash(), |
| "path": latestRev.GetNode().GetProxy().getFullPath(), |
| "operation": latestRev.GetNode().GetProxy().Operation.String(), |
| }) |
| } |
| } |
| |
| // Reserve the path to prevent others to modify while we reload from persistence |
| log.Debugw("reserve-and-lock-path", log.Fields{"key": latestRev.GetHash(), "path": pathLock}) |
| pac = PAC().ReservePath(latestRev.GetNode().GetProxy().getFullPath(), |
| latestRev.GetNode().GetProxy(), pathLock) |
| pac.lock() |
| latestRev.GetNode().GetProxy().Operation = PROXY_WATCH |
| pac.SetProxy(latestRev.GetNode().GetProxy()) |
| |
| // Load changes and apply to memory |
| latestRev.LoadFromPersistence(latestRev.GetName(), "", blobs) |
| |
| log.Debugw("release-and-unlock-path", log.Fields{"key": latestRev.GetHash(), "path": pathLock}) |
| pac.getProxy().Operation = PROXY_GET |
| pac.unlock() |
| PAC().ReleasePath(pathLock) |
| |
| } else { |
| // This block should be reached only if coming from a non-proxied request |
| log.Debugw("revision-with-no-proxy", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName()}) |
| |
| // Load changes and apply to memory |
| latestRev.LoadFromPersistence(latestRev.GetName(), "", blobs) |
| } |
| } |
| |
| default: |
| log.Debugw("unhandled-event", log.Fields{"key": latestRev.GetHash(), "watch": latestRev.GetName(), "type": event.EventType}) |
| } |
| } |
| } |
| |
| Watches().Cache.Delete(pr.GetName() + "-" + pr.GetHash()) |
| |
| log.Debugw("exiting-watch", log.Fields{"key": pr.GetHash(), "watch": pr.GetName()}) |
| } |
| |
| // UpdateData modifies the information in the data model and saves it in the persistent storage |
| func (pr *PersistedRevision) UpdateData(data interface{}, branch *Branch) Revision { |
| log.Debugw("updating-persisted-data", log.Fields{"hash": pr.GetHash()}) |
| |
| newNPR := pr.Revision.UpdateData(data, branch) |
| |
| newPR := &PersistedRevision{ |
| Revision: newNPR, |
| Compress: pr.Compress, |
| kvStore: pr.kvStore, |
| events: pr.events, |
| isWatched: pr.isWatched, |
| } |
| |
| if newPR.GetHash() != pr.GetHash() { |
| newPR.isStored = false |
| pr.Drop(branch.Txid, false) |
| pr.Drop(branch.Txid, false) |
| } else { |
| newPR.isStored = true |
| } |
| |
| return newPR |
| } |
| |
| // UpdateChildren modifies the children of a revision and of a specific component and saves it in the persistent storage |
| func (pr *PersistedRevision) UpdateChildren(name string, children []Revision, |
| branch *Branch) Revision { |
| log.Debugw("updating-persisted-children", log.Fields{"hash": pr.GetHash()}) |
| |
| newNPR := pr.Revision.UpdateChildren(name, children, branch) |
| |
| newPR := &PersistedRevision{ |
| Revision: newNPR, |
| Compress: pr.Compress, |
| kvStore: pr.kvStore, |
| events: pr.events, |
| isWatched: pr.isWatched, |
| } |
| |
| if newPR.GetHash() != pr.GetHash() { |
| newPR.isStored = false |
| pr.Drop(branch.Txid, false) |
| } else { |
| newPR.isStored = true |
| } |
| |
| return newPR |
| } |
| |
| // UpdateAllChildren modifies the children for all components of a revision and saves it in the peristent storage |
| func (pr *PersistedRevision) UpdateAllChildren(children map[string][]Revision, branch *Branch) Revision { |
| log.Debugw("updating-all-persisted-children", log.Fields{"hash": pr.GetHash()}) |
| |
| newNPR := pr.Revision.UpdateAllChildren(children, branch) |
| |
| newPR := &PersistedRevision{ |
| Revision: newNPR, |
| Compress: pr.Compress, |
| kvStore: pr.kvStore, |
| events: pr.events, |
| isWatched: pr.isWatched, |
| } |
| |
| if newPR.GetHash() != pr.GetHash() { |
| newPR.isStored = false |
| pr.Drop(branch.Txid, false) |
| } else { |
| newPR.isStored = true |
| } |
| |
| return newPR |
| } |
| |
| // Drop takes care of eliminating a revision hash that is no longer needed |
| // and its associated config when required |
| func (pr *PersistedRevision) Drop(txid string, includeConfig bool) { |
| pr.Revision.Drop(txid, includeConfig) |
| } |
| |
| // Drop takes care of eliminating a revision hash that is no longer needed |
| // and its associated config when required |
| func (pr *PersistedRevision) StorageDrop(txid string, includeConfig bool) { |
| log.Debugw("dropping-revision", |
| log.Fields{"txid": txid, "hash": pr.GetHash(), "config-hash": pr.GetConfig().Hash}) |
| |
| pr.mutex.Lock() |
| defer pr.mutex.Unlock() |
| if pr.kvStore != nil && txid == "" { |
| if pr.isStored { |
| if pr.isWatched { |
| pr.kvStore.DeleteWatch(pr.GetName(), pr.events) |
| pr.isWatched = false |
| } |
| |
| if err := pr.kvStore.Delete(pr.GetName()); err != nil { |
| log.Errorw("failed-to-remove-revision", log.Fields{"hash": pr.GetHash(), "error": err.Error()}) |
| } else { |
| pr.isStored = false |
| } |
| } |
| |
| } else { |
| if includeConfig { |
| log.Debugw("attempted-to-remove-transacted-revision-config", log.Fields{"hash": pr.GetConfig().Hash, "txid": txid}) |
| } |
| log.Debugw("attempted-to-remove-transacted-revision", log.Fields{"hash": pr.GetHash(), "txid": txid}) |
| } |
| |
| pr.Revision.Drop(txid, includeConfig) |
| } |
| |
| // verifyPersistedEntry validates if the provided data is available or not in memory and applies updates as required |
| func (pr *PersistedRevision) verifyPersistedEntry(data interface{}, typeName string, keyName string, keyValue string, txid string) (response Revision) { |
| // Parent which holds the current node entry |
| parent := pr.GetBranch().Node.Root |
| |
| // Get a copy of the parent's children |
| children := make([]Revision, len(parent.GetBranch(NONE).Latest.GetChildren(typeName))) |
| copy(children, parent.GetBranch(NONE).Latest.GetChildren(typeName)) |
| |
| // Verify if a child with the provided key value can be found |
| if childIdx, childRev := pr.GetNode().findRevByKey(children, keyName, keyValue); childRev != nil { |
| // A child matching the provided key exists in memory |
| // Verify if the data differs from what was retrieved from persistence |
| // Also check if we are treating a newer revision of the data or not |
| if childRev.GetData().(proto.Message).String() != data.(proto.Message).String() { |
| log.Debugw("revision-data-is-different", log.Fields{ |
| "key": childRev.GetHash(), |
| "name": childRev.GetName(), |
| "data": childRev.GetData(), |
| }) |
| |
| // |
| // Data has changed; replace the child entry and update the parent revision |
| // |
| |
| // BEGIN Lock child -- prevent any incoming changes |
| childRev.GetBranch().LatestLock.Lock() |
| |
| // Update child |
| updatedChildRev := childRev.UpdateData(data, childRev.GetBranch()) |
| |
| updatedChildRev.GetNode().SetProxy(childRev.GetNode().GetProxy()) |
| updatedChildRev.SetupWatch(updatedChildRev.GetName()) |
| updatedChildRev.SetLastUpdate() |
| |
| // Update cache |
| GetRevCache().Cache.Store(updatedChildRev.GetName(), updatedChildRev) |
| childRev.Drop(txid, false) |
| |
| childRev.GetBranch().LatestLock.Unlock() |
| // END lock child |
| |
| // Update child entry |
| children[childIdx] = updatedChildRev |
| |
| // BEGIN lock parent -- Update parent |
| parent.GetBranch(NONE).LatestLock.Lock() |
| |
| updatedRev := parent.GetBranch(NONE).Latest.UpdateChildren(typeName, children, parent.GetBranch(NONE)) |
| parent.GetBranch(NONE).Node.makeLatest(parent.GetBranch(NONE), updatedRev, nil) |
| |
| parent.GetBranch(NONE).LatestLock.Unlock() |
| // END lock parent |
| |
| // Drop the previous child revision |
| parent.GetBranch(NONE).Latest.ChildDrop(typeName, childRev.GetHash()) |
| |
| if updatedChildRev != nil { |
| log.Debugw("verify-persisted-entry--adding-child", log.Fields{ |
| "key": updatedChildRev.GetHash(), |
| "name": updatedChildRev.GetName(), |
| "data": updatedChildRev.GetData(), |
| }) |
| response = updatedChildRev |
| } |
| } else { |
| // Data is the same. Continue to the next entry |
| log.Debugw("same-revision-data", log.Fields{ |
| "key": childRev.GetHash(), |
| "name": childRev.GetName(), |
| "data": childRev.GetData(), |
| }) |
| if childRev != nil { |
| log.Debugw("keeping-same-revision-data", log.Fields{ |
| "key": childRev.GetHash(), |
| "name": childRev.GetName(), |
| "data": childRev.GetData(), |
| }) |
| |
| // Update timestamp to reflect when it was last read and to reset tracked timeout |
| childRev.SetLastUpdate() |
| GetRevCache().Cache.Store(childRev.GetName(), childRev) |
| response = childRev |
| } |
| } |
| |
| } else { |
| // There is no available child with that key value. |
| // Create a new child and update the parent revision. |
| log.Debugw("no-such-revision-entry", log.Fields{ |
| "key": keyValue, |
| "name": typeName, |
| "data": data, |
| }) |
| |
| // BEGIN child lock |
| pr.GetBranch().LatestLock.Lock() |
| |
| // Construct a new child node with the retrieved persistence data |
| childRev = pr.GetBranch().Node.MakeNode(data, txid).Latest(txid) |
| |
| // We need to start watching this entry for future changes |
| childRev.SetName(typeName + "/" + keyValue) |
| childRev.SetupWatch(childRev.GetName()) |
| |
| pr.GetBranch().LatestLock.Unlock() |
| // END child lock |
| |
| // |
| // Add the child to the parent revision |
| // |
| |
| // BEGIN parent lock |
| parent.GetBranch(NONE).LatestLock.Lock() |
| children = append(children, childRev) |
| updatedRev := parent.GetBranch(NONE).Latest.UpdateChildren(typeName, children, parent.GetBranch(NONE)) |
| updatedRev.GetNode().SetProxy(parent.GetBranch(NONE).Node.GetProxy()) |
| parent.GetBranch(NONE).Node.makeLatest(parent.GetBranch(NONE), updatedRev, nil) |
| parent.GetBranch(NONE).LatestLock.Unlock() |
| // END parent lock |
| |
| // Child entry is valid and can be included in the response object |
| if childRev != nil { |
| log.Debugw("adding-revision-to-response", log.Fields{ |
| "key": childRev.GetHash(), |
| "name": childRev.GetName(), |
| "data": childRev.GetData(), |
| }) |
| response = childRev |
| } |
| } |
| |
| return response |
| } |
| |
| // LoadFromPersistence retrieves data from kv store at the specified location and refreshes the memory |
| // by adding missing entries, updating changed entries and ignoring unchanged ones |
| func (pr *PersistedRevision) LoadFromPersistence(path string, txid string, blobs map[string]*kvstore.KVPair) []Revision { |
| pr.mutex.Lock() |
| defer pr.mutex.Unlock() |
| |
| log.Debugw("loading-from-persistence", log.Fields{"path": path, "txid": txid}) |
| |
| var response []Revision |
| |
| for strings.HasPrefix(path, "/") { |
| path = path[1:] |
| } |
| |
| if pr.kvStore != nil && path != "" { |
| if blobs == nil || len(blobs) == 0 { |
| log.Debugw("retrieve-from-kv", log.Fields{"path": path, "txid": txid}) |
| blobs, _ = pr.kvStore.List(path) |
| } |
| |
| partition := strings.SplitN(path, "/", 2) |
| name := partition[0] |
| |
| var nodeType interface{} |
| if len(partition) < 2 { |
| path = "" |
| nodeType = pr.GetBranch().Node.Type |
| } else { |
| path = partition[1] |
| nodeType = pr.GetBranch().Node.Root.Type |
| } |
| |
| field := ChildrenFields(nodeType)[name] |
| |
| if field != nil && field.IsContainer { |
| log.Debugw("parsing-data-blobs", log.Fields{ |
| "path": path, |
| "name": name, |
| "size": len(blobs), |
| }) |
| |
| for _, blob := range blobs { |
| output := blob.Value.([]byte) |
| |
| data := reflect.New(field.ClassType.Elem()) |
| |
| if err := proto.Unmarshal(output, data.Interface().(proto.Message)); err != nil { |
| log.Errorw("failed-to-unmarshal", log.Fields{ |
| "path": path, |
| "txid": txid, |
| "error": err, |
| }) |
| } else if path == "" { |
| if field.Key != "" { |
| log.Debugw("no-path-with-container-key", log.Fields{ |
| "path": path, |
| "txid": txid, |
| "data": data.Interface(), |
| }) |
| |
| // Retrieve the key identifier value from the data structure |
| // based on the field's key attribute |
| _, key := GetAttributeValue(data.Interface(), field.Key, 0) |
| |
| if entry := pr.verifyPersistedEntry(data.Interface(), name, field.Key, key.String(), txid); entry != nil { |
| response = append(response, entry) |
| } |
| } else { |
| log.Debugw("path-with-no-container-key", log.Fields{ |
| "path": path, |
| "txid": txid, |
| "data": data.Interface(), |
| }) |
| } |
| |
| } else if field.Key != "" { |
| log.Debugw("path-with-container-key", log.Fields{ |
| "path": path, |
| "txid": txid, |
| "data": data.Interface(), |
| }) |
| // The request is for a specific entry/id |
| partition := strings.SplitN(path, "/", 2) |
| key := partition[0] |
| if len(partition) < 2 { |
| path = "" |
| } else { |
| path = partition[1] |
| } |
| keyValue := field.KeyFromStr(key) |
| |
| if entry := pr.verifyPersistedEntry(data.Interface(), name, field.Key, keyValue.(string), txid); entry != nil { |
| response = append(response, entry) |
| } |
| } |
| } |
| |
| log.Debugw("no-more-data-blobs", log.Fields{"path": path, "name": name}) |
| } else { |
| log.Debugw("cannot-process-field", log.Fields{ |
| "type": pr.GetBranch().Node.Type, |
| "name": name, |
| }) |
| } |
| } |
| |
| return response |
| } |