| /* |
| * Copyright 2020-present Open Networking Foundation |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| //Package main -> this is the entry point of the OpenOnuAdapter |
| package main |
| |
| import ( |
| "context" |
| "errors" |
| "fmt" |
| "io/ioutil" |
| "os" |
| "os/signal" |
| "strings" |
| "syscall" |
| "time" |
| |
| "github.com/opencord/voltha-lib-go/v4/pkg/adapters" |
| "github.com/opencord/voltha-lib-go/v4/pkg/adapters/adapterif" |
| com "github.com/opencord/voltha-lib-go/v4/pkg/adapters/common" |
| conf "github.com/opencord/voltha-lib-go/v4/pkg/config" |
| "github.com/opencord/voltha-lib-go/v4/pkg/db/kvstore" |
| "github.com/opencord/voltha-lib-go/v4/pkg/events" |
| "github.com/opencord/voltha-lib-go/v4/pkg/events/eventif" |
| "github.com/opencord/voltha-lib-go/v4/pkg/kafka" |
| "github.com/opencord/voltha-lib-go/v4/pkg/log" |
| "github.com/opencord/voltha-lib-go/v4/pkg/probe" |
| "github.com/opencord/voltha-lib-go/v4/pkg/version" |
| ic "github.com/opencord/voltha-protos/v4/go/inter_container" |
| "github.com/opencord/voltha-protos/v4/go/voltha" |
| |
| "github.com/opencord/voltha-openonu-adapter-go/internal/pkg/config" |
| ac "github.com/opencord/voltha-openonu-adapter-go/internal/pkg/onuadaptercore" |
| ) |
| |
| type adapter struct { |
| //defaultAppName string |
| instanceID string |
| config *config.AdapterFlags |
| iAdapter adapters.IAdapter // from Voltha interface adapters |
| kafkaClient kafka.Client |
| kvClient kvstore.Client |
| kip kafka.InterContainerProxy |
| coreProxy adapterif.CoreProxy |
| adapterProxy adapterif.AdapterProxy |
| eventProxy eventif.EventProxy |
| halted bool |
| exitChannel chan int |
| receiverChannels []<-chan *ic.InterContainerMessage //from inter-container |
| } |
| |
| func newAdapter(cf *config.AdapterFlags) *adapter { |
| var a adapter |
| a.instanceID = cf.InstanceID |
| a.config = cf |
| a.halted = false |
| a.exitChannel = make(chan int, 1) |
| a.receiverChannels = make([]<-chan *ic.InterContainerMessage, 0) |
| return &a |
| } |
| |
| func (a *adapter) start(ctx context.Context) error { |
| logger.Info(ctx, "Starting Core Adapter components") |
| var err error |
| |
| var p *probe.Probe |
| if value := ctx.Value(probe.ProbeContextKey); value != nil { |
| if _, ok := value.(*probe.Probe); ok { |
| p = value.(*probe.Probe) |
| p.RegisterService( |
| ctx, |
| "message-bus", |
| "kv-store", |
| "container-proxy", |
| "core-request-handler", |
| "register-with-core", |
| ) |
| } |
| } |
| |
| // Setup KV Client |
| logger.Debugw(ctx, "create-kv-client", log.Fields{"kvstore": a.config.KVStoreType}) |
| if err = a.setKVClient(ctx); err != nil { |
| logger.Fatalw(ctx, "error-setting-kv-client", log.Fields{"error": err}) |
| } |
| |
| if p != nil { |
| p.UpdateStatus(ctx, "kv-store", probe.ServiceStatusRunning) |
| } |
| |
| // Setup Log Config |
| cm := conf.NewConfigManager(ctx, a.kvClient, a.config.KVStoreType, a.config.KVStoreAddress, a.config.KVStoreTimeout) |
| go conf.StartLogLevelConfigProcessing(cm, ctx) |
| |
| // Setup Kafka Client |
| if a.kafkaClient, err = newKafkaClient(ctx, "sarama", a.config.KafkaAdapterAddress); err != nil { |
| logger.Fatalw(ctx, "Unsupported-common-client", log.Fields{"error": err}) |
| } |
| |
| if p != nil { |
| p.UpdateStatus(ctx, "message-bus", probe.ServiceStatusRunning) |
| } |
| |
| // Start the common InterContainer Proxy - retries as per program arguments or indefinitely per default |
| if a.kip, err = a.startInterContainerProxy(ctx, a.config.KafkaReconnectRetries); err != nil { |
| logger.Fatalw(ctx, "error-starting-inter-container-proxy", log.Fields{"error": err}) |
| //aborting the complete processing here (does notmake sense after set Retry number [else use -1 for infinite]) |
| return err |
| } |
| |
| // Create the core proxy to handle requests to the Core |
| a.coreProxy = com.NewCoreProxy(ctx, a.kip, a.config.Topic, a.config.CoreTopic) |
| |
| logger.Debugw(ctx, "create adapter proxy", log.Fields{"CoreTopic": a.config.CoreTopic}) |
| a.adapterProxy = com.NewAdapterProxy(ctx, a.kip, a.config.CoreTopic, cm.Backend) |
| |
| // Create the event proxy to post events to KAFKA |
| a.eventProxy = events.NewEventProxy(events.MsgClient(a.kafkaClient), events.MsgTopic(kafka.Topic{Name: a.config.EventTopic})) |
| |
| // Create the open ONU interface adapter |
| if a.iAdapter, err = a.startVolthaInterfaceAdapter(ctx, a.kip, a.coreProxy, a.adapterProxy, a.eventProxy, |
| a.config, cm); err != nil { |
| logger.Fatalw(ctx, "error-starting-volthaInterfaceAdapter for OpenOnt", log.Fields{"error": err}) |
| } |
| |
| // Register the core request handler |
| if err = a.setupRequestHandler(ctx, a.instanceID, a.iAdapter); err != nil { |
| logger.Fatalw(ctx, "error-setting-core-request-handler", log.Fields{"error": err}) |
| } |
| |
| // Register this adapter to the Core - retries indefinitely |
| if err = a.registerWithCore(ctx, -1); err != nil { |
| logger.Fatalw(ctx, "error-registering-with-core", log.Fields{"error": err}) |
| } |
| |
| // check the readiness and liveliness and update the probe status |
| a.checkServicesReadiness(ctx) |
| return err |
| } |
| |
| func (a *adapter) stop(ctx context.Context) { |
| // Stop leadership tracking |
| a.halted = true |
| |
| // send exit signal |
| a.exitChannel <- 0 |
| |
| // Cleanup - applies only if we had a kvClient |
| if a.kvClient != nil { |
| // Release all reservations |
| if err := a.kvClient.ReleaseAllReservations(ctx); err != nil { |
| logger.Infow(ctx, "fail-to-release-all-reservations", log.Fields{"error": err}) |
| } |
| // Close the DB connection |
| a.kvClient.Close(ctx) |
| } |
| |
| if a.kip != nil { |
| a.kip.Stop(ctx) |
| } |
| } |
| |
| // ############################################# |
| // Adapter Utility methods ##### begin ######### |
| |
| func newKVClient(ctx context.Context, storeType, address string, timeout time.Duration) (kvstore.Client, error) { |
| logger.Infow(ctx, "kv-store-type", log.Fields{"store": storeType}) |
| switch storeType { |
| case "consul": |
| return kvstore.NewConsulClient(ctx, address, timeout) |
| case "etcd": |
| return kvstore.NewEtcdClient(ctx, address, timeout, log.FatalLevel) |
| } |
| return nil, errors.New("unsupported-kv-store") |
| } |
| |
| func newKafkaClient(ctx context.Context, clientType, addr string) (kafka.Client, error) { |
| |
| logger.Infow(ctx, "common-client-type", log.Fields{"client": clientType}) |
| |
| switch clientType { |
| case "sarama": |
| return kafka.NewSaramaClient( |
| kafka.Address(addr), |
| kafka.ProducerReturnOnErrors(true), |
| kafka.ProducerReturnOnSuccess(true), |
| kafka.ProducerMaxRetries(6), |
| kafka.ProducerRetryBackoff(time.Millisecond*30), |
| kafka.MetadatMaxRetries(15)), nil |
| } |
| |
| return nil, errors.New("unsupported-client-type") |
| } |
| |
| func (a *adapter) setKVClient(ctx context.Context) error { |
| client, err := newKVClient(ctx, a.config.KVStoreType, a.config.KVStoreAddress, a.config.KVStoreTimeout) |
| if err != nil { |
| a.kvClient = nil |
| logger.Errorw(ctx, "error-starting-KVClient", log.Fields{"error": err}) |
| return err |
| } |
| a.kvClient = client |
| return nil |
| } |
| |
| func (a *adapter) startInterContainerProxy(ctx context.Context, retries int) (kafka.InterContainerProxy, error) { |
| logger.Infow(ctx, "starting-intercontainer-messaging-proxy", log.Fields{"addr": a.config.KafkaAdapterAddress, "topic": a.config.Topic}) |
| var err error |
| /* address config update acc. to [VOL-2736] */ |
| kip := kafka.NewInterContainerProxy( |
| kafka.InterContainerAddress(a.config.KafkaAdapterAddress), |
| kafka.MsgClient(a.kafkaClient), |
| kafka.DefaultTopic(&kafka.Topic{Name: a.config.Topic})) |
| count := 0 |
| for { |
| if err = kip.Start(ctx); err != nil { |
| logger.Warnw(ctx, "error-starting-messaging-proxy", log.Fields{"error": err, "retry": retries, "count": count}) |
| if retries == count { |
| return nil, err |
| } |
| count++ |
| // Take a nap before retrying |
| time.Sleep(2 * time.Second) |
| } else { |
| break |
| } |
| } |
| probe.UpdateStatusFromContext(ctx, "container-proxy", probe.ServiceStatusRunning) |
| logger.Info(ctx, "common-messaging-proxy-created") |
| return kip, nil |
| } |
| |
| func (a *adapter) startVolthaInterfaceAdapter(ctx context.Context, kip kafka.InterContainerProxy, |
| cp adapterif.CoreProxy, ap adapterif.AdapterProxy, ep eventif.EventProxy, |
| cfg *config.AdapterFlags, cm *conf.ConfigManager) (*ac.OpenONUAC, error) { |
| var err error |
| sAcONU := ac.NewOpenONUAC(ctx, a.kip, cp, ap, ep, a.kvClient, cfg, cm) |
| |
| if err = sAcONU.Start(ctx); err != nil { |
| logger.Fatalw(ctx, "error-starting-OpenOnuAdapterCore", log.Fields{"error": err}) |
| return nil, err |
| } |
| |
| logger.Info(ctx, "open-ont-OpenOnuAdapterCore-started") |
| return sAcONU, nil |
| } |
| |
| func (a *adapter) setupRequestHandler(ctx context.Context, coreInstanceID string, iadapter adapters.IAdapter) error { |
| logger.Info(ctx, "setting-request-handler") |
| requestProxy := com.NewRequestHandlerProxy(coreInstanceID, iadapter, a.coreProxy) |
| if err := a.kip.SubscribeWithRequestHandlerInterface(ctx, kafka.Topic{Name: a.config.Topic}, requestProxy); err != nil { |
| logger.Errorw(ctx, "request-handler-setup-failed", log.Fields{"error": err}) |
| return err |
| |
| } |
| probe.UpdateStatusFromContext(ctx, "core-request-handler", probe.ServiceStatusRunning) |
| logger.Info(ctx, "request-handler-setup-done") |
| return nil |
| } |
| |
| func (a *adapter) registerWithCore(ctx context.Context, retries int) error { |
| adapterID := fmt.Sprintf("brcm_openomci_onu_%d", a.config.CurrentReplica) |
| vendorIdsList := strings.Split(a.config.OnuVendorIds, ",") |
| logger.Infow(ctx, "registering-with-core", log.Fields{ |
| "adapterID": adapterID, |
| "currentReplica": a.config.CurrentReplica, |
| "totalReplicas": a.config.TotalReplicas, |
| "onuVendorIds": vendorIdsList, |
| }) |
| adapterDescription := &voltha.Adapter{ |
| Id: adapterID, // Unique name for the device type ->exact type required for OLT comm???? |
| Vendor: "VOLTHA OpenONUGo", |
| Version: version.VersionInfo.Version, |
| Endpoint: a.config.Topic, |
| Type: "brcm_openomci_onu", |
| CurrentReplica: int32(a.config.CurrentReplica), |
| TotalReplicas: int32(a.config.TotalReplicas), |
| } |
| types := []*voltha.DeviceType{{Id: "brcm_openomci_onu", |
| VendorIds: vendorIdsList, |
| Adapter: "brcm_openomci_onu", // Name of the adapter that handles device type |
| AcceptsBulkFlowUpdate: false, // Currently openolt adapter does not support bulk flow handling |
| AcceptsAddRemoveFlowUpdates: true}} |
| deviceTypes := &voltha.DeviceTypes{Items: types} |
| count := 0 |
| for { |
| if err := a.coreProxy.RegisterAdapter(context.TODO(), adapterDescription, deviceTypes); err != nil { |
| logger.Warnw(ctx, "registering-with-core-failed", log.Fields{"error": err}) |
| if retries == count { |
| return err |
| } |
| count++ |
| // Take a nap before retrying |
| time.Sleep(2 * time.Second) |
| } else { |
| break |
| } |
| } |
| probe.UpdateStatusFromContext(ctx, "register-with-core", probe.ServiceStatusRunning) |
| logger.Info(ctx, "registered-with-core") |
| return nil |
| } |
| |
| /** |
| This function checks the liveliness and readiness of the kakfa and kv-client services |
| and update the status in the probe. |
| */ |
| func (a *adapter) checkServicesReadiness(ctx context.Context) { |
| // checks the kafka readiness |
| go a.checkKafkaReadiness(ctx) |
| |
| // checks the kv-store readiness |
| go a.checkKvStoreReadiness(ctx) |
| } |
| |
| /** |
| This function checks the liveliness and readiness of the kv-store service |
| and update the status in the probe. |
| */ |
| func (a *adapter) checkKvStoreReadiness(ctx context.Context) { |
| // dividing the live probe interval by 2 to get updated status every 30s |
| timeout := a.config.LiveProbeInterval / 2 |
| kvStoreChannel := make(chan bool, 1) |
| |
| // Default false to check the liveliness. |
| kvStoreChannel <- false |
| for { |
| timeoutTimer := time.NewTimer(timeout) |
| select { |
| case liveliness := <-kvStoreChannel: |
| if !liveliness { |
| // kv-store not reachable or down, updating the status to not ready state |
| probe.UpdateStatusFromContext(ctx, "kv-store", probe.ServiceStatusNotReady) |
| timeout = a.config.NotLiveProbeInterval |
| } else { |
| // kv-store is reachable , updating the status to running state |
| probe.UpdateStatusFromContext(ctx, "kv-store", probe.ServiceStatusRunning) |
| timeout = a.config.LiveProbeInterval / 2 |
| } |
| // Check if the timer has expired or not |
| if !timeoutTimer.Stop() { |
| <-timeoutTimer.C |
| } |
| case <-timeoutTimer.C: |
| // Check the status of the kv-store |
| logger.Info(ctx, "kv-store liveliness-recheck") |
| if a.kvClient.IsConnectionUp(ctx) { |
| kvStoreChannel <- true |
| } else { |
| kvStoreChannel <- false |
| } |
| } |
| } |
| } |
| |
| /** |
| This function checks the liveliness and readiness of the kafka service |
| and update the status in the probe. |
| */ |
| func (a *adapter) checkKafkaReadiness(ctx context.Context) { |
| livelinessChannel := a.kafkaClient.EnableLivenessChannel(ctx, true) |
| healthinessChannel := a.kafkaClient.EnableHealthinessChannel(ctx, true) |
| timeout := a.config.LiveProbeInterval |
| for { |
| timeoutTimer := time.NewTimer(timeout) |
| |
| select { |
| case healthiness := <-healthinessChannel: |
| if !healthiness { |
| // logger.Fatal will call os.Exit(1) to terminate |
| logger.Fatal(ctx, "Kafka service has become unhealthy") |
| } |
| case liveliness := <-livelinessChannel: |
| if !liveliness { |
| // kafka not reachable or down, updating the status to not ready state |
| probe.UpdateStatusFromContext(ctx, "message-bus", probe.ServiceStatusNotReady) |
| timeout = a.config.NotLiveProbeInterval |
| } else { |
| // kafka is reachable , updating the status to running state |
| probe.UpdateStatusFromContext(ctx, "message-bus", probe.ServiceStatusRunning) |
| timeout = a.config.LiveProbeInterval |
| } |
| // Check if the timer has expired or not |
| if !timeoutTimer.Stop() { |
| <-timeoutTimer.C |
| } |
| case <-timeoutTimer.C: |
| logger.Info(ctx, "kafka-proxy-liveness-recheck") |
| // send the liveness probe in a goroutine; we don't want to deadlock ourselves as |
| // the liveness probe may wait (and block) writing to our channel. |
| err := a.kafkaClient.SendLiveness(ctx) |
| if err != nil { |
| // Catch possible error case if sending liveness after Sarama has been stopped. |
| logger.Warnw(ctx, "error-kafka-send-liveness", log.Fields{"error": err}) |
| } |
| } |
| } |
| } |
| |
| // Adapter Utility methods ##### end ######### |
| // ############################################# |
| |
| func getVerifiedCodeVersion(ctx context.Context) string { |
| if version.VersionInfo.Version == "unknown-version" { |
| content, err := ioutil.ReadFile("VERSION") |
| if err == nil { |
| return string(content) |
| } |
| logger.Error(ctx, "'VERSION'-file not readable") |
| } |
| return version.VersionInfo.Version |
| } |
| |
| func printVersion(appName string) { |
| fmt.Println(appName) |
| fmt.Println(version.VersionInfo.String(" ")) |
| } |
| |
| func printBanner() { |
| fmt.Println(" ____ ____ ___ ___ _ ") |
| fmt.Println(" / __ \\ / __ \\| \\ \\ | | | | |") |
| fmt.Println(" | | | |_ __ ___ _ __ | | | | |\\ \\ | | | | | ____ ____") |
| fmt.Println(" | | | | '_ \\ / _ \\ '_ \\ | | | | | \\ \\ | | | | | / '_ \\ / _' \\") |
| fmt.Println(" | |__| | |_) | __/| | | || |__| | | \\ \\| | \\__/ || (__) | (__) |") |
| fmt.Println(" \\___ /| .__/ \\___|_| |_| \\____/|_| \\___|______| \\.___ |\\___./") |
| fmt.Println(" | | __| |") |
| fmt.Println(" |_| |____/") |
| fmt.Println(" ") |
| } |
| |
| func waitForExit(ctx context.Context) int { |
| signalChannel := make(chan os.Signal, 1) |
| signal.Notify(signalChannel, |
| syscall.SIGHUP, |
| syscall.SIGINT, |
| syscall.SIGTERM, |
| syscall.SIGQUIT) |
| |
| exitChannel := make(chan int) |
| |
| go func() { |
| select { |
| case <-ctx.Done(): |
| logger.Infow(ctx, "Adapter run aborted due to internal errors", log.Fields{"context": "done"}) |
| exitChannel <- 2 |
| case s := <-signalChannel: |
| switch s { |
| case syscall.SIGHUP, |
| syscall.SIGINT, |
| syscall.SIGTERM, |
| syscall.SIGQUIT: |
| logger.Infow(ctx, "closing-signal-received", log.Fields{"signal": s}) |
| exitChannel <- 0 |
| default: |
| logger.Infow(ctx, "unexpected-signal-received", log.Fields{"signal": s}) |
| exitChannel <- 1 |
| } |
| } |
| }() |
| |
| code := <-exitChannel |
| return code |
| } |
| |
| func main() { |
| start := time.Now() |
| ctx, cancel := context.WithCancel(context.Background()) |
| defer cancel() |
| |
| cf := config.NewAdapterFlags() |
| defaultAppName := cf.InstanceID + "_" + getVerifiedCodeVersion(ctx) |
| cf.ParseCommandArguments() |
| |
| // Setup logging |
| |
| logLevel, err := log.StringToLogLevel(cf.LogLevel) |
| if err != nil { |
| logger.Fatalf(ctx, "Cannot setup logging, %s", err) |
| } |
| |
| // Setup default logger - applies for packages that do not have specific logger set |
| if _, err := log.SetDefaultLogger(log.JSON, logLevel, log.Fields{"instanceId": cf.InstanceID}); err != nil { |
| logger.With(log.Fields{"error": err}).Fatal(ctx, "Cannot setup logging") |
| } |
| |
| // Update all loggers (provisioned via init) with a common field |
| if err := log.UpdateAllLoggers(log.Fields{"instanceId": cf.InstanceID}); err != nil { |
| logger.With(log.Fields{"error": err}).Fatal(ctx, "Cannot setup logging") |
| } |
| |
| log.SetAllLogLevel(logLevel) |
| |
| realMain(ctx) //fatal on httpListen(0,6060) ... |
| |
| defer func() { |
| _ = log.CleanUp() |
| }() |
| // Print version / build information and exit |
| if cf.DisplayVersionOnly { |
| printVersion(defaultAppName) |
| return |
| } |
| logger.Infow(ctx, "config", log.Fields{"StartName": defaultAppName}) |
| logger.Infow(ctx, "config", log.Fields{"BuildVersion": version.VersionInfo.String(" ")}) |
| logger.Infow(ctx, "config", log.Fields{"Arguments": os.Args[1:]}) |
| |
| // Print banner if specified |
| if cf.Banner { |
| printBanner() |
| } |
| |
| logger.Infow(ctx, "config", log.Fields{"config": *cf}) |
| |
| ad := newAdapter(cf) |
| |
| p := &probe.Probe{} |
| logger.Infow(ctx, "resources", log.Fields{"Context": ctx, "Adapter": ad.instanceID, "ProbeCoreState": p.GetStatus("register-with-core")}) |
| |
| go p.ListenAndServe(ctx, fmt.Sprintf("%s:%d", ad.config.ProbeHost, ad.config.ProbePort)) |
| logger.Infow(ctx, "probeState", log.Fields{"ProbeCoreState": p.GetStatus("register-with-core")}) |
| |
| probeCtx := context.WithValue(ctx, probe.ProbeContextKey, p) |
| |
| closer, err := log.GetGlobalLFM().InitTracingAndLogCorrelation(cf.TraceEnabled, cf.TraceAgentAddress, cf.LogCorrelationEnabled) |
| if err != nil { |
| logger.Warnw(ctx, "unable-to-initialize-tracing-and-log-correlation-module", log.Fields{"error": err}) |
| } else { |
| defer log.TerminateTracing(closer) |
| } |
| |
| go func() { |
| err := ad.start(probeCtx) |
| // If this operation returns an error |
| // cancel all operations using this context |
| if err != nil { |
| cancel() |
| } |
| }() |
| |
| code := waitForExit(ctx) |
| logger.Infow(ctx, "received-a-closing-signal", log.Fields{"code": code}) |
| |
| // Cleanup before leaving |
| ad.stop(ctx) |
| |
| elapsed := time.Since(start) |
| logger.Infow(ctx, "run-time", log.Fields{"Name": "openadapter", "time": elapsed / time.Microsecond}) |
| //logger.Infow(ctx,"run-time", log.Fields{"instanceId": ad.config.InstanceID, "time": elapsed / time.Second}) |
| } |