blob: e7723fda95b54549f05e8aef15625328e186d479 [file] [log] [blame]
/*
* Copyright 2019-present Ciena Corporation
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package commands
import (
"context"
"fmt"
"github.com/fullstorydev/grpcurl"
flags "github.com/jessevdk/go-flags"
"github.com/jhump/protoreflect/dynamic"
"github.com/opencord/voltctl/pkg/format"
"github.com/opencord/voltctl/pkg/model"
"google.golang.org/grpc"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/tools/clientcmd"
"strings"
)
type SetLogLevelOutput struct {
ComponentName string
Status string
Error string
}
type SetLogLevelOpts struct {
OutputOptions
Package string `short:"p" long:"package" description:"Package name to set log level"`
Args struct {
Level string
Component []string
} `positional-args:"yes" required:"yes"`
}
type GetLogLevelsOpts struct {
ListOutputOptions
Args struct {
Component []string
} `positional-args:"yes" required:"yes"`
}
type ListLogLevelsOpts struct {
ListOutputOptions
}
type LogLevelOpts struct {
SetLogLevel SetLogLevelOpts `command:"set"`
GetLogLevels GetLogLevelsOpts `command:"get"`
ListLogLevels ListLogLevelsOpts `command:"list"`
}
var logLevelOpts = LogLevelOpts{}
const (
DEFAULT_LOGLEVELS_FORMAT = "table{{ .ComponentName }}\t{{.PackageName}}\t{{.Level}}"
DEFAULT_SETLOGLEVEL_FORMAT = "table{{ .ComponentName }}\t{{.Status}}\t{{.Error}}"
)
func RegisterLogLevelCommands(parent *flags.Parser) {
_, err := parent.AddCommand("loglevel", "loglevel commands", "Get and set log levels", &logLevelOpts)
if err != nil {
panic(err)
}
}
func MapListAppend(m map[string][]string, name string, item string) {
list, okay := m[name]
if okay {
m[name] = append(list, item)
} else {
m[name] = []string{item}
}
}
/*
* A roundabout way of going of using the LogLevel enum to map from
* a string to an integer that we can pass into the dynamic
* proto.
*
* TODO: There's probably an easier way.
*/
func LogLevelStringToInt(logLevelString string) (int32, error) {
ProcessGlobalOptions() // required for GetMethod()
/*
* Use GetMethod() to get us a descriptor on the proto file we're
* interested in.
*/
descriptor, _, err := GetMethod("update-log-level")
if err != nil {
return 0, err
}
/*
* Map string LogLevel to enumerated type LogLevel
* We have descriptor from above, which is a DescriptorSource
* We can use FindSymbol to get at the message
*/
loggingSymbol, err := descriptor.FindSymbol("common.LogLevel")
if err != nil {
return 0, err
}
/*
* LoggingSymbol is a Descriptor, but not a MessageDescrptior,
* so we can't look at it's fields yet. Go back to the file,
* call FindMessage to get the Message, then we can get the
* embedded enum.
*/
loggingFile := loggingSymbol.GetFile()
logLevelMessage := loggingFile.FindMessage("common.LogLevel")
logLevelEnumType := logLevelMessage.GetNestedEnumTypes()[0]
enumLogLevel := logLevelEnumType.FindValueByName(logLevelString)
if enumLogLevel == nil {
return 0, fmt.Errorf("Unknown log level %s", logLevelString)
}
return enumLogLevel.GetNumber(), nil
}
// Validate a list of component names and throw an error if any one of them is bad.
func ValidateComponentNames(kube_to_arouter map[string][]string, names []string) error {
var badNames []string
for _, name := range names {
_, ok := kube_to_arouter[name]
if !ok {
badNames = append(badNames, name)
}
}
if len(badNames) > 0 {
allowedNames := make([]string, len(kube_to_arouter))
i := 0
for k := range kube_to_arouter {
allowedNames[i] = k
i++
}
return fmt.Errorf("Unknown component(s): %s.\n (Allowed values for component names: \n %s)",
strings.Join(badNames, ", "),
strings.Join(allowedNames, ",\n "))
} else {
return nil
}
}
func BuildKubernetesNameMap() (map[string][]string, map[string]string, error) {
kube_to_arouter := make(map[string][]string)
arouter_to_kube := make(map[string]string)
// use the current context in kubeconfig
config, err := clientcmd.BuildConfigFromFlags("", GlobalOptions.K8sConfig)
if err != nil {
return nil, nil, err
}
// create the clientset
clientset, err := kubernetes.NewForConfig(config)
if err != nil {
return nil, nil, err
}
pods, err := clientset.CoreV1().Pods("").List(metav1.ListOptions{
LabelSelector: "app.kubernetes.io/part-of=voltha",
})
if err != nil {
return nil, nil, err
}
if len(pods.Items) == 0 {
return nil, nil, fmt.Errorf("No Voltha pods found in Kubernetes -- verify pod is setup")
}
for _, pod := range pods.Items {
app, ok := pod.Labels["app"]
if !ok {
continue
}
var arouter_name string
switch app {
case "voltha-api-server":
/*
* Assumes a single api_server for now.
* TODO: Make labeling changes in charts to be able to derive name from labels
*/
arouter_name = "api_server0.api_server01"
case "rw-core":
affinity_group, ok := pod.Labels["affinity-group"]
if !ok {
Warn.Printf("rwcore %s lacks affinity-group label", pod.Name)
continue
}
affinity_group_core_id, ok := pod.Labels["affinity-group-core-id"]
if !ok {
Warn.Printf("rwcore %s lacks affinity-group-core-id label", pod.Name)
continue
}
arouter_name = "vcore" + affinity_group + ".vcore" + affinity_group + affinity_group_core_id
case "ro-core":
/*
* Assumes a single rocore for now.
* TODO: Make labeling changes in charts to be able to derive name from labels
*/
arouter_name = "ro_vcore0.ro_vcore01"
default:
// skip this pod as it's not relevant
continue
}
// Multiple ways to identify the component
// 1) The pod name. One pod name maps to exactly one pod.
arouter_to_kube[arouter_name] = pod.Name
MapListAppend(kube_to_arouter, pod.Name, arouter_name)
// 2) The kubernetes component name. A single component (i.e. "core") may map to multiple pods.
component, ok := pod.Labels["app.kubernetes.io/component"]
if ok {
MapListAppend(kube_to_arouter, component, arouter_name)
}
// 3) The voltha app label. A single app (i.e. "rwcore") may map to multiple pods.
MapListAppend(kube_to_arouter, app, arouter_name)
}
return kube_to_arouter, arouter_to_kube, nil
}
func (options *SetLogLevelOpts) Execute(args []string) error {
if len(options.Args.Component) == 0 {
return fmt.Errorf("Please specify at least one component")
}
kube_to_arouter, arouter_to_kube, err := BuildKubernetesNameMap()
if err != nil {
return err
}
var output []SetLogLevelOutput
// Validate component names, throw error now to avoid doing partial work
err = ValidateComponentNames(kube_to_arouter, options.Args.Component)
if err != nil {
return err
}
// Validate and map the logLevel string to an integer, throw error now to avoid doing partial work
intLogLevel, err := LogLevelStringToInt(options.Args.Level)
if err != nil {
return err
}
for _, kubeComponentName := range options.Args.Component {
var descriptor grpcurl.DescriptorSource
var conn *grpc.ClientConn
var method string
componentNameList := kube_to_arouter[kubeComponentName]
for _, componentName := range componentNameList {
conn, err = NewConnection()
if err != nil {
return err
}
defer conn.Close()
if strings.HasPrefix(componentName, "api_server") {
// apiserver's UpdateLogLevel is in the afrouter.Configuration gRPC package
descriptor, method, err = GetMethod("apiserver-update-log-level")
} else {
descriptor, method, err = GetMethod("update-log-level")
}
if err != nil {
return err
}
ctx, cancel := context.WithTimeout(context.Background(), GlobalConfig.Grpc.Timeout)
defer cancel()
ll := make(map[string]interface{})
ll["component_name"] = componentName
ll["package_name"] = options.Package
ll["level"] = intLogLevel
h := &RpcEventHandler{
Fields: map[string]map[string]interface{}{"common.Logging": ll},
}
err = grpcurl.InvokeRPC(ctx, descriptor, conn, method, []string{}, h, h.GetParams)
if err != nil {
return err
}
if h.Status != nil && h.Status.Err() != nil {
output = append(output, SetLogLevelOutput{ComponentName: arouter_to_kube[componentName], Status: "Failure", Error: h.Status.Err().Error()})
continue
}
output = append(output, SetLogLevelOutput{ComponentName: arouter_to_kube[componentName], Status: "Success"})
}
}
outputFormat := CharReplacer.Replace(options.Format)
if outputFormat == "" {
outputFormat = GetCommandOptionWithDefault("loglevel-set", "format", DEFAULT_SETLOGLEVEL_FORMAT)
}
result := CommandResult{
Format: format.Format(outputFormat),
OutputAs: toOutputType(options.OutputAs),
NameLimit: options.NameLimit,
Data: output,
}
GenerateOutput(&result)
return nil
}
func (options *GetLogLevelsOpts) getLogLevels(methodName string, args []string) error {
if len(options.Args.Component) == 0 {
return fmt.Errorf("Please specify at least one component")
}
kube_to_arouter, arouter_to_kube, err := BuildKubernetesNameMap()
if err != nil {
return err
}
var data []model.LogLevel
// Validate component names, throw error now to avoid doing partial work
err = ValidateComponentNames(kube_to_arouter, options.Args.Component)
if err != nil {
return err
}
for _, kubeComponentName := range options.Args.Component {
var descriptor grpcurl.DescriptorSource
var conn *grpc.ClientConn
var method string
componentNameList := kube_to_arouter[kubeComponentName]
for _, componentName := range componentNameList {
conn, err = NewConnection()
if err != nil {
return err
}
defer conn.Close()
if strings.HasPrefix(componentName, "api_server") {
// apiserver's UpdateLogLevel is in the afrouter.Configuration gRPC package
descriptor, method, err = GetMethod("apiserver-get-log-levels")
} else {
descriptor, method, err = GetMethod("get-log-levels")
}
if err != nil {
return err
}
ctx, cancel := context.WithTimeout(context.Background(), GlobalConfig.Grpc.Timeout)
defer cancel()
ll := make(map[string]interface{})
ll["component_name"] = componentName
h := &RpcEventHandler{
Fields: map[string]map[string]interface{}{"common.LoggingComponent": ll},
}
err = grpcurl.InvokeRPC(ctx, descriptor, conn, method, []string{}, h, h.GetParams)
if err != nil {
return err
}
if h.Status != nil && h.Status.Err() != nil {
return h.Status.Err()
}
d, err := dynamic.AsDynamicMessage(h.Response)
if err != nil {
return err
}
items, err := d.TryGetFieldByName("items")
if err != nil {
return err
}
for _, item := range items.([]interface{}) {
logLevel := model.LogLevel{}
logLevel.PopulateFrom(item.(*dynamic.Message))
logLevel.ComponentName = arouter_to_kube[logLevel.ComponentName]
data = append(data, logLevel)
}
}
}
outputFormat := CharReplacer.Replace(options.Format)
if outputFormat == "" {
outputFormat = GetCommandOptionWithDefault(methodName, "format", DEFAULT_LOGLEVELS_FORMAT)
}
orderBy := options.OrderBy
if orderBy == "" {
orderBy = GetCommandOptionWithDefault(methodName, "order", "")
}
result := CommandResult{
Format: format.Format(outputFormat),
Filter: options.Filter,
OrderBy: orderBy,
OutputAs: toOutputType(options.OutputAs),
NameLimit: options.NameLimit,
Data: data,
}
GenerateOutput(&result)
return nil
}
func (options *GetLogLevelsOpts) Execute(args []string) error {
return options.getLogLevels("loglevel-get", args)
}
func (options *ListLogLevelsOpts) Execute(args []string) error {
var getOptions GetLogLevelsOpts
var podNames []string
_, arouter_to_kube, err := BuildKubernetesNameMap()
if err != nil {
return err
}
for _, podName := range arouter_to_kube {
podNames = append(podNames, podName)
}
// Just call GetLogLevels with a list of podnames that includes everything relevant.
getOptions.ListOutputOptions = options.ListOutputOptions
getOptions.Args.Component = podNames
return getOptions.getLogLevels("loglevel-list", args)
}