Enabling monitoring in experimental-multi-stack scale job
Change-Id: I09efef324be815852443267ea4b792786aee4a45
diff --git a/jjb/pipeline/voltha-scale-multi-stack.groovy b/jjb/pipeline/voltha-scale-multi-stack.groovy
index 7b4a949..bdb938c 100644
--- a/jjb/pipeline/voltha-scale-multi-stack.groovy
+++ b/jjb/pipeline/voltha-scale-multi-stack.groovy
@@ -184,7 +184,7 @@
steps {
sh '''
if [ ${withMonitoring} = true ] ; then
- helm install nem-monitoring cord/nem-monitoring \
+ helm install -n $INFRA_NS nem-monitoring cord/nem-monitoring \
-f $HOME/voltha-scale/grafana.yaml \
--set prometheus.alertmanager.enabled=false,prometheus.pushgateway.enabled=false \
--set kpi_exporter.enabled=false,dashboards.xos=false,dashboards.onos=false,dashboards.aaa=false,dashboards.voltha=false
@@ -451,7 +451,7 @@
helm ls --all-namespaces > $LOG_FOLDER/helm-list.txt || true
# store information on the running pods
- kubectl get pods -o wide > $LOG_FOLDER/pods.txt || true
+ kubectl get pods --all-namespaces -o wide > $LOG_FOLDER/pods.txt || true
kubectl get pods --all-namespaces -o jsonpath="{range .items[*].status.containerStatuses[*]}{.image}{'\\n'}" | sort | uniq | tee $LOG_FOLDER/pod-images.txt || true
kubectl get pods --all-namespaces -o jsonpath="{range .items[*].status.containerStatuses[*]}{.imageID}{'\\n'}" | sort | uniq | tee $LOG_FOLDER/pod-imagesId.txt || true