Merge "[AETHER-264] [AETHER-209] Migrate OMEC postmerge job to dev cluster"
diff --git a/jjb/pipeline/voltha-scale-test.groovy b/jjb/pipeline/voltha-scale-test.groovy
index 3ae1ce8..ba9e535 100644
--- a/jjb/pipeline/voltha-scale-test.groovy
+++ b/jjb/pipeline/voltha-scale-test.groovy
@@ -38,7 +38,7 @@
     DEPLOY_K8S="no"
     CONFIG_SADIS="external"
     WITH_KAFKA="kafka.default.svc.cluster.local"
-    WITH_ETCD="etcd-cluster-client.default.svc.cluster.local"
+    WITH_ETCD="etcd.default.svc.cluster.local"
     VOLTHA_ETCD_PORT=9999
 
     // install everything in the default namespace
@@ -85,7 +85,7 @@
               kill -9 \$P_ID
             fi
 
-            for hchart in \$(helm list -q | grep -E -v 'docker-registry|kafkacat|etcd-operator');
+            for hchart in \$(helm list -q | grep -E -v 'docker-registry|kafkacat');
             do
                 echo "Purging chart: \${hchart}"
                 helm delete --purge "\${hchart}"
@@ -150,7 +150,9 @@
         sh '''
         helm install -n kafka incubator/kafka --version 0.13.3 --set replicas=3 --set persistence.enabled=false --set zookeeper.replicaCount=3 --set zookeeper.persistence.enabled=false --version=0.15.3
 
-        helm install --set clusterName=etcd-cluster --set autoCompactionRetention=1 --set clusterSize=3 --set defaults.log_level=WARN --namespace default -n etcd-cluster onf/voltha-etcd-cluster ${extraHelmFlags}
+        # the ETCD chart use "auth" for resons different than BBsim, so strip that away
+        ETCD_FLAGS=$(echo ${extraHelmFlags} | sed -e 's/--set auth=false / /g') | sed -e 's/--set auth=true / /g'
+        helm install -f $WORKSPACE/kind-voltha/minimal-values.yaml --set etcd.replicas=3 -n etcd incubator/etcd $ETCD_FLAGS
 
         if [ ${withMonitoring} = true ] ; then
           helm install -n nem-monitoring cord/nem-monitoring \
@@ -198,6 +200,9 @@
             cd $WORKSPACE/kind-voltha/
 
             ./voltha up
+
+            # Forward the ETCD port onto $VOLTHA_ETCD_PORT
+            _TAG=etcd-port-forward kubectl port-forward --address 0.0.0.0 -n default service/etcd $VOLTHA_ETCD_PORT:2379&
           """
         }
         // bbsim-sadis server takes a while to cache the subscriber entries
@@ -243,11 +248,10 @@
           if [ ${withMibTemplate} = true ] ; then
             rm -f BBSM-12345123451234512345-00000000000001-v1.json
             wget https://raw.githubusercontent.com/opencord/voltha-openonu-adapter/master/templates/BBSM-12345123451234512345-00000000000001-v1.json
-            cat BBSM-12345123451234512345-00000000000001-v1.json | kubectl exec -it $(kubectl get pods | grep etcd-cluster | awk 'NR==1{print $1}') etcdctl put service/voltha/omci_mibs/templates/BBSM/12345123451234512345/00000000000001
+            cat BBSM-12345123451234512345-00000000000001-v1.json | kubectl exec -it $(kubectl get pods -l app=etcd | awk 'NR==2{print $1}') etcdctl put service/voltha/omci_mibs/templates/BBSM/12345123451234512345/00000000000001
           fi
 
           # Set extra openolt-adapter logs
-          _TAG=etcd-port-forward kubectl port-forward --address 0.0.0.0 -n default service/etcd-cluster-client 9999:2379&
           voltctl log level set INFO adapter-open-olt
           voltctl log level set WARN adapter-open-olt#github.com/opencord/voltha-lib-go/v3/pkg/db
           voltctl log level set WARN adapter-open-olt#github.com/opencord/voltha-lib-go/v3/pkg/probe
@@ -461,6 +465,8 @@
           sh '''
           voltctl device list -o json > $WORKSPACE/logs/device-list.json
           python -m json.tool $WORKSPACE/logs/device-list.json > $WORKSPACE/logs/voltha-devices-list.json
+          rm $WORKSPACE/logs/device-list.json
+          voltctl device list > $WORKSPACE/logs/voltha-devices-list.txt
 
           printf '%s\n' $(voltctl device list | grep olt | awk '{print $1}') | xargs -I@ bash -c "voltctl device flows @ > $WORKSPACE/logs/voltha-device-flows-@.txt"
           printf '%s\n' $(voltctl device list | grep olt | awk '{print $1}') | xargs -I@ bash -c "voltctl device port list --format 'table{{.PortNo}}\t{{.Label}}\t{{.Type}}\t{{.AdminState}}\t{{.OperStatus}}' @ > $WORKSPACE/logs/voltha-device-ports-@.txt"
diff --git a/jjb/voltha-scale.yaml b/jjb/voltha-scale.yaml
index a02386c..2456580 100644
--- a/jjb/voltha-scale.yaml
+++ b/jjb/voltha-scale.yaml
@@ -95,21 +95,23 @@
       #     withIgmp: false
 
       # jobs for 512 ONUs with a 8 openonu-adapters and clustered ONOS
-      # - 'voltha-scale-measurements':
-      #     name: 'voltha-scale-measurements-1-16-32-activation'
-      #     build-node: 'onf-pod1-head-node'
-      #     time-trigger: "H H/4 * * *"
-      #     olts: 1
-      #     pons: 16
-      #     onus: 32
-      #     withFlows: false
-      #     provisionSubscribers: false
-      #     withEapol: false
-      #     withDhcp: false
-      #     withIgmp: false
-      #     openonuAdapterReplicas: 8
-      #     onosReplicas: 3
-      #     atomixReplicas: 3
+      - 'voltha-scale-measurements':
+          name: 'voltha-scale-measurements-1-16-32-activation'
+          build-node: 'onf-pod1-head-node'
+          time-trigger: "H H/4 * * *"
+          olts: 1
+          pons: 16
+          onus: 32
+          withFlows: false
+          provisionSubscribers: false
+          withEapol: false
+          withDhcp: false
+          withIgmp: false
+          openonuAdapterReplicas: 8
+          onosReplicas: 3
+          atomixReplicas: 3
+          extraHelmFlags: '--set limits.memory=8192Mi '
+
       - 'voltha-scale-measurements':
           name: 'voltha-scale-measurements-1-16-32-att-flows'
           build-node: 'onf-pod1-head-node'
@@ -125,7 +127,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set auth=false --set memoryLimit=8192Mi '
+          extraHelmFlags: '--set auth=false --set limits.memory=8192Mi '
 
 
       - 'voltha-scale-measurements':
@@ -143,7 +145,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi --set defaults.rw_core.timeout=30s'
+          extraHelmFlags: '--set limits.memory=8192Mi --set defaults.rw_core.timeout=30s'
 
 
       # jobs for DT with 512 ONUs with a 8 openonu-adapters and clustered ONOS.
@@ -163,7 +165,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi '
+          extraHelmFlags: '--set limits.memory=8192Mi '
 
 
       # jobs for 1024 ONUs with a 8 openonu-adapters and clustered ONOS (2 OLTs)
@@ -183,7 +185,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi '
+          extraHelmFlags: '--set limits.memory=8192Mi '
 
       - 'voltha-scale-measurements':
           disabled: true
@@ -201,7 +203,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi --set auth=false '
+          extraHelmFlags: '--set limits.memory=8192Mi --set auth=false '
 
       - 'voltha-scale-measurements':
           disabled: true
@@ -219,7 +221,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi --set defaults.rw_core.timeout=30s'
+          extraHelmFlags: '--set limits.memory=8192Mi --set defaults.rw_core.timeout=30s'
 
       # jobs for DT with 512 ONUs with a 8 openonu-adapters and clustered ONOS.
       - 'voltha-scale-measurements':
@@ -238,7 +240,7 @@
           openonuAdapterReplicas: 8
           onosReplicas: 3
           atomixReplicas: 3
-          extraHelmFlags: '--set memoryLimit=8192Mi --set defaults.rw_core.timeout=30s'
+          extraHelmFlags: '--set limits.memory=8192Mi --set defaults.rw_core.timeout=30s'
 
       # development pipeline
       - 'voltha-scale-measurements-dev':