SEBA-171 Fabric synchronizer resync when kubernetes even received

Change-Id: I66f8aeee81002cc4a248fb5b3bcfa37877c55988
diff --git a/xos/synchronizer/config.yaml b/xos/synchronizer/config.yaml
index c845316..c55603a 100644
--- a/xos/synchronizer/config.yaml
+++ b/xos/synchronizer/config.yaml
@@ -21,6 +21,7 @@
 sys_dir: "/opt/xos/synchronizers/fabric/sys"
 models_dir: "/opt/xos/synchronizers/fabric/models"
 model_policies_dir: "/opt/xos/synchronizers/fabric/model_policies"
+event_steps_dir: "/opt/xos/synchronizers/fabric/event_steps"
 logging:
   version: 1
   handlers:
diff --git a/xos/synchronizer/event_steps/kubernetes_event.py b/xos/synchronizer/event_steps/kubernetes_event.py
new file mode 100644
index 0000000..d571485
--- /dev/null
+++ b/xos/synchronizer/event_steps/kubernetes_event.py
@@ -0,0 +1,74 @@
+
+# Copyright 2017-present Open Networking Foundation
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+import json
+import os
+import sys
+from synchronizers.new_base.eventstep import EventStep
+from synchronizers.new_base.modelaccessor import model_accessor
+from synchronizers.new_base.modelaccessor import FabricService, Switch, Service
+from xosconfig import Config
+from multistructlog import create_logger
+
+log = create_logger(Config().get('logging'))
+
+class KubernetesPodDetailsEventStep(EventStep):
+    topics = ["xos.kubernetes.pod-details"]
+    technology = "kafka"
+
+    def __init__(self, *args, **kwargs):
+        super(KubernetesPodDetailsEventStep, self).__init__(*args, **kwargs)
+
+    @staticmethod
+    def get_fabric_onos(service):
+        service = Service.objects.get(id=service.id)
+        # get the onos_fabric service
+        fabric_onos = [s.leaf_model for s in service.provider_services if "onos" in s.name.lower()]
+
+        if len(fabric_onos) == 0:
+            raise Exception('Cannot find ONOS service in provider_services of Fabric')
+
+        return fabric_onos[0]
+
+    def process_event(self, event):
+        value = json.loads(event.value)
+
+        if (value.get("status") != "created"):
+            return
+
+        if "labels" not in value:
+            return
+
+        xos_service = value["labels"].get("xos_service")
+        if not xos_service:
+            return
+
+        for fabric_service in FabricService.objects.all():
+            onos_service = KubernetesPodDetailsEventStep.get_fabric_onos(fabric_service)
+            if (onos_service.name != xos_service):
+                continue
+
+            for switch in Switch.objects.all():
+                log.info("Dirtying Switch", switch=switch)
+                switch.backend_code=0
+                switch.backend_status="resynchronize due to kubernetes event"
+                switch.save(update_fields=["updated", "backend_code", "backend_status"], always_update_timestamp=True)
+
+                for port in switch.ports.all():
+                    log.info("Dirtying SwitchPort", port=port)
+                    port.backend_code=0
+                    port.backend_status="resynchronize due to kubernetes event"
+                    port.save(update_fields=["updated", "backend_code", "backend_status"], always_update_timestamp=True)
diff --git a/xos/synchronizer/event_steps/test_kubernetes_event.py b/xos/synchronizer/event_steps/test_kubernetes_event.py
new file mode 100644
index 0000000..22b0c95
--- /dev/null
+++ b/xos/synchronizer/event_steps/test_kubernetes_event.py
@@ -0,0 +1,248 @@
+# Copyright 2017-present Open Networking Foundation
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import unittest
+import json
+import functools
+from mock import patch, call, Mock, PropertyMock
+import requests_mock
+
+import os, sys
+
+# Hack to load synchronizer framework
+test_path=os.path.abspath(os.path.dirname(os.path.realpath(__file__)))
+xos_dir=os.path.join(test_path, "../../..")
+if not os.path.exists(os.path.join(test_path, "new_base")):
+    xos_dir=os.path.join(test_path, "../../../../../../orchestration/xos/xos")
+    services_dir = os.path.join(xos_dir, "../../xos_services")
+sys.path.append(xos_dir)
+sys.path.append(os.path.join(xos_dir, 'synchronizers', 'new_base'))
+# END Hack to load synchronizer framework
+
+# generate model from xproto
+def get_models_fn(service_name, xproto_name):
+    name = os.path.join(service_name, "xos", xproto_name)
+    if os.path.exists(os.path.join(services_dir, name)):
+        return name
+    else:
+        name = os.path.join(service_name, "xos", "synchronizer", "models", xproto_name)
+        if os.path.exists(os.path.join(services_dir, name)):
+            return name
+    raise Exception("Unable to find service=%s xproto=%s" % (service_name, xproto_name))
+# END generate model from xproto
+
+class TestKubernetesEvent(unittest.TestCase):
+
+    def setUp(self):
+        global DeferredException
+
+        self.sys_path_save = sys.path
+        sys.path.append(xos_dir)
+        sys.path.append(os.path.join(xos_dir, 'synchronizers', 'new_base'))
+
+        # Setting up the config module
+        from xosconfig import Config
+        config = os.path.join(test_path, "../test_config.yaml")
+        Config.clear()
+        Config.init(config, "synchronizer-config-schema.yaml")
+        # END Setting up the config module
+
+        from synchronizers.new_base.mock_modelaccessor_build import build_mock_modelaccessor
+        build_mock_modelaccessor(xos_dir, services_dir, [
+            get_models_fn("fabric", "fabric.xproto"),
+            get_models_fn("onos-service", "onos.xproto"),
+        ])
+        import synchronizers.new_base.mock_modelaccessor
+        reload(synchronizers.new_base.mock_modelaccessor) # in case nose2 loaded it in a previous test        
+        import synchronizers.new_base.modelaccessor
+        reload(synchronizers.new_base.modelaccessor)      # in case nose2 loaded it in a previous test         
+        from kubernetes_event import model_accessor
+        from mock_modelaccessor import MockObjectList
+
+        from kubernetes_event import KubernetesPodDetailsEventStep
+
+        # import all class names to globals
+        for (k, v) in model_accessor.all_model_classes.items():
+            globals()[k] = v
+
+        self.event_step = KubernetesPodDetailsEventStep
+
+        self.onos = ONOSService(name="myonos",
+                                id=1111,
+                                rest_hostname = "onos-url",
+                                rest_port = "8181",
+                                rest_username = "karaf",
+                                rest_password = "karaf",
+                                backend_code=1,
+                                backend_status="succeeded")
+
+        self.fabric_service = FabricService(name="fabric",
+                                                   id=1112,
+                                                   backend_code=1,
+                                                   backend_status="succeeded",
+                                                   provider_services=[self.onos])
+
+        self.switch = Switch(name="switch1",
+                             backend_code=1,
+                             backend_status="succeeded")
+
+        self.port1 = SwitchPort(name="switch1port1",
+                                switch=self.switch,
+                                backend_code=1,
+                                backend_status="succeeded")
+
+        self.port2 = SwitchPort(name="switch1port2",
+                                switch=self.switch,
+                                backend_code=1,
+                                backend_status="succeeded")
+
+        self.switch.ports = MockObjectList([self.port1, self.port2])
+
+        self.log = Mock()
+
+    def tearDown(self):
+        sys.path = self.sys_path_save
+
+    def test_process_event(self):
+        with patch.object(FabricService.objects, "get_items") as fabric_service_objects, \
+                patch.object(Service.objects, "get_items") as service_objects, \
+                patch.object(Switch.objects, "get_items") as switch_objects, \
+                patch.object(Switch, "save", autospec=True) as switch_save, \
+                patch.object(SwitchPort, "save", autospec=True) as switchport_save:
+            fabric_service_objects.return_value = [self.fabric_service]
+            service_objects.return_value = [self.onos, self.fabric_service]
+            switch_objects.return_value = [self.switch]
+
+            event_dict = {"status": "created",
+                          "labels": {"xos_service": "myonos"}}
+            event = Mock()
+            event.value = json.dumps(event_dict)
+
+            step = self.event_step(log=self.log)
+            step.process_event(event)
+
+            self.assertEqual(self.switch.backend_code, 0)
+            self.assertEqual(self.switch.backend_status, "resynchronize due to kubernetes event")
+
+            switch_save.assert_called_with(self.switch, update_fields=["updated", "backend_code", "backend_status"],
+                                            always_update_timestamp=True)
+
+            self.assertEqual(self.port1.backend_code, 0)
+            self.assertEqual(self.port1.backend_status, "resynchronize due to kubernetes event")
+
+            self.assertEqual(self.port2.backend_code, 0)
+            self.assertEqual(self.port2.backend_status, "resynchronize due to kubernetes event")
+
+            switchport_save.assert_has_calls([call(self.port1, update_fields=["updated", "backend_code", "backend_status"],
+                                            always_update_timestamp=True),
+                                       call(self.port2, update_fields=["updated", "backend_code", "backend_status"],
+                                            always_update_timestamp=True)])
+
+    def test_process_event_unknownstatus(self):
+        with patch.object(FabricService.objects, "get_items") as fabric_service_objects, \
+                patch.object(Service.objects, "get_items") as service_objects, \
+                patch.object(Switch.objects, "get_items") as switch_objects, \
+                patch.object(Switch, "save") as switch_save, \
+                patch.object(SwitchPort, "save") as switchport_save:
+            fabric_service_objects.return_value = [self.fabric_service]
+            service_objects.return_value = [self.onos, self.fabric_service]
+            switch_objects.return_value = [self.switch]
+
+            event_dict = {"status": "something_else",
+                          "labels": {"xos_service": "myonos"}}
+            event = Mock()
+            event.value = json.dumps(event_dict)
+
+            step = self.event_step(log=self.log)
+            step.process_event(event)
+
+            self.assertEqual(self.switch.backend_code, 1)
+            self.assertEqual(self.switch.backend_status, "succeeded")
+
+            switch_save.assert_not_called()
+
+            self.assertEqual(self.port1.backend_code, 1)
+            self.assertEqual(self.port1.backend_status, "succeeded")
+
+            self.assertEqual(self.port2.backend_code, 1)
+            self.assertEqual(self.port2.backend_status, "succeeded")
+
+            switchport_save.assert_not_called()
+
+    def test_process_event_unknownservice(self):
+        with patch.object(FabricService.objects, "get_items") as fabric_service_objects, \
+                patch.object(Service.objects, "get_items") as service_objects, \
+                patch.object(Switch.objects, "get_items") as switch_objects, \
+                patch.object(Switch, "save") as switch_save, \
+                patch.object(SwitchPort, "save") as switchport_save:
+            fabric_service_objects.return_value = [self.fabric_service]
+            service_objects.return_value = [self.onos, self.fabric_service]
+            switch_objects.return_value = [self.switch]
+
+            event_dict = {"status": "created",
+                          "labels": {"xos_service": "something_else"}}
+            event = Mock()
+            event.value = json.dumps(event_dict)
+
+            step = self.event_step(log=self.log)
+            step.process_event(event)
+
+            self.assertEqual(self.switch.backend_code, 1)
+            self.assertEqual(self.switch.backend_status, "succeeded")
+
+            switch_save.assert_not_called()
+
+            self.assertEqual(self.port1.backend_code, 1)
+            self.assertEqual(self.port1.backend_status, "succeeded")
+
+            self.assertEqual(self.port2.backend_code, 1)
+            self.assertEqual(self.port2.backend_status, "succeeded")
+
+            switchport_save.assert_not_called()
+
+    def test_process_event_nolabels(self):
+        with patch.object(FabricService.objects, "get_items") as fabric_service_objects, \
+                patch.object(Service.objects, "get_items") as service_objects, \
+                patch.object(Switch.objects, "get_items") as switch_objects, \
+                patch.object(Switch, "save") as switch_save, \
+                patch.object(SwitchPort, "save") as switchport_save:
+            fabric_service_objects.return_value = [self.fabric_service]
+            service_objects.return_value = [self.onos, self.fabric_service]
+            switch_objects.return_value = [self.switch]
+
+            event_dict = {"status": "created"}
+            event = Mock()
+            event.value = json.dumps(event_dict)
+
+            step = self.event_step(log=self.log)
+            step.process_event(event)
+
+            self.assertEqual(self.switch.backend_code, 1)
+            self.assertEqual(self.switch.backend_status, "succeeded")
+
+            switch_save.assert_not_called()
+
+            self.assertEqual(self.port1.backend_code, 1)
+            self.assertEqual(self.port1.backend_status, "succeeded")
+
+            self.assertEqual(self.port2.backend_code, 1)
+            self.assertEqual(self.port2.backend_status, "succeeded")
+
+            switchport_save.assert_not_called()
+
+if __name__ == '__main__':
+    unittest.main()
+
+
+
diff --git a/xos/unittest.cfg b/xos/unittest.cfg
index 974e092..f28ec79 100644
--- a/xos/unittest.cfg
+++ b/xos/unittest.cfg
@@ -3,6 +3,7 @@
 code-directories=synchronizer
                  steps
                  model_policies
+                 event_steps
 [coverage]
 always-on = True
 coverage = synchronizer