blob: 992e1b6c563a5c331db297cc77d09ff8142aa0e3 [file] [log] [blame]
#
# Copyright 2017 the original author or authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
"""
Broadcom OLT/ONU adapter.
"""
from uuid import uuid4
import structlog
from twisted.internet import reactor, task
from twisted.internet.defer import DeferredQueue, inlineCallbacks, returnValue
from zope.interface import implementer
from common.utils.asleep import asleep
from voltha.adapters.interface import IAdapterInterface
from voltha.core.logical_device_agent import mac_str_to_tuple
import voltha.core.flow_decomposer as fd
from voltha.protos import third_party
from voltha.protos.adapter_pb2 import Adapter
from voltha.protos.adapter_pb2 import AdapterConfig
from voltha.protos.common_pb2 import LogLevel, OperStatus, ConnectStatus, \
AdminState
from voltha.protos.device_pb2 import DeviceType, DeviceTypes, Port, Image
from voltha.protos.health_pb2 import HealthStatus
from voltha.protos.logical_device_pb2 import LogicalPort
from voltha.protos.openflow_13_pb2 import OFPPS_LIVE, OFPPF_FIBER, OFPPF_1GB_FD, OFPPS_LINK_DOWN
from voltha.protos.openflow_13_pb2 import OFPXMC_OPENFLOW_BASIC, ofp_port, \
ofp_port_stats
from voltha.protos.bbf_fiber_base_pb2 import VEnetConfig, VOntaniConfig
from voltha.protos.bbf_fiber_traffic_descriptor_profile_body_pb2 import \
TrafficDescriptorProfileData
from voltha.protos.bbf_fiber_tcont_body_pb2 import TcontsConfigData
from voltha.protos.bbf_fiber_gemport_body_pb2 import GemportsConfigData
from common.frameio.frameio import hexify
from voltha.extensions.omci.omci import *
from voltha.registry import registry
_ = third_party
log = structlog.get_logger()
MANAGEMENT_VLAN = 4090
BRDCM_DEFAULT_VLAN = 4091
ADMIN_STATE_LOCK = 1
ADMIN_STATE_UNLOCK = 0
RESERVED_VLAN_ID = 4095
FLOW_TYPE_EAPOL = 34958
@implementer(IAdapterInterface)
class BroadcomOnuAdapter(object):
name = 'broadcom_onu'
supported_device_types = [
DeviceType(
id=name,
vendor_ids=['NONE', 'BBSM'],
adapter=name,
accepts_bulk_flow_update=True,
accepts_add_remove_flow_updates=True
)
]
def __init__(self, adapter_agent, config):
self.adapter_agent = adapter_agent
self.config = config
self.descriptor = Adapter(
id=self.name,
vendor='Voltha project',
version='0.46',
config=AdapterConfig(log_level=LogLevel.INFO)
)
self.devices_handlers = dict() # device_id -> BroadcomOnuHandler()
# register for adapter messages
self.adapter_agent.register_for_inter_adapter_messages()
def start(self):
log.debug('starting')
log.info('started')
def stop(self):
log.debug('stopping')
log.info('stopped')
def adapter_descriptor(self):
return self.descriptor
def device_types(self):
return DeviceTypes(items=self.supported_device_types)
def health(self):
return HealthStatus(state=HealthStatus.HealthState.HEALTHY)
def change_master_state(self, master):
raise NotImplementedError()
def adopt_device(self, device):
log.info('adopt_device', device_id=device.id)
self.devices_handlers[device.id] = BroadcomOnuHandler(self, device.id)
reactor.callLater(0, self.devices_handlers[device.id].activate, device)
return device
def reconcile_device(self, device):
log.info('reconcile-device', device_id=device.id)
self.devices_handlers[device.id] = BroadcomOnuHandler(self, device.id)
reactor.callLater(0, self.devices_handlers[device.id].reconcile, device)
def abandon_device(self, device):
log.info('abadon-device - Not implemented')
raise NotImplementedError()
def disable_device(self, device):
log.info('disable-onu-device', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.disable(device)
def reenable_device(self, device):
log.info('reenable-onu-device', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.reenable(device)
def reboot_device(self, device):
log.info('reboot-device', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.reboot()
def download_image(self, device, request):
raise NotImplementedError()
def get_image_download_status(self, device, request):
raise NotImplementedError()
def cancel_image_download(self, device, request):
raise NotImplementedError()
def activate_image_update(self, device, request):
raise NotImplementedError()
def revert_image_update(self, device, request):
raise NotImplementedError()
def self_test_device(self, device):
"""
This is called to Self a device based on a NBI call.
:param device: A Voltha.Device object.
:return: Will return result of self test
"""
log.info('self-test-device - Not implemented', device=device.id)
raise NotImplementedError()
def delete_device(self, device):
log.info('delete-device', device_id=device.id, device_handlers=self.devices_handlers)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
log.debug('calling-handler-delete', handler=handler)
handler.delete(device)
del self.devices_handlers[device.id]
return
def get_device_details(self, device):
raise NotImplementedError()
def update_pm_config(self, device, pm_configs):
raise NotImplementedError()
def update_flows_bulk(self, device, flows, groups):
'''
log.info('bulk-flow-update', device_id=device.id,
flows=flows, groups=groups)
'''
assert len(groups.items) == 0
handler = self.devices_handlers[device.id]
return handler.update_flow_table(device, flows.items)
def update_flows_incrementally(self, device, flow_changes, group_changes):
log.info('incremental-flow-update', device_id=device.id,
flows=flow_changes, groups=group_changes)
# For now, there is no support for group changes
assert len(group_changes.to_add.items) == 0
assert len(group_changes.to_remove.items) == 0
handler = self.devices_handlers[device.id]
# Remove flows
if len(flow_changes.to_remove.items) != 0:
handler.remove_from_flow_table(flow_changes.to_remove.items)
# Add flows
if len(flow_changes.to_add.items) != 0:
handler.add_to_flow_table(flow_changes.to_add.items)
def send_proxied_message(self, proxy_address, msg):
log.debug('send-proxied-message', proxy_address=proxy_address, msg=msg)
def receive_proxied_message(self, proxy_address, msg):
log.debug('receive-proxied-message', proxy_address=proxy_address,
device_id=proxy_address.device_id, msg=hexify(msg))
# Device_id from the proxy_address is the olt device id. We need to
# get the onu device id using the port number in the proxy_address
device = self.adapter_agent. \
get_child_device_with_proxy_address(proxy_address)
if device:
handler = self.devices_handlers[device.id]
handler.receive_message(msg)
def receive_packet_out(self, logical_device_id, egress_port_no, msg):
log.info('packet-out', logical_device_id=logical_device_id,
egress_port_no=egress_port_no, msg_len=len(msg))
def receive_inter_adapter_message(self, msg):
log.debug('receive_inter_adapter_message', msg=msg)
proxy_address = msg['proxy_address']
assert proxy_address is not None
# Device_id from the proxy_address is the olt device id. We need to
# get the onu device id using the port number in the proxy_address
device = self.adapter_agent. \
get_child_device_with_proxy_address(proxy_address)
if device:
handler = self.devices_handlers[device.id]
handler.event_messages.put(msg)
else:
log.error("device-not-found")
def create_interface(self, device, data):
log.debug('create-interface', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.create_interface(data)
def update_interface(self, device, data):
log.debug('update-interface', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.update_interface(data)
def remove_interface(self, device, data):
log.debug('remove-interface', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.remove_interface(data)
def receive_onu_detect_state(self, device_id, state):
raise NotImplementedError()
def create_tcont(self, device, tcont_data, traffic_descriptor_data):
log.debug('create-tcont', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.create_tcont(tcont_data, traffic_descriptor_data)
def update_tcont(self, device, tcont_data, traffic_descriptor_data):
log.info('update_tcont not implemented in onu')
def remove_tcont(self, device, tcont_data, traffic_descriptor_data):
log.debug('remove-tcont', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.remove_tcont(tcont_data, traffic_descriptor_data)
def create_gemport(self, device, data):
log.debug('create-gemport', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.create_gemport(data)
def update_gemport(self, device, data):
raise NotImplementedError()
def remove_gemport(self, device, data):
log.debug('remove-gemport', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.remove_gemport(data)
def create_multicast_gemport(self, device, data):
log.debug('create-multicast-gemport', device_id=device.id)
if device.id in self.devices_handlers:
handler = self.devices_handlers[device.id]
if handler is not None:
handler.create_multicast_gemport(data)
def update_multicast_gemport(self, device, data):
raise NotImplementedError()
def remove_multicast_gemport(self, device, data):
raise NotImplementedError()
def create_multicast_distribution_set(self, device, data):
raise NotImplementedError()
def update_multicast_distribution_set(self, device, data):
raise NotImplementedError()
def remove_multicast_distribution_set(self, device, data):
raise NotImplementedError()
def suppress_alarm(self, filter):
raise NotImplementedError()
def unsuppress_alarm(self, filter):
raise NotImplementedError()
class BroadcomOnuHandler(object):
def __init__(self, adapter, device_id):
self.adapter = adapter
self.adapter_agent = adapter.adapter_agent
self.device_id = device_id
self.log = structlog.get_logger(device_id=device_id)
self.incoming_messages = DeferredQueue()
self.event_messages = DeferredQueue()
self.proxy_address = None
self.tx_id = 0
self.flow_map = dict()
# Need to query ONU for number of supported uni ports
# For now, temporarily set number of ports to 1 - port #2
self.uni_ports = (1, 2, 3, 4, 5)
self.flow_config_in_progress = False
# Handle received ONU event messages
reactor.callLater(0, self.handle_onu_events)
def receive_message(self, msg):
self.incoming_messages.put(msg)
@inlineCallbacks
def handle_onu_events(self):
event_msg = yield self.event_messages.get()
if event_msg['event'] == 'activation-completed':
if event_msg['event_data']['activation_successful'] == True:
yield self.message_exchange()
device = self.adapter_agent.get_device(self.device_id)
device.connect_status = ConnectStatus.REACHABLE
device.oper_status = OperStatus.ACTIVE
self.adapter_agent.update_device(device)
else:
device = self.adapter_agent.get_device(self.device_id)
self.disable_ports(device)
device.connect_status = ConnectStatus.UNREACHABLE
device.oper_status = OperStatus.FAILED
self.adapter_agent.update_device(device)
self.flow_map.clear()
elif event_msg['event'] == 'deactivation-completed':
device = self.adapter_agent.get_device(self.device_id)
device.oper_status = OperStatus.DISCOVERED
self.adapter_agent.update_device(device)
elif event_msg['event'] == 'deactivate-onu':
device = self.adapter_agent.get_device(self.device_id)
self.disable_ports(device)
device.connect_status = ConnectStatus.UNREACHABLE
device.oper_status = OperStatus.DISCOVERED
self.adapter_agent.update_device(device)
self.flow_map.clear()
elif (event_msg['event'] == 'olt-reboot'):
device = self.adapter_agent.get_device(self.device_id)
device.connect_status = ConnectStatus.UNREACHABLE
self.adapter_agent.update_device(device)
self.flow_map.clear()
elif event_msg['event'] == 'ranging-completed':
if event_msg['event_data']['ranging_successful'] == True:
device = self.adapter_agent.get_device(self.device_id)
device.oper_status = OperStatus.ACTIVATING
self.adapter_agent.update_device(device)
else:
device = self.adapter_agent.get_device(self.device_id)
device.oper_status = OperStatus.FAILED
self.adapter_agent.update_device(device)
elif event_msg['event'] == 'olt-disabled':
self.adapter_agent.disable_all_ports(self.device_id)
device = self.adapter_agent.get_device(self.device_id)
device.connect_status = ConnectStatus.UNREACHABLE
self.adapter_agent.update_device(device)
self.flow_map.clear()
elif event_msg['event'] == 'olt-enabled':
self.adapter_agent.enable_all_ports(self.device_id)
device = self.adapter_agent.get_device(self.device_id)
device.connect_status = ConnectStatus.REACHABLE
self.adapter_agent.update_device(device)
elif event_msg['event'] == 'create-tcont':
tcont = TcontsConfigData()
tcont.alloc_id = event_msg['event_data']['alloc_id']
self.create_tcont(tcont, traffic_descriptor_data=None)
elif event_msg['event'] == 'create-venet':
venet = VEnetConfig(name=event_msg['event_data']['uni_name'])
venet.interface.name = event_msg['event_data']['interface_name']
self.create_interface(venet)
elif event_msg['event'] == 'create-gemport':
gem_port = GemportsConfigData()
gem_port.gemport_id = event_msg['event_data']['gemport_id']
self.create_gemport(gem_port)
# Handle next event
reactor.callLater(0, self.handle_onu_events)
def activate(self, device):
self.log.info('activating')
# first we verify that we got parent reference and proxy info
assert device.parent_id
assert device.proxy_address.device_id
#assert device.proxy_address.channel_id # c-vid
# register for proxied messages right away
self.proxy_address = device.proxy_address
self.adapter_agent.register_for_proxied_messages(device.proxy_address)
# populate device info
device.root = True
device.vendor = 'Broadcom'
device.model = 'n/a'
device.hardware_version = 'to be filled'
device.firmware_version = 'to be filled'
device.images.image.extend([
Image(version="to be filled")
])
device.connect_status = ConnectStatus.REACHABLE
self.adapter_agent.update_device(device)
self.adapter_agent.add_port(device.id, Port(
port_no=100,
label='PON port',
type=Port.PON_ONU,
admin_state=AdminState.ENABLED,
oper_status=OperStatus.ACTIVE,
peers=[
Port.PeerPort(
device_id=device.parent_id,
port_no=device.parent_port_no
)
]
))
parent_device = self.adapter_agent.get_device(device.parent_id)
logical_device_id = parent_device.parent_id
assert logical_device_id
device = self.adapter_agent.get_device(device.id)
device.oper_status = OperStatus.DISCOVERED
self.adapter_agent.update_device(device)
self.log.info('activated')
def reconcile(self, device):
self.log.debug('reconciling-broadcom-onu-device-starts')
# first we verify that we got parent reference and proxy info
assert device.parent_id
assert device.proxy_address.device_id
# register for proxied messages right away
self.proxy_address = device.proxy_address
self.adapter_agent.register_for_proxied_messages(device.proxy_address)
# TODO: Query ONU current status after reconcile and update.
# To be addressed in future commits.
self.log.info('reconciling-broadcom-onu-device-ends')
def update_logical_port(self, logical_device_id, port_id, state):
try:
self.log.debug('updating-logical-port', logical_port_id=port_id,
logical_device_id=logical_device_id, state=state)
logical_port = self.adapter_agent.get_logical_port(logical_device_id,
port_id)
logical_port.ofp_port.state = state
self.adapter_agent.update_logical_port(logical_device_id,
logical_port)
except Exception as e:
self.log.exception("exception-updating-port",e=e)
def delete(self, device):
self.log.info('delete-onu', device=device)
parent_device = self.adapter_agent.get_device(device.parent_id)
if parent_device.type == 'openolt':
parent_adapter = registry('adapter_loader').get_agent(parent_device.adapter).adapter
self.log.debug('parent-adapter-delete-onu', onu_device=device,
parent_device=parent_device,
parent_adapter=parent_adapter)
try:
parent_adapter.delete_child_device(parent_device.id, device)
except AttributeError:
self.log.debug('parent-device-delete-child-not-implemented')
@inlineCallbacks
def update_flow_table(self, device, flows):
# Twisted code is not inherently thread safe. This API could
# be invoked again by reactor while previous one is in progress.
# Since we maintain some stateful information here, we better
# synchronize parallel invocations on this API.
yield self._wait_for_previous_update_flow_to_finish()
self.flow_config_in_progress = True
#
# We need to proxy through the OLT to get to the ONU
# Configuration from here should be using OMCI
#
try:
# Calculates flows_to_adds and flows_to_delete using
# incoming_flows and current_flows.
current_flows = set(self.flow_map.keys())
# Only adding those flows to incoming_flows which are having cookie.
incoming_flows = set(flow.cookie for flow in flows if flow.cookie)
flows_to_add = incoming_flows.difference(current_flows)
flows_to_delete = current_flows.difference(incoming_flows)
# Sends request to delete flows for ONU flows in flows_to_delete list.
for cookie in flows_to_delete:
if cookie in self.flow_map:
c_tag = self.flow_map[cookie]["vlan_id"]
self.log.debug("flow-to-delete-cookie",
cookie=cookie, c_tag=c_tag)
# Deleting flow from ONU.
yield self._delete_onu_flow(self.flow_map[cookie])
# Removing flow_map entry for deleted flow.
del self.flow_map[cookie]
else:
self.log.info('ignoring-cookie_received', cookie=cookie)
# If flow is not in flow_to_add, no need to add flow to ONU.
for flow in flows:
if flow.cookie not in flows_to_add:
self.log.debug("flow-not-to-be-added", cookie=flow.cookie)
continue
# Adding flow to ONU.
yield self._add_onu_flow(flow)
except Exception as e:
self.log.exception('failed-to-update-flow-table', e=e)
self.flow_config_in_progress = False
@inlineCallbacks
def add_to_flow_table(self, flows):
"""
This function is called for update_flows_incrementally to add
only delta flows to ONU.
:param flows: flows to add to ONU.
"""
yield self._wait_for_previous_update_flow_to_finish()
self.flow_config_in_progress = True
self.log.debug('add-to-flow-table', flows=flows)
try:
for flow in flows:
# if incoming flow contains cookie, then add to ONU
if flow.cookie:
# Adds flow to ONU.
yield self._add_onu_flow(flow)
except Exception as e:
self.log.exception('failed-to-add-to-flow-table', e=e)
self.flow_config_in_progress = False
@inlineCallbacks
def remove_from_flow_table(self, flows):
"""
This function is called for update_flow_incrementally to delete
only delta flows from ONU.
:param flows: flows to delete from ONU
"""
yield self._wait_for_previous_update_flow_to_finish()
self.flow_config_in_progress = True
self.log.debug('remove-from-flow-table', flows=flows)
try:
cookies = [flow.cookie for flow in flows]
for cookie in cookies:
if cookie in self.flow_map:
c_tag = self.flow_map[cookie]["vlan_id"]
self.log.debug("remove-from-flow-table",
cookie=cookie, c_tag=c_tag)
# Deleting flow from ONU.
yield self._delete_onu_flow(self.flow_map[cookie])
# Removing flow_map entry for deleted flow.
del self.flow_map[cookie]
else:
self.log.error('ignoring-cookie_received', cookie=cookie)
except Exception as e:
self.log.exception('failed-to-remove-from-flow-table', e=e)
self.flow_config_in_progress = False
def get_tx_id(self):
self.tx_id += 1
return self.tx_id
def send_omci_message(self, frame):
_frame = hexify(str(frame))
self.log.debug('send-omci-message-%s' % _frame)
device = self.adapter_agent.get_device(self.device_id)
try:
self.adapter_agent.send_proxied_message(device.proxy_address, _frame)
except Exception as e:
self.log.warn('send-omci-message-exception', exc=str(e))
def send_get_circuit_pack(self, entity_id=0):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciGet.message_id,
omci_message=OmciGet(
entity_class=CircuitPack.class_id,
entity_id=entity_id,
attributes_mask=CircuitPack.mask_for('vendor_id')
)
)
self.send_omci_message(frame)
def send_mib_reset(self, entity_id=0):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciMibReset.message_id,
omci_message=OmciMibReset(
entity_class=OntData.class_id,
entity_id=entity_id
)
)
self.send_omci_message(frame)
def send_create_gal_ethernet_profile(self,
entity_id,
max_gem_payload_size):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=GalEthernetProfile.class_id,
entity_id=entity_id,
data=dict(
max_gem_payload_size=max_gem_payload_size
)
)
)
self.send_omci_message(frame)
def send_set_admin_state(self,
entity_id,
admin_state):
data = dict(
administrative_state=admin_state
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=OntG.class_id,
entity_id=entity_id,
attributes_mask=OntG.mask_for(*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_tcont(self,
entity_id,
alloc_id):
data = dict(
alloc_id=alloc_id
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=Tcont.class_id,
entity_id=entity_id,
attributes_mask=Tcont.mask_for(*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_create_8021p_mapper_service_profile(self,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=Ieee8021pMapperServiceProfile.class_id,
entity_id=entity_id,
data=dict(
tp_pointer=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_0=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_1=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_2=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_3=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_4=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_5=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_6=OmciNullPointer,
interwork_tp_pointer_for_p_bit_priority_7=OmciNullPointer
)
)
)
self.send_omci_message(frame)
def send_create_mac_bridge_service_profile(self,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=MacBridgeServiceProfile.class_id,
entity_id=entity_id,
data=dict(
spanning_tree_ind=False,
learning_ind=True,
priority=0x8000,
max_age=20 * 256,
hello_time=2 * 256,
forward_delay=15 * 256,
unknown_mac_address_discard=True
)
)
)
self.send_omci_message(frame)
def send_create_gem_port_network_ctp(self,
entity_id,
port_id,
tcont_id,
direction,
tm):
_directions = {"upstream": 1, "downstream": 2, "bi-directional": 3}
if _directions.has_key(direction):
_direction = _directions[direction]
else:
self.log.error('invalid-gem-port-direction', direction=direction)
raise ValueError('Invalid GEM port direction: {_dir}'.format(_dir=direction))
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=GemPortNetworkCtp.class_id,
entity_id=entity_id,
data=dict(
port_id=port_id,
tcont_pointer=tcont_id,
direction=_direction,
traffic_management_pointer_upstream=tm
)
)
)
self.send_omci_message(frame)
def send_create_multicast_gem_interworking_tp(self,
entity_id,
gem_port_net_ctp_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=MulticastGemInterworkingTp.class_id,
entity_id=entity_id,
data=dict(
gem_port_network_ctp_pointer=gem_port_net_ctp_id,
interworking_option=0,
service_profile_pointer=0x1
)
)
)
self.send_omci_message(frame)
def send_create_gem_inteworking_tp(self,
entity_id,
gem_port_net_ctp_id,
service_profile_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=GemInterworkingTp.class_id,
entity_id=entity_id,
data=dict(
gem_port_network_ctp_pointer=gem_port_net_ctp_id,
interworking_option=5,
service_profile_pointer=service_profile_id,
interworking_tp_pointer=0x0,
gal_profile_pointer=0x1
)
)
)
self.send_omci_message(frame)
def send_delete_omci_mesage(self,
class_id,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciDelete.message_id,
omci_message=OmciDelete(
entity_class=class_id,
entity_id=entity_id
)
)
self.send_omci_message(frame)
def send_set_8021p_mapper_service_profile(self,
entity_id,
interwork_tp_id_0_6=None,
interwork_tp_id_7=None):
data = dict()
if interwork_tp_id_0_6 is not None:
data['interwork_tp_pointer_for_p_bit_priority_0']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_1']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_2']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_3']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_4']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_5']=interwork_tp_id_0_6
data['interwork_tp_pointer_for_p_bit_priority_6']=interwork_tp_id_0_6
if interwork_tp_id_7 is not None:
data['interwork_tp_pointer_for_p_bit_priority_0']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_1']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_2']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_3']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_4']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_5']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_6']=interwork_tp_id_7
data['interwork_tp_pointer_for_p_bit_priority_7']=interwork_tp_id_7
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=Ieee8021pMapperServiceProfile.class_id,
entity_id=entity_id,
attributes_mask=Ieee8021pMapperServiceProfile.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_create_mac_bridge_port_configuration_data(self,
entity_id,
bridge_id,
port_id,
tp_type,
tp_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=MacBridgePortConfigurationData.class_id,
entity_id=entity_id,
data=dict(
bridge_id_pointer=bridge_id,
port_num=port_id,
tp_type=tp_type,
tp_pointer=tp_id
)
)
)
self.send_omci_message(frame)
def send_create_vlan_tagging_filter_data(self,
entity_id,
vlan_id,
fwd_operation):
vlan_filter_list = [0] * 12
vlan_filter_list[0] = vlan_id
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=VlanTaggingFilterData.class_id,
entity_id=entity_id,
data=dict(
vlan_filter_list=vlan_filter_list,
forward_operation=fwd_operation,
number_of_entries=1
)
)
)
self.send_omci_message(frame)
def send_set_vlan_tagging_filter_data(self,
entity_id,
vlan_id):
vlan_filter_list = [0] * 12
vlan_filter_list[0] = vlan_id
data = dict(
vlan_filter_list=vlan_filter_list,
forward_operation=0x10,
number_of_entries=1
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=VlanTaggingFilterData.class_id,
entity_id=entity_id,
attributes_mask=VlanTaggingFilterData.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_delete_vlan_tagging_filter_data(self,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciDelete.message_id,
omci_message=OmciDelete(
entity_class=VlanTaggingFilterData.class_id,
entity_id=entity_id
)
)
self.send_omci_message(frame)
def send_create_extended_vlan_tagging_operation_configuration_data(self,
entity_id,
assoc_type,
assoc_me):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=
ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id,
data=dict(
association_type=assoc_type,
associated_me_pointer=assoc_me
)
)
)
self.send_omci_message(frame)
def send_set_extended_vlan_tagging_operation_tpid_configuration_data(self,
entity_id,
input_tpid,
output_tpid):
data = dict(
input_tpid=input_tpid,
output_tpid=output_tpid,
downstream_mode=0, # inverse of upstream
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=
ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id,
attributes_mask=
ExtendedVlanTaggingOperationConfigurationData.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_extended_vlan_tagging_operation_vlan_configuration_data_untagged(self,
entity_id,
filter_inner_vid,
treatment_inner_vid):
data = dict(
received_frame_vlan_tagging_operation_table=
VlanTaggingOperation(
filter_outer_priority=15,
filter_outer_vid=4096,
filter_outer_tpid_de=0,
filter_inner_priority=15,
filter_inner_vid=filter_inner_vid,
filter_inner_tpid_de=0,
filter_ether_type=0,
treatment_tags_to_remove=0,
treatment_outer_priority=15,
treatment_outer_vid=0,
treatment_outer_tpid_de=0,
treatment_inner_priority=0,
treatment_inner_vid=treatment_inner_vid,
treatment_inner_tpid_de=4
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=
ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id,
attributes_mask=
ExtendedVlanTaggingOperationConfigurationData.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_delete_extended_vlan_tagging_operation_vlan_configuration_data_untagged(self,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciDelete.message_id,
omci_message=OmciDelete(
entity_class=ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id
)
)
self.send_omci_message(frame)
def send_set_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(self,
entity_id,
filter_inner_priority,
filter_inner_vid,
filter_inner_tpid_de,
treatment_tags_to_remove,
treatment_inner_priority,
treatment_inner_vid):
data = dict(
received_frame_vlan_tagging_operation_table=
VlanTaggingOperation(
filter_outer_priority=15,
filter_outer_vid=4096,
filter_outer_tpid_de=0,
filter_inner_priority=filter_inner_priority,
filter_inner_vid=filter_inner_vid,
filter_inner_tpid_de=filter_inner_tpid_de,
filter_ether_type=0,
treatment_tags_to_remove=treatment_tags_to_remove,
treatment_outer_priority=15,
treatment_outer_vid=0,
treatment_outer_tpid_de=0,
treatment_inner_priority=treatment_inner_priority,
treatment_inner_vid=treatment_inner_vid,
treatment_inner_tpid_de=4
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=
ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id,
attributes_mask=
ExtendedVlanTaggingOperationConfigurationData.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_extended_vlan_tagging_operation_vlan_configuration_data_double_tag(self,
entity_id,
filter_outer_vid,
filter_inner_vid,
treatment_tags_to_remove
):
# Note: The default values below are meaningful if transparent handling is
# needed for double vlan tagged packets with matching inner and outer vlans.
# If any other handling is needed for double vlan tagged packets, the default
# values may not work and the function needs to be adapted accordingly.
data = dict(
received_frame_vlan_tagging_operation_table=
VlanTaggingOperation(
filter_outer_priority=14,
filter_outer_vid=filter_outer_vid,
filter_outer_tpid_de=0,
filter_inner_priority=14,
filter_inner_vid=filter_inner_vid,
filter_inner_tpid_de=0,
filter_ether_type=0,
treatment_tags_to_remove=treatment_tags_to_remove,
treatment_outer_priority=15,
treatment_outer_vid=0, # N/A
treatment_outer_tpid_de=0, # N/A
treatment_inner_priority=15,
treatment_inner_vid=0, # N/A
treatment_inner_tpid_de=0 # N/A
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=
ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id,
attributes_mask=
ExtendedVlanTaggingOperationConfigurationData.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_delete_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(self,
entity_id):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciDelete.message_id,
omci_message=OmciDelete(
entity_class=ExtendedVlanTaggingOperationConfigurationData.class_id,
entity_id=entity_id
)
)
self.send_omci_message(frame)
def send_create_multicast_operations_profile(self,
entity_id,
igmp_ver):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=
MulticastOperationsProfile.class_id,
entity_id=entity_id,
data=dict(
igmp_version=igmp_ver,
igmp_function=0,
immediate_leave=0
)
)
)
self.send_omci_message(frame)
def send_set_multicast_operations_profile_acl_row0(self,
entity_id,
acl_table,
row_key,
gem_port,
vlan,
src_ip,
dst_ip_start,
dst_ip_end):
row0 = AccessControlRow0(
set_ctrl=1,
row_part_id=0,
test=0,
row_key=row_key,
gem_port_id=gem_port,
vlan_id=vlan,
src_ip=src_ip,
dst_ip_start=dst_ip_start,
dst_ip_end=dst_ip_end,
ipm_group_bw=0
)
if acl_table == 'dynamic':
data = dict(
dynamic_access_control_list_table=row0
)
else:
data = dict(
static_access_control_list_table=row0
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastOperationsProfile.class_id,
entity_id=entity_id,
attributes_mask=MulticastOperationsProfile.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_multicast_operations_profile_ds_igmp_mcast_tci(self,
entity_id,
ctrl_type,
tci):
data = dict(
ds_igmp_mcast_tci=
DownstreamIgmpMulticastTci(
ctrl_type=ctrl_type,
tci=tci
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastOperationsProfile.class_id,
entity_id=entity_id,
attributes_mask=MulticastOperationsProfile.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_create_multicast_subscriber_config_info(self,
entity_id,
me_type,
mcast_oper_profile):
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciCreate.message_id,
omci_message=OmciCreate(
entity_class=
MulticastSubscriberConfigInfo.class_id,
entity_id=entity_id,
data=dict(
me_type=me_type,
mcast_operations_profile_pointer=mcast_oper_profile
)
)
)
self.send_omci_message(frame)
def send_set_multicast_subscriber_config_info(self,
entity_id,
max_groups=0,
max_mcast_bw=0,
bw_enforcement=0):
data = dict(
max_simultaneous_groups=max_groups,
max_multicast_bandwidth=max_mcast_bw,
bandwidth_enforcement=bw_enforcement
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastSubscriberConfigInfo.class_id,
entity_id=entity_id,
attributes_mask=MulticastSubscriberConfigInfo.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_multicast_service_package(self,
entity_id,
row_key,
vid_uni,
max_groups,
max_mcast_bw,
mcast_oper_profile):
data = dict(
multicast_service_package_table=
MulticastServicePackage(
set_ctrl=1,
row_key=row_key,
vid_uni=vid_uni,
max_simultaneous_groups=max_groups,
max_multicast_bw=max_mcast_bw,
mcast_operations_profile_pointer=mcast_oper_profile
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastSubscriberConfigInfo.class_id,
entity_id=entity_id,
attributes_mask=MulticastSubscriberConfigInfo.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_multicast_allowed_preview_groups_row0(self,
entity_id,
row_key,
src_ip,
vlan_id_ani,
vlan_id_uni):
data = dict(
allowed_preview_groups_table=
AllowedPreviewGroupsRow0(
set_ctrl=1,
row_part_id=0,
row_key=row_key,
src_ip=src_ip,
vlan_id_ani=vlan_id_ani,
vlan_id_uni=vlan_id_uni
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastSubscriberConfigInfo.class_id,
entity_id=entity_id,
attributes_mask=MulticastSubscriberConfigInfo.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_set_multicast_allowed_preview_groups_row1(self,
entity_id,
row_key,
dst_ip,
duration,
time_left):
data = dict(
allowed_preview_groups_table=
AllowedPreviewGroupsRow1(
set_ctrl=1,
row_part_id=1,
row_key=row_key,
dst_ip=dst_ip,
duration=duration,
time_left=time_left
)
)
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciSet.message_id,
omci_message=OmciSet(
entity_class=MulticastSubscriberConfigInfo.class_id,
entity_id=entity_id,
attributes_mask=MulticastSubscriberConfigInfo.mask_for(
*data.keys()),
data=data
)
)
self.send_omci_message(frame)
def send_reboot(self):
self.log.info('send omci reboot message')
frame = OmciFrame(
transaction_id=self.get_tx_id(),
message_type=OmciReboot.message_id,
omci_message=OmciReboot(
entity_class=OntG.class_id,
entity_id=0
)
)
self.send_omci_message(frame)
@inlineCallbacks
def wait_for_response(self):
self.log.debug('wait-for-response')
try:
response = yield self.incoming_messages.get()
self.log.debug('got-response')
resp = OmciFrame(response)
resp.show()
returnValue(resp)
except Exception as e:
returnValue(None)
self.log.info('wait-for-response-exception', exc=str(e))
@inlineCallbacks
def message_exchange(self):
# reset incoming message queue
while self.incoming_messages.pending:
_ = yield self.incoming_messages.get()
mcvid = MANAGEMENT_VLAN
cvid = BRDCM_DEFAULT_VLAN
# construct message
# MIB Reset - OntData - 0
self.send_mib_reset()
yield self.wait_for_response()
# Create AR - GalEthernetProfile - 1
self.send_create_gal_ethernet_profile(1, 48)
yield self.wait_for_response()
# MAC Bridge Service config
# Create AR - MacBridgeServiceProfile - 513
self.send_create_mac_bridge_service_profile(0x201)
yield self.wait_for_response()
# Mapper Service config
# Create AR - 802.1pMapperServiceProfile - 32769
self.send_create_8021p_mapper_service_profile(0x8001)
yield self.wait_for_response()
# MAC Bridge Port config
# Create AR - MacBridgePortConfigData - 8450 - 513 - 3 - 3 - 32769
self.send_create_mac_bridge_port_configuration_data(0x2102, 0x201, 3, 3, 0x8001)
yield self.wait_for_response()
# VLAN Tagging Filter config
# Create AR - VlanTaggingFilterData - 8450 - c-vid
# As per G.988 - Table 9.3.11-1 - Forward operation attribute values
self.send_create_vlan_tagging_filter_data(0x2102, cvid, 0x10)
yield self.wait_for_response()
for port_id in self.uni_ports:
# Extended VLAN Tagging Operation config
# Create AR - ExtendedVlanTaggingOperationConfigData - 514 - 2 - 0x102(Uni-Port-Num)
self.send_create_extended_vlan_tagging_operation_configuration_data(0x200 + port_id, 2, 0x100 + port_id)
yield self.wait_for_response()
# Set AR - ExtendedVlanTaggingOperationConfigData - 514 - 8100 - 8100
self.send_set_extended_vlan_tagging_operation_tpid_configuration_data(0x200 + port_id, 0x8100, 0x8100)
yield self.wait_for_response()
# Create AR - MacBridgePortConfigData - Entity_id -
# bridge ID -
# port num -
# tp_type -
# IEEE MApper poniter
self.send_create_mac_bridge_port_configuration_data(0x200 + port_id, 0x201, port_id, 1, 0x100 + port_id)
yield self.wait_for_response()
# Set AR - ExtendedVlanTaggingOperationConfigData
# 514 - RxVlanTaggingOperationTable - add VLAN <cvid> to priority tagged pkts - c-vid
# To be removed once DPU supports sending vlan4090p3 instead of vlan4090p0
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(0x200 + port_id, 0, mcvid, 0, 1, 7, mcvid)
yield self.wait_for_response()
# Set AR - ExtendedVlanTaggingOperationConfigData
# 514 - RxVlanTaggingOperationTable - add VLAN <cvid> to priority tagged pkts - c-vid
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(0x200 + port_id, 3, mcvid, 0, 1, 7, mcvid)
yield self.wait_for_response()
# Set AR - ExtendedVlanTaggingOperationConfigData
# 514 - RxVlanTaggingOperationTable - add VLAN <cvid> to untagged pkts - c-vid
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_untagged(0x200 + port_id, 0x1000, cvid)
yield self.wait_for_response()
# Multicast related MEs
# Set AR - MulticastOperationsProfile - Dynamic Access Control List table
# Create AR - MacBridgePortConfigData - 9000 - 513 - 6 - 6 - 6
self.send_create_mac_bridge_port_configuration_data(0x2328, 0x201, 6, 6, 6)
yield self.wait_for_response()
# Multicast Operation Profile config
# Create AR - MulticastOperationsProfile
self.send_create_multicast_operations_profile(0x201, 3)
yield self.wait_for_response()
# Multicast Subscriber config
# Create AR - MulticastSubscriberConfigInfo
self.send_create_multicast_subscriber_config_info(0x201, 0, 0x201)
yield self.wait_for_response()
# Create AR - GemPortNetworkCtp - 260 - 4000 - 0 Multicast
self.send_create_gem_port_network_ctp(0x104, 0x0FA0, 0, "downstream", 0)
yield self.wait_for_response()
# Multicast GEM Interworking config Multicast
# Create AR - MulticastGemInterworkingTp - 6 - 260
self.send_create_multicast_gem_interworking_tp(0x6, 0x104)
yield self.wait_for_response()
self.send_set_multicast_operations_profile_acl_row0(0x201,
'dynamic',
0,
0x0fa0,
0x0fa0,
'0.0.0.0',
'224.0.0.0',
'239.255.255.255')
yield self.wait_for_response()
# Multicast Operation Profile config
# Set AR - MulticastOperationsProfile - Downstream IGMP Multicast TCI
self.send_set_multicast_operations_profile_ds_igmp_mcast_tci(0x201, 4, cvid)
yield self.wait_for_response()
def add_uni_port(self, device, parent_logical_device_id,
name, parent_port_num=None):
self.log.info('adding-logical-port', device_id=device.id,
logical_device_id=parent_logical_device_id,
name=name)
if parent_port_num is not None:
uni = parent_port_num
port_no = parent_port_num
else:
uni = self.uni_ports[0]
port_no = device.proxy_address.channel_id + uni
# register physical ports
uni_port = Port(
port_no=uni,
label='uni-'+str(uni),
type=Port.ETHERNET_UNI,
admin_state=AdminState.ENABLED,
oper_status=OperStatus.ACTIVE
)
self.adapter_agent.add_port(device.id, uni_port)
# add uni port to logical device
cap = OFPPF_1GB_FD | OFPPF_FIBER
self.adapter_agent.add_logical_port(parent_logical_device_id,
LogicalPort(
id='uni-{}'.format(port_no),
ofp_port=ofp_port(
port_no=port_no,
hw_addr=mac_str_to_tuple('00:00:00:%02x:%02x:%02x' %
(device.proxy_address.onu_id & 0xff,
(port_no >> 8) & 0xff,
port_no & 0xff)),
#name='uni-{}'.format(port_no),
name=name,
config=0,
state=OFPPS_LIVE,
curr=cap,
advertised=cap,
peer=cap,
curr_speed=OFPPF_1GB_FD,
max_speed=OFPPF_1GB_FD
),
device_id=device.id,
device_port_no=uni_port.port_no
))
def del_uni_port(self, device, parent_logical_device_id,
name, parent_port_num=None):
self.log.debug('del-uni-port', device_id=device.id,
logical_device_id=parent_logical_device_id,
name=name)
if parent_port_num is not None:
uni = parent_port_num
port_no = parent_port_num
else:
uni = self.uni_ports[0]
port_no = device.proxy_address.channel_id + uni
# register physical ports
ports = self.adapter_agent.get_ports(self.device_id, Port.ETHERNET_UNI)
for port in ports:
if port.label == 'uni-'+str(uni):
break
self.adapter_agent.delete_port(self.device_id, port)
self.adapter_agent.delete_logical_port_by_id(parent_logical_device_id,
'uni-{}'.format(port_no))
def delete_v_ont_ani(self, data):
self.log.info('deleting-v_ont_ani')
device = self.adapter_agent.get_device(self.device_id)
self.proxy_address = device.proxy_address
self.adapter_agent.unregister_for_proxied_messages(device.proxy_address)
ports = self.adapter_agent.get_ports(self.device_id, Port.PON_ONU)
if ports is not None:
for port in ports:
if port.label == 'PON port':
self.adapter_agent.delete_port(self.device_id, port)
break
# construct message
# MIB Reset - OntData - 0
if device.connect_status != ConnectStatus.REACHABLE:
self.log.error('device-unreachable')
return
self.send_mib_reset()
# It is observed that the device is already deleted before the response
# is received. So, there is no point waiting for response.
# Also, currently there is no response validation or timeout.
# Until we move to the OpenOMCI framework, it is ok to ignore this
# response for now.
# yield self.wait_for_response()
def create_interface(self, data):
if isinstance(data, VEnetConfig):
parent_port_num = None
onu_device = self.adapter_agent.get_device(self.device_id)
ports = self.adapter_agent.get_ports(onu_device.parent_id, Port.ETHERNET_UNI)
parent_port_num = None
for port in ports:
if port.label == data.interface.name:
parent_port_num = port.port_no
break
parent_device = self.adapter_agent.get_device(onu_device.parent_id)
logical_device_id = parent_device.parent_id
assert logical_device_id
self.add_uni_port(onu_device, logical_device_id,
data.name, parent_port_num)
if parent_port_num is None:
self.log.error("matching-parent-uni-port-num-not-found")
return
onu_ports = self.adapter_agent.get_ports(self.device_id, Port.PON_ONU)
if onu_ports:
# To-Do :
# Assumed only one PON port and UNI port per ONU.
pon_port = onu_ports[0]
else:
self.log.error("No-Pon-port-configured-yet")
return
self.adapter_agent.delete_port_reference_from_parent(self.device_id,
pon_port)
pon_port.peers[0].device_id = onu_device.parent_id
pon_port.peers[0].port_no = parent_port_num
pon_port.admin_state = AdminState.ENABLED
pon_port.oper_status = OperStatus.ACTIVE
self.adapter_agent.add_port_reference_to_parent(self.device_id,
pon_port)
else:
self.log.info('Not-handled-Yet')
return
def update_interface(self, data):
self.log.info('Not-Implemented-yet')
return
def remove_interface(self, data):
if isinstance(data, VEnetConfig):
parent_port_num = None
onu_device = self.adapter_agent.get_device(self.device_id)
ports = self.adapter_agent.get_ports(onu_device.parent_id, Port.ETHERNET_UNI)
parent_port_num = None
for port in ports:
if port.label == data.interface.name:
parent_port_num = port.port_no
break
parent_device = self.adapter_agent.get_device(onu_device.parent_id)
logical_device_id = parent_device.parent_id
assert logical_device_id
self.del_uni_port(onu_device, logical_device_id,
data.name, parent_port_num)
if isinstance(data, VOntaniConfig):
self.delete_v_ont_ani(data)
else:
self.log.info('not-handled-yet')
return
@inlineCallbacks
def create_gemport(self, data):
self.log.debug('create-gemport')
gem_port = GemportsConfigData()
gem_port.CopyFrom(data)
if gem_port.tcont_ref is None:
self.log.error('recevied-null-gem-port-data')
else:
#To-Do Need to see how the valuse 0x8001 is derived
self.send_create_gem_port_network_ctp(gem_port.gemport_id,
gem_port.gemport_id, 0x8001,
"bi-directional", 0x100)
yield self.wait_for_response()
# GEM Interworking config
# Create AR - GemInterworkingTp - Gem_port,TP_pointer -
# Gem port CTP pointer -
# Mapper service profile id
self.send_create_gem_inteworking_tp(gem_port.gemport_id,
gem_port.gemport_id, 0x8001)
yield self.wait_for_response()
# Mapper Service Profile config
# Set AR - 802.1pMapperServiceProfile - Mapper_ profile_id -
# gem_port_tp pointer
if gem_port.traffic_class == 2:
self.send_set_8021p_mapper_service_profile(0x8001,
interwork_tp_id_7=gem_port.gemport_id)
else:
self.send_set_8021p_mapper_service_profile(0x8001,
interwork_tp_id_0_6=gem_port.gemport_id)
yield self.wait_for_response()
@inlineCallbacks
def remove_gemport(self, data):
self.log.debug('remove-gemport')
gem_port = GemportsConfigData()
gem_port.CopyFrom(data)
device = self.adapter_agent.get_device(self.device_id)
if device.connect_status != ConnectStatus.REACHABLE:
self.log.error('device-unreachable')
return
self.send_set_8021p_mapper_service_profile(0x8001,
interwork_tp_id_0_6=0xFFFF,
interwork_tp_id_7=0xFFFF)
yield self.wait_for_response()
self.send_delete_omci_mesage(GemInterworkingTp.class_id,
gem_port.gemport_id)
yield self.wait_for_response()
#To-Do Need to see how the valuse 0x8001 is derived
self.send_delete_omci_mesage(GemPortNetworkCtp.class_id,
gem_port.gemport_id)
yield self.wait_for_response()
@inlineCallbacks
def create_tcont(self, tcont_data, traffic_descriptor_data):
self.log.debug('create-tcont')
tcont = TcontsConfigData()
tcont.CopyFrom(tcont_data)
if tcont.interface_reference is not None:
self.log.debug('tcont', tcont=tcont.alloc_id)
self.send_set_tcont(0x8001, tcont.alloc_id)
yield self.wait_for_response()
else:
self.log.info('recevied-null-tcont-data', tcont=tcont.alloc_id)
@inlineCallbacks
def remove_tcont(self, tcont_data, traffic_descriptor_data):
self.log.debug('remove-tcont')
device = self.adapter_agent.get_device(self.device_id)
if device.connect_status != ConnectStatus.REACHABLE:
self.log.error('device-unreachable')
return
self.send_set_tcont(0x8001, 0xFFFF)
yield self.wait_for_response()
def create_multicast_gemport(self, data):
self.log.info('Send relevant OMCI message - Not implemented yet')
@inlineCallbacks
def disable(self, device):
try:
self.log.info('sending-admin-state-lock-towards-device', device=device)
self.send_set_admin_state(0x0000, ADMIN_STATE_LOCK)
yield self.wait_for_response()
# Disable all ports on that device
self.disable_ports(device)
device.oper_status = OperStatus.UNKNOWN
device.connect_status = ConnectStatus.UNREACHABLE
self.adapter_agent.update_device(device)
except Exception as e:
log.exception('exception-in-onu-disable', exception=e)
@inlineCallbacks
def reenable(self, device):
try:
self.log.info('sending-admin-state-unlock-towards-device', device=device)
self.send_set_admin_state(0x0000, ADMIN_STATE_UNLOCK)
yield self.wait_for_response()
self.enable_ports(device)
device.oper_status = OperStatus.ACTIVE
device.connect_status = ConnectStatus.REACHABLE
self.adapter_agent.update_device(device)
except Exception as e:
log.exception('exception-in-onu-reenable', exception=e)
@inlineCallbacks
def reboot(self):
self.log.info('reboot-device')
device = self.adapter_agent.get_device(self.device_id)
if device.connect_status != ConnectStatus.REACHABLE:
self.log.error("device-unreacable")
return
try:
self.send_reboot()
response = yield self.wait_for_response()
if response is not None:
omci_response = response.getfieldval("omci_message")
success_code = omci_response.getfieldval("success_code")
if success_code == 0:
self.log.debug("reboot-command-processed-successfully")
# Update the device connection and operation status
device = self.adapter_agent.get_device(self.device_id)
device.connect_status = ConnectStatus.UNREACHABLE
device.oper_status = OperStatus.DISCOVERED
self.adapter_agent.update_device(device)
self.disable_ports(device)
self.flow_map.clear()
else:
self.log.error("reboot-failed", success_code=success_code)
else:
self.log.error("error-in-processing-reboot-response")
except Exception as e:
self.log.error('wait-for-response-exception', exc=str(e))
def disable_ports(self, device):
self.log.info('disable-ports', device_id=self.device_id)
# Disable all ports on that device
self.adapter_agent.disable_all_ports(self.device_id)
parent_device = self.adapter_agent.get_device(device.parent_id)
assert parent_device
logical_device_id = parent_device.parent_id
assert logical_device_id
ports = self.adapter_agent.get_ports(device.id, Port.ETHERNET_UNI)
for port in ports:
port_id = 'uni-{}'.format(port.port_no)
try:
lgcl_port = self.adapter_agent.get_logical_port(logical_device_id, port_id)
lgcl_port.ofp_port.state = OFPPS_LINK_DOWN
self.adapter_agent.update_logical_port(logical_device_id, lgcl_port)
except KeyError:
self.log.info('logical-port-not-found', device_id=self.device_id,
portid=port_id)
def enable_ports(self, device):
self.log.info('enable-ports', device_id=self.device_id)
# Enable all ports on that device
self.adapter_agent.enable_all_ports(self.device_id)
parent_device = self.adapter_agent.get_device(device.parent_id)
assert parent_device
logical_device_id = parent_device.parent_id
assert logical_device_id
ports = self.adapter_agent.get_ports(device.id, Port.ETHERNET_UNI)
for port in ports:
port_id = 'uni-{}'.format(port.port_no)
try:
lgcl_port = self.adapter_agent.get_logical_port(logical_device_id, port_id)
lgcl_port.ofp_port.state = OFPPS_LIVE
self.adapter_agent.update_logical_port(logical_device_id, lgcl_port)
except KeyError:
self.log.info('logical-port-not-found', device_id=self.device_id,
portid=port_id)
@inlineCallbacks
def _add_onu_flow(self, flow):
_type = None
_port = None
_vlan_vid = None
_udp_dst = None
_udp_src = None
_ipv4_dst = None
_ipv4_src = None
_metadata = None
_output = None
_push_tpid = None
_field = None
_set_vlan_vid = None
self.log.info('add-flow', flow=flow)
def is_downstream(port):
return port == 100 # Need a better way
def is_upstream(port):
return not is_downstream(port)
try:
_in_port = fd.get_in_port(flow)
assert _in_port is not None
if is_downstream(_in_port):
self.log.info('downstream-flow')
elif is_upstream(_in_port):
self.log.info('upstream-flow')
else:
raise Exception('port should be 1 or 2 by our convention')
_out_port = fd.get_out_port(flow) # may be None
self.log.info('out-port', out_port=_out_port)
for field in fd.get_ofb_fields(flow):
if field.type == fd.ETH_TYPE:
_type = field.eth_type
self.log.info('field-type-eth-type',
eth_type=_type)
elif field.type == fd.IP_PROTO:
_proto = field.ip_proto
self.log.info('field-type-ip-proto',
ip_proto=_proto)
elif field.type == fd.IN_PORT:
_port = field.port
self.log.info('field-type-in-port',
in_port=_port)
elif field.type == fd.VLAN_VID:
_vlan_vid = field.vlan_vid & 0xfff
self.log.info('field-type-vlan-vid',
vlan=_vlan_vid)
elif field.type == fd.VLAN_PCP:
_vlan_pcp = field.vlan_pcp
self.log.info('field-type-vlan-pcp',
pcp=_vlan_pcp)
elif field.type == fd.UDP_DST:
_udp_dst = field.udp_dst
self.log.info('field-type-udp-dst',
udp_dst=_udp_dst)
elif field.type == fd.UDP_SRC:
_udp_src = field.udp_src
self.log.info('field-type-udp-src',
udp_src=_udp_src)
elif field.type == fd.IPV4_DST:
_ipv4_dst = field.ipv4_dst
self.log.info('field-type-ipv4-dst',
ipv4_dst=_ipv4_dst)
elif field.type == fd.IPV4_SRC:
_ipv4_src = field.ipv4_src
self.log.info('field-type-ipv4-src',
ipv4_dst=_ipv4_src)
elif field.type == fd.METADATA:
_metadata = field.table_metadata
self.log.info('field-type-metadata',
metadata=_metadata)
else:
raise NotImplementedError('field.type={}'.format(
field.type))
for action in fd.get_actions(flow):
if action.type == fd.OUTPUT:
_output = action.output.port
self.log.info('action-type-output',
output=_output, in_port=_in_port)
elif action.type == fd.POP_VLAN:
self.log.info('action-type-pop-vlan',
in_port=_in_port)
elif action.type == fd.PUSH_VLAN:
_push_tpid = action.push.ethertype
self.log.info('action-type-push-vlan',
push_tpid=_push_tpid, in_port=_in_port)
if action.push.ethertype != 0x8100:
self.log.error('unhandled-tpid',
ethertype=action.push.ethertype)
elif action.type == fd.SET_FIELD:
_field = action.set_field.field.ofb_field
assert (action.set_field.field.oxm_class ==
OFPXMC_OPENFLOW_BASIC)
self.log.info('action-type-set-field',
field=_field, in_port=_in_port)
if _field.type == fd.VLAN_VID:
_set_vlan_vid = _field.vlan_vid & 0xfff
self.log.info('set-field-type-vlan-vid',vlan=_set_vlan_vid)
else:
self.log.error('unsupported-action-set-field-type',
field_type=_field.type)
else:
self.log.error('unsupported-action-type',
action_type=action.type, in_port=_in_port)
#
# All flows created from ONU adapter should be OMCI based
#
if _vlan_vid == 0 and _set_vlan_vid != None and _set_vlan_vid != 0:
# allow priority tagged packets
# Set AR - ExtendedVlanTaggingOperationConfigData
# 514 - RxVlanTaggingOperationTable - add VLAN <cvid> to priority tagged pkts - c-vid
if _set_vlan_vid != RESERVED_VLAN_ID:
# As per G.988 - Table 9.3.11-1 - Forward operation attribute values
# Forward action of 0x10 allows VID Investigation
if _type != FLOW_TYPE_EAPOL:
self.log.info("Triggering-extended-vlan-configurations",vlan=_set_vlan_vid,eth_type=_type)
self.send_set_vlan_tagging_filter_data(0x2102, _set_vlan_vid)
#self.send_create_vlan_tagging_filter_data(0x2102, _set_vlan_vid, 0x10)
yield self.wait_for_response()
for port_id in self.uni_ports:
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_untagged(0x200 + port_id, 0x1000, _set_vlan_vid)
yield self.wait_for_response()
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(0x200 + port_id, 8, 0, 0, 1, 8, _set_vlan_vid)
yield self.wait_for_response()
else:
self.log.info("Use-vlan-4091-for-eapol",eth_type=_type)
else:
# As per G.988 - Table 9.3.11-1 - Forward operation attribute values
# Forward action of 0x00 does not perform VID Investigation for transparent vlan case
self.send_delete_vlan_tagging_filter_data(0x2102)
yield self.wait_for_response()
self.send_create_vlan_tagging_filter_data(0x2102, _set_vlan_vid, 0x00)
yield self.wait_for_response()
for port_id in self.uni_ports:
self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(0x200 + port_id, 14, 4096, 0, 0, 15, 0)
yield self.wait_for_response()
#self.send_set_extended_vlan_tagging_operation_vlan_configuration_data_double_tag( \
# 0x200 + port_id, 4096, 4096, 0)
#yield self.wait_for_response()
# Create the entry in the internal flow table
self.flow_map[flow.cookie] = {"vlan_id": _set_vlan_vid,
"eth_type":_type,
"vlan_tagging_filter_data_entity_id": 0x2102,
"extended_vlan_tagging_filter_data": [0x200 + port_id for port_id in self.uni_ports]}
except Exception as e:
self.log.exception('failed-to-install-flow', e=e, flow=flow)
@inlineCallbacks
def _delete_onu_flow(self, flow_map_value):
# Deletes ONU flows.
try:
if flow_map_value["vlan_id"] != RESERVED_VLAN_ID:
for extended_vlan_tagging_filter_data in flow_map_value["extended_vlan_tagging_filter_data"]:
if flow_map_value["eth_type"] != FLOW_TYPE_EAPOL:
self.send_delete_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(
extended_vlan_tagging_filter_data)
yield self.wait_for_response()
self.send_delete_extended_vlan_tagging_operation_vlan_configuration_data_untagged(
extended_vlan_tagging_filter_data)
yield self.wait_for_response()
else:
for extended_vlan_tagging_filter_data in flow_map_value["extended_vlan_tagging_filter_data"]:
self.send_delete_extended_vlan_tagging_operation_vlan_configuration_data_single_tag(
extended_vlan_tagging_filter_data)
yield self.wait_for_response()
for port_id in self.uni_ports:
# Extended VLAN Tagging Operation config
# Create AR - ExtendedVlanTaggingOperationConfigData - 514 - 2 - 0x102(Uni-Port-Num)
self.send_create_extended_vlan_tagging_operation_configuration_data(0x200 + port_id, 2, 0x100 + port_id)
yield self.wait_for_response()
# Set AR - ExtendedVlanTaggingOperationConfigData - 514 - 8100 - 8100
self.send_set_extended_vlan_tagging_operation_tpid_configuration_data(0x200 + port_id, 0x8100, 0x8100)
yield self.wait_for_response()
except Exception as e:
self.log.exception('failed-to-delete-flow', e=e, flow_map_value=flow_map_value)
@inlineCallbacks
def _wait_for_previous_update_flow_to_finish(self):
while self.flow_config_in_progress:
# non-blocking wait for 200ms
yield asleep(0.2)
return