This configuration can be used to set up a CORD development environment. It does the following:
NOTE: This configuration is under active development and is not yet finished! Some features are not fully working yet.
The configuration uses XOS to set up an end-to-end dataplane for development of the XOS services and ONOS apps used in CORD. It abstracts away most of the complexity of the CORD hardware using virtual networks and Open vSwitch (OvS) switches. At a high level the dataplane looks like this:
olt virtualbng ---- ---- ONOS ONOS | | client ----> CPqD ----> vCPE ----> OvS ----> Internet 1 2 3 4
On the datapath are two OvS switches, controlled by the olt
and virtualbng
ONOS applications. Once all the pieces are in place, the client at left should be able to obtain an IP address via DHCP from the vCPE and send packets out to the Internet.
All of the components in the above diagram (i.e., client, OvS switches, ONOS, and vCPE) currently run in distinct VMs created by XOS. The numbers in the diagram correspond to networks set up by XOS:
The configuration is intended to be run on CloudLab. It launches an XOS container on Cloudlab that runs the XOS develserver. The container is left running in the background.
To get started on CloudLab:
ctl:~$ git clone https://github.com/open-cloud/xos.git ctl:~$ cd xos/xos/configurations/cord/ ctl:~/xos/xos/configurations/cord$ make
Running make
in this directory creates the XOS Docker container and runs the TOSCA engine with cord.yaml
to configure XOS with the CORD services. In addition, a number of VMs are created:
virtualbng
app loadedolt
app loadedvirtualbng
app as controllerolt
app as controllerOnce all the VMs are up and the ONOS apps are configured, XOS should be able to get an address mapping from the virtualbng
ONOS app for the vCPE. To verify that it has received an IP address mapping, look at the Routeable subnet: field in the appropriate Vbng tenant object in XOS. It should contain an IP address in the 10.254.0.0/24 subnet.
After launching the ONOS apps, it is necessary to configure software switches along the dataplane so that ONOS can control them. To do this, from the cord
configuration directory:
ctl:~/xos/xos/configurations/cord$ cd dataplane/ ctl:~/xos/xos/configurations/cord/dataplane$ ./gen-inventory.sh > hosts ctl:~/xos/xos/configurations/cord/dataplane$ ansible-playbook -i hosts dataplane.yaml
To setup the dataplane for containers on bare metal, perform these steps in addition to the above (note: make sure to sudo when running the playbook):
ctl:~/xos/xos/configurations/cord/dataplane$ ./generate-bm.sh > hosts-bm ctl:~/xos/xos/configurations/cord/dataplane$ sudo ansible-playbook -i hosts-bm dataplane-bm.yaml
Check that the vCPE container has started, by going into the XOS UI, selecting 'Services', 'service_vcpe', 'Administration', 'Vcpe Tenants', and make sure there's a green icon next to the vCPE.
If the vCPE Tenant is still red, then the Instance could be exponentially backed-off due to errors while trying to sync before dataplane.yaml was run. You can reset the exponential backoff by tracking down the vCPE Instance (Slices->mysite_vcpe->Instances, and find the Instance associated with the vCPE Tenant) and hitting the save button.
Now SSH into ONOS running the OLT app (see below) and activate the subscriber:
onos> add-subscriber-access of:0000000000000001 1 432
At this point the client should be able to get an IP address from the vCPE via DHCP. To set up the IP address and default route on the client:
client:$ sudo route del default gw 10.11.10.5 client:$ sudo dhclient br-sub
Once dhclient
returns, the client should now be able to surf the Internet through the dataplane.
To make it easy to log into the various VMs that make up the dataplane, add entries for them into /etc/hosts
on the ctl node. As root, run:
ctl:~/xos/xos/configurations/cord/dataplane$ ./gen-etc-hosts.sh >> /etc/hosts
For example, to log into the client:
ctl:~$ ssh ubuntu@client
ONOS apps are run inside Docker containers hosted in VMs. All ports exposed by the ONOS container are forwarded to the outside, and can be accessed from the ctl node over the flat-lan-1-net
network. Assuming that /etc/hosts
has been configured as described above, it is possible to SSH to the ONOS running the virtualbng
app as follows (password is karaf):
$ ssh -p 8101 karaf@onos_vbng Password authentication Password: Welcome to Open Network Operating System (ONOS)! ____ _ ______ ____ / __ \/ |/ / __ \/ __/ / /_/ / / /_/ /\ \ \____/_/|_/\____/___/ Hit '<tab>' for a list of available commands and '[cmd] --help' for help on a specific command. Hit '<ctrl-d>' or type 'system:shutdown' or 'logout' to shutdown ONOS. onos>
For instance, to check the IP address mappings managed by the virtualbng
app:
onos> vbngs Private IP - Public IP 10.0.1.3 - 10.254.0.129
virtualbng.json
are the actual compute nodes used in your experiment.virtualbng
app: $ ssh ubuntu@onos-vbng "sudo docker restart ONOS"