tree: a9804acc047b760a014f0e4084bf84c0f7820296 [path history] [tgz]
  1. Dockerfile.cord
  2. Makefile
  3. Makefile.inside
  4. README-VTN.md
  5. README.md
  6. ceilometer.yaml
  7. cord.yaml
  8. dataplane/
  9. docker-compose.yml
  10. make-virtualbng-json.sh
  11. make-vtn-networkconfig-json.sh
  12. xos_cord_config
cord-deprecated/README.md

CORD development environment

This configuration can be used to set up a CORD development environment. It does the following:

  • Sets up a basic dataplane for testing end-to-end packet flow between a subscriber client and the Internet
  • Brings up ONOS apps for controlling the dataplane: virtualbng, olt
  • Configures XOS with the CORD services: vCPE, vBNG, vOLT

NOTE: This configuration is stale and likely not working at present. If you are looking to evaluate and/or contribute to CORD, you should look instead at the cord-pod configuration. Almost all CORD developers have transitioned to cord-pod.

End-to-end dataplane

The configuration uses XOS to set up an end-to-end dataplane for development of the XOS services and ONOS apps used in CORD. It abstracts away most of the complexity of the CORD hardware using virtual networks and Open vSwitch (OvS) switches. At a high level the dataplane looks like this:

             olt                 virtualbng
             ----                  ----
             ONOS                  ONOS
              |                     |
client ----> CPqD ----> vCPE ----> OvS ----> Internet
         1         2          3         4

On the datapath are two OvS switches, controlled by the olt and virtualbng ONOS applications. Once all the pieces are in place, the client at left should be able to obtain an IP address via DHCP from the vCPE and send packets out to the Internet.

All of the components in the above diagram (i.e., client, OvS switches, ONOS, and vCPE) currently run in distinct VMs created by XOS. The numbers in the diagram correspond to networks set up by XOS:

  1. subscriber_network
  2. lan_network
  3. wan_network
  4. public_network

How to run it

The configuration is intended to be run on CloudLab. It launches an XOS container on Cloudlab that runs the XOS develserver. The container is left running in the background.

To get started on CloudLab:

  • Create an experiment using the OpenStack-CORD profile. (You can also use the OpenStack profile, but choose Kilo with two compute nodes and disable security groups.)
  • Wait until you get an email from CloudLab with title "OpenStack Instance Finished Setting Up".
  • Login to the ctl node of your experiment and run:
ctl:~$ git clone https://github.com/open-cloud/xos.git
ctl:~$ cd xos/xos/configurations/cord/
ctl:~/xos/xos/configurations/cord$ make

Running make in this directory creates the XOS Docker container and runs the TOSCA engine with cord.yaml to configure XOS with the CORD services. In addition, a number of VMs are created:

  1. Slice mysite_onos: runs the ONOS Docker container with virtualbng app loaded
  2. Slice mysite_onos: runs the ONOS Docker container with olt app loaded
  3. Slice mysite_vbng: for running OvS with the virtualbng app as controller
  4. Slice mysite_volt: for running the CPqD switch with the olt app as controller
  5. Slice mysite_clients: a subscriber client for end-to-end testing
  6. Slice mysite_vcpe: runs the vCPE Docker container (if not using containers on bare metal)

Once all the VMs are up and the ONOS apps are configured, XOS should be able to get an address mapping from the virtualbng ONOS app for the vCPE. To verify that it has received an IP address mapping, look at the Routeable subnet: field in the appropriate Vbng tenant object in XOS. It should contain an IP address in the 10.254.0.0/24 subnet.

After launching the ONOS apps, it is necessary to configure software switches along the dataplane so that ONOS can control them. To do this, from the cord configuration directory:

ctl:~/xos/xos/configurations/cord$ cd dataplane/
ctl:~/xos/xos/configurations/cord/dataplane$ ./gen-inventory.sh > hosts
ctl:~/xos/xos/configurations/cord/dataplane$ ansible-playbook -i hosts dataplane.yaml

To setup the dataplane for containers on bare metal, perform these steps in addition to the above (note: make sure to sudo when running the playbook):

ctl:~/xos/xos/configurations/cord/dataplane$ ./generate-bm.sh > hosts-bm   
ctl:~/xos/xos/configurations/cord/dataplane$ sudo ansible-playbook -i hosts-bm dataplane-bm.yaml

Check that the vCPE container has started, by going into the XOS UI, selecting 'Services', 'service_vcpe', 'Administration', 'Vcpe Tenants', and make sure there's a green icon next to the vCPE.

If the vCPE Tenant is still red, then the Instance could be exponentially backed-off due to errors while trying to sync before dataplane.yaml was run. You can reset the exponential backoff by tracking down the vCPE Instance (Slices->mysite_vcpe->Instances, and find the Instance associated with the vCPE Tenant) and hitting the save button.

Now SSH into ONOS running the OLT app (see below) and activate the subscriber:

onos> add-subscriber-access of:0000000000000001 1 432

At this point the client should be able to get an IP address from the vCPE via DHCP. To set up the IP address and default route on the client:

client:$ sudo route del default gw 10.11.10.5
client:$ sudo dhclient br-sub

Once dhclient returns, the client should now be able to surf the Internet through the dataplane.

Setting up /etc/hosts

To make it easy to log into the various VMs that make up the dataplane, add entries for them into /etc/hosts on the ctl node. As root, run:

ctl:~/xos/xos/configurations/cord/dataplane$ ./gen-etc-hosts.sh >> /etc/hosts

For example, to log into the client:

ctl:~$ ssh ubuntu@client

How to log into ONOS

ONOS apps are run inside Docker containers hosted in VMs. All ports exposed by the ONOS container are forwarded to the outside, and can be accessed from the ctl node over the flat-lan-1-net network. Assuming that /etc/hosts has been configured as described above, it is possible to SSH to the ONOS running the virtualbng app as follows (password is karaf):

$ ssh -p 8101 karaf@onos_vbng
Password authentication
Password:
Welcome to Open Network Operating System (ONOS)!
     ____  _  ______  ____
    / __ \/ |/ / __ \/ __/
   / /_/ /    / /_/ /\ \
   \____/_/|_/\____/___/


Hit '<tab>' for a list of available commands
and '[cmd] --help' for help on a specific command.
Hit '<ctrl-d>' or type 'system:shutdown' or 'logout' to shutdown ONOS.

onos>

For instance, to check the IP address mappings managed by the virtualbng app:

onos> vbngs
   Private IP - Public IP
   10.0.1.3 - 10.254.0.129

Troubleshooting

Problem: No external connectivity from vCPE container

  1. Make sure the hosts listed in virtualbng.json are the actual compute nodes used in your experiment.
  2. Try rebooting the ONOS container running the virtualbng app: $ ssh ubuntu@onos-vbng "sudo docker restart ONOS"