Installation

The networking-odl repository includes integration with DevStack that enables creation of a simple OpenDaylight (ODL) development and test environment. This document discusses what is required for manual installation and integration into a production OpenStack deployment tool of conventional architectures that include the following types of nodes:

  • Controller - Runs OpenStack control plane services such as REST APIs and databases.
  • Network - Provides connectivity between provider (public) and project (private) networks. Services provided include layer-3 (routing), DHCP, and metadata agents. Layer-3 agent is optional. When using netvirt (vpnservice) DHCP/metadata are optional.
  • Compute - Runs the hypervisor and layer-2 agent for the Networking service.

ODL Installation

http://docs.opendaylight.org provides manual and general documentation for ODL

Review the following documentation regardless of install scenario:

Choose and review one of the following installation scenarios:

  • GBP with OpenStack. OpenDaylight Group Based Policy allows users to express network configuration in a declarative rather than imperative way. Often described as asking for “what you want”, rather than “how you can do it”, Group Based Policy achieves this by implementing an Intent System. The Intent System is a process around an intent driven data model and contains no domain specifics but is capable of addressing multiple semantic definitions of intent.
  • OVSDB with OpenStack. OpenDaylight OVSDB allows users to take advantage of Network Virtualization using OpenDaylight SDN capabilities whilst utilizing OpenvSwitch. The stack includes a Neutron Northbound, a Network Virtualization layer, an OVSDB southbound plugin, and an OpenFlow southbound plugin.
  • VTN with OpenStack. OpenDaylight Virtual Tenant Network (VTN) is an application that provides multi-tenant virtual network on an SDN controller. VTN Manager is implemented as one plugin to the OpenDaylight controller and provides a REST interface to create/update/delete VTN components. It provides an implementation of Openstack L2 Network Functions API.

Networking-odl Installation

# sudo pip install networking-odl

Note

pip need to be installed before running above command.

Networking-odl Configuration

All related neutron services need to be restarted after configuration change.

  1. Configure Openstack neutron server. The neutron server implements ODL as an ML2 driver. Edit the /etc/neutron/neutron.conf file:

    • Enable the ML2 core plug-in.

      [DEFAULT]
      ...
      core_plugin = neutron.plugins.ml2.plugin.Ml2Plugin
      
    • (Optional) Enable ODL L3 router, if QoS feature is desired, then qos should be appended to service_plugins

      [DEFAULT]
      ...
      service_plugins = odl-router
      
  2. Configure the ML2 plug-in. Edit the /etc/neutron/plugins/ml2/ml2_conf.ini file:

    • Configure the ODL mechanism driver, network type drivers, self-service (tenant) network types, and enable extension drivers(optional).

      [ml2]
      ...
      mechanism_drivers = opendaylight_v2
      type_drivers = local,flat,vlan,vxlan
      tenant_network_types = vxlan
      extension_drivers = port_security, qos
      

    ..note:

    The enabling of extension_driver qos is optional, it should be
    enabled if service_plugins for qos is also enabled.
    
    • Configure the vxlan range.

      [ml2_type_vxlan]
      ...
      vni_ranges = 1:1000
      
    • Optionally, enable support for VLAN provider and self-service networks on one or more physical networks. If you specify only the physical network, only administrative (privileged) users can manage VLAN networks. Additionally specifying a VLAN ID range for a physical network enables regular (non-privileged) users to manage VLAN networks. The Networking service allocates the VLAN ID for each self-service network using the VLAN ID range for the physical network.

      [ml2_type_vlan]
      ...
      network_vlan_ranges = PHYSICAL_NETWORK:MIN_VLAN_ID:MAX_VLAN_ID
      

      Replace PHYSICAL_NETWORK with the physical network name and optionally define the minimum and maximum VLAN IDs. Use a comma to separate each physical network.

      For example, to enable support for administrative VLAN networks on the physnet1 network and self-service VLAN networks on the physnet2 network using VLAN IDs 1001 to 2000:

      network_vlan_ranges = physnet1,physnet2:1001:2000
      
    • Enable security groups.

      [securitygroup]
      ...
      enable_security_group = true
      
    • Configure ML2 ODL

      [ml2_odl]
      
      ...
      username = <ODL_USERNAME>
      password = <ODL_PASSWORD>
      url = http://<ODL_IP_ADDRESS>:<ODL_PORT>/controller/nb/v2/neutron
      port_binding_controller = pseudo-agentdb-binding
      

Compute/network nodes

Each compute/network node runs the OVS services. If compute/network nodes are already configured to run with Neutron ML2 OVS driver, more steps are necessary. OVSDB with OpenStack can be referred to.

  1. Install the openvswitch packages.

  2. Start the OVS service.

    Using the systemd unit:

    # systemctl start openvswitch
    

    Using the ovs-ctl script:

    # /usr/share/openvswitch/scripts/ovs-ctl start
    
  3. Configure OVS to use ODL as a manager.

    # ovs-vsctl set-manager tcp:${ODL_IP_ADDRESS}:6640
    

    Replace ODL_IP_ADDRESS with the IP address of ODL controller node

  4. Set host OVS configurations if port_binding_controller is pseudo-agent

    # sudo neutron-odl-ovs-hostconfig
    
  5. Verify the OVS service.

    # ovs-vsctl show
    

..note:

After setting config files, you have to restart the neutron server
if you are using screen then it can be directly started from q-svc
window or you can use service neutron-server restart, latter may or
may not work depending on OS you are using.