Now neutron provides only ovs based mechanism driver since linuxbridge driver was removed. The null driver is intended for testing and the openvswitch interface driver may be the only driver used in real deployments (unless an external plugin is used). Updating the default so that the option becomes optional, which may help users deploy neutron more simply. Change-Id: I0510b76d38168c06468856bfb58c7d561b14868b
8.9 KiB
Networking Option 2: Self-service networks
Install and configure the Networking components on the controller node.
Install the components
# apt install neutron-server neutron-plugin-ml2 \
neutron-openvswitch-agent neutron-l3-agent neutron-dhcp-agent \
neutron-metadata-agent
Configure the server component
Edit the
/etc/neutron/neutron.conffile and complete the following actions:In the
[database]section, configure database access:[database] # ... connection = mysql+pymysql://neutron:NEUTRON_DBPASS@controller/neutronReplace
NEUTRON_DBPASSwith the password you chose for the database.Note
Comment out or remove any other
connectionoptions in the[database]section.In the
[DEFAULT]section, enable the Modular Layer 2 (ML2) plug-in and router service:[DEFAULT] # ... core_plugin = ml2 service_plugins = routerIn the
[DEFAULT]section, configureRabbitMQmessage queue access:[DEFAULT] # ... transport_url = rabbit://openstack:RABBIT_PASS@controllerReplace
RABBIT_PASSwith the password you chose for theopenstackaccount in RabbitMQ.In the
[DEFAULT]and[keystone_authtoken]sections, configure Identity service access:[DEFAULT] # ... auth_strategy = keystone [keystone_authtoken] # ... www_authenticate_uri = http://controller:5000 auth_url = http://controller:5000 memcached_servers = controller:11211 auth_type = password project_domain_name = Default user_domain_name = Default project_name = service username = neutron password = NEUTRON_PASSReplace
NEUTRON_PASSwith the password you chose for theneutronuser in the Identity service.Note
Comment out or remove any other options in the
[keystone_authtoken]section.In the
[DEFAULT]and[nova]sections, configure Networking to notify Compute of network topology changes:[DEFAULT] # ... notify_nova_on_port_status_changes = true notify_nova_on_port_data_changes = true [nova] # ... auth_url = http://controller:5000 auth_type = password project_domain_name = Default user_domain_name = Default region_name = RegionOne project_name = service username = nova password = NOVA_PASSReplace
NOVA_PASSwith the password you chose for thenovauser in the Identity service.
In the
[oslo_concurrency]section, configure the lock path:[oslo_concurrency] # ... lock_path = /var/lib/neutron/tmp
Configure the Modular Layer 2 (ML2) plug-in
The ML2 plug-in uses the Linux bridge mechanism to build layer-2 (bridging and switching) virtual networking infrastructure for instances.
- Edit the
/etc/neutron/plugins/ml2/ml2_conf.inifile and complete the following actions:In the
[ml2]section, enable flat, VLAN, and VXLAN networks:[ml2] # ... type_drivers = flat,vlan,vxlanIn the
[ml2]section, enable VXLAN self-service networks:[ml2] # ... tenant_network_types = vxlanIn the
[ml2]section, enable the Linux bridge and layer-2 population mechanisms:[ml2] # ... mechanism_drivers = openvswitch,l2populationWarning
After you configure the ML2 plug-in, removing values in the
type_driversoption can lead to database inconsistency.Note
The Linux bridge agent only supports VXLAN overlay networks.
In the
[ml2]section, enable the port security extension driver:[ml2] # ... extension_drivers = port_securityIn the
[ml2_type_flat]section, configure the provider virtual network as a flat network:[ml2_type_flat] # ... flat_networks = providerIn the
[ml2_type_vxlan]section, configure the VXLAN network identifier range for self-service networks:[ml2_type_vxlan] # ... vni_ranges = 1:1000
Configure the Open vSwitch agent
The Linux bridge agent builds layer-2 (bridging and switching) virtual networking infrastructure for instances and handles security groups.
- Edit the
/etc/neutron/plugins/ml2/openvswitch_agent.inifile and complete the following actions:In the
[ovs]section, map the provider virtual network to the provider physical bridge and configure the IP address of the physical network interface that handles overlay networks:[ovs] bridge_mappings = provider:PROVIDER_BRIDGE_NAME local_ip = OVERLAY_INTERFACE_IP_ADDRESSReplace
PROVIDER_BRIDGE_NAMEwith the name of the bridge connected to the underlying provider physical network. Seeenvironment-networking-ubuntuand../admin/deploy-ovs-providerfor more information.Also replace
OVERLAY_INTERFACE_IP_ADDRESSwith the IP address of the underlying physical network interface that handles overlay networks. The example architecture uses the management interface to tunnel traffic to the other nodes. Therefore, replaceOVERLAY_INTERFACE_IP_ADDRESSwith the management IP address of the controller node. Seeenvironment-networking-ubuntufor more information.Ensure
PROVIDER_BRIDGE_NAMEexternal bridge is created andPROVIDER_INTERFACE_NAMEis added to that bridge# ovs-vsctl add-br $PROVIDER_BRIDGE_NAME # ovs-vsctl add-port $PROVIDER_BRIDGE_NAME $PROVIDER_INTERFACE_NAMEIn the
[agent]section, enable VXLAN overlay networks and enable layer-2 population:[agent] tunnel_types = vxlan l2_population = trueIn the
[securitygroup]section, enable security groups and configure the Open vSwitch native or the hybrid iptables firewall driver:[securitygroup] # ... enable_security_group = true firewall_driver = openvswitch #firewall_driver = iptables_hybridIn the case of using the hybrid iptables firewall driver, ensure your Linux operating system kernel supports network bridge filters by verifying all the following
sysctlvalues are set to1:net.bridge.bridge-nf-call-iptables net.bridge.bridge-nf-call-ip6tablesTo enable networking bridge support, typically the
br_netfilterkernel module needs to be loaded. Check your operating system's documentation for additional details on enabling this module.
Configure the layer-3 agent
The Layer-3 (L3) agent provides routing and NAT services for self-service virtual networks.
- Edit the
/etc/neutron/l3_agent.inifile in case additional customization is needed.
Configure the DHCP agent
The DHCP agent provides DHCP services for virtual networks.
- Edit the
/etc/neutron/dhcp_agent.inifile and complete the following actions:In the
[DEFAULT]section, configure Dnsmasq DHCP driver, and enable isolated metadata so instances on provider networks can access metadata over the network:[DEFAULT] # ... dhcp_driver = neutron.agent.linux.dhcp.Dnsmasq enable_isolated_metadata = true
Return to Networking controller node configuration.