neutron/tools/ovn_migration
Arnau Verdaguer 2869d008e2 [OVN][Migration] More robust checks on stop-agents
On OSP17 there are some ml2 ovs services that are not
present on some computes eventhought is defined on the
ansible service facts.

This patch will ensure that only those services that are
actually running will be stopped.

Change-Id: I94f0832d09d263837262ada109a567e864915a1a
2022-04-27 10:07:33 +02:00
..
infrared/tripleo-ovn-migration [ovn-migration] Add debug information to create-resources.sh.j2 scripts 2022-03-04 08:58:05 +00:00
tripleo_environment [OVN][Migration] More robust checks on stop-agents 2022-04-27 10:07:33 +02:00
hosts.sample Migrate the OVN migration scripts 2020-03-31 15:35:35 +01:00
migrate-to-ovn.yml Migrate the OVN migration scripts 2020-03-31 15:35:35 +01:00
README.rst Migrate the OVN migration scripts 2020-03-31 15:35:35 +01:00

Migration from ML2/OVS to ML2/OVN

Proof-of-concept ansible script for migrating an OpenStack deployment that uses ML2/OVS to OVN.

If you have a tripleo ML2/OVS deployment then please see the folder tripleo_environment

Prerequisites:

  1. Ansible 2.2 or greater.
  2. ML2/OVS must be using the OVS firewall driver.

To use:

  1. Create an ansible inventory with the expected set of groups and variables as indicated by the hosts-sample file.

  2. Run the playbook:

    $ ansible-playbook migrate-to-ovn.yml -i hosts

Testing Status:

  • Tested on an RDO cloud on CentOS 7.3 based on Ocata.
  • The cloud had 3 controller nodes and 6 compute nodes.
  • Observed network downtime was 10 seconds.
  • The "--forks 10" option was used with ansible-playbook to ensure that commands could be run across the entire environment in parallel.

MTU:

  • If migrating an ML2/OVS deployment using VXLAN tenant networks to an OVN deployment using Geneve for tenant networks, we have an unresolved issue around MTU. The VXLAN overhead is 30 bytes. OVN with Geneve has an overhead of 38 bytes. We need the tenant networks MTU adjusted for OVN and then we need all VMs to receive the updated MTU value through DHCP before the migration can take place. For testing purposes, we've just hacked the Neutron code to indicate that the VXLAN overhead was 38 bytes instead of 30, bypassing the issue at migration time.