Retire Packaging Deb project repos
This commit is part of a series to retire the Packaging Deb project. Step 2 is to remove all content from the project repos, replacing it with a README notification where to find ongoing work, and how to recover the repo if needed at some future point (as in https://docs.openstack.org/infra/manual/drivers.html#retiring-a-project). Change-Id: Iba382dfd8c4949b5881a2213bf73f3ddafe979a1
This commit is contained in:
parent
6673921476
commit
80a2beef51
@ -1,7 +0,0 @@
|
||||
[run]
|
||||
branch = True
|
||||
source = networking-arista
|
||||
omit = networking-arista/tests/*,networking-arista/openstack/*
|
||||
|
||||
[report]
|
||||
ignore_errors = True
|
53
.gitignore
vendored
53
.gitignore
vendored
@ -1,53 +0,0 @@
|
||||
*.py[cod]
|
||||
|
||||
# C extensions
|
||||
*.so
|
||||
|
||||
# Packages
|
||||
*.egg
|
||||
*.egg-info
|
||||
dist
|
||||
build
|
||||
eggs
|
||||
parts
|
||||
bin
|
||||
var
|
||||
sdist
|
||||
develop-eggs
|
||||
.installed.cfg
|
||||
lib
|
||||
lib64
|
||||
|
||||
# Installer logs
|
||||
pip-log.txt
|
||||
|
||||
# Unit test / coverage reports
|
||||
.coverage
|
||||
.tox
|
||||
nosetests.xml
|
||||
.testrepository
|
||||
.venv
|
||||
|
||||
# Translations
|
||||
*.mo
|
||||
|
||||
# Mr Developer
|
||||
.mr.developer.cfg
|
||||
.project
|
||||
.pydevproject
|
||||
|
||||
# Complexity
|
||||
output/*.html
|
||||
output/*/index.html
|
||||
|
||||
# Sphinx
|
||||
doc/build
|
||||
|
||||
# pbr generates these
|
||||
AUTHORS
|
||||
ChangeLog
|
||||
|
||||
# Editors
|
||||
*~
|
||||
.*.swp
|
||||
.*sw?
|
@ -1,4 +0,0 @@
|
||||
[gerrit]
|
||||
host=review.openstack.org
|
||||
port=29418
|
||||
project=openstack/networking-arista.git
|
6
.mailmap
6
.mailmap
@ -1,6 +0,0 @@
|
||||
# Format is:
|
||||
# <preferred e-mail> <other e-mail 1>
|
||||
# <preferred e-mail> <other e-mail 2>
|
||||
Sukhdev Kapur <sukhdev@arista.com> <sukhdevkapur@gmail.com>
|
||||
Shashank Hegde <shashank@arista.com>
|
||||
Andre Pech <apech@arista.com>
|
@ -1,8 +0,0 @@
|
||||
[DEFAULT]
|
||||
test_command=OS_STDOUT_CAPTURE=${OS_STDOUT_CAPTURE:-1} \
|
||||
OS_STDERR_CAPTURE=${OS_STDERR_CAPTURE:-1} \
|
||||
OS_TEST_TIMEOUT=${OS_TEST_TIMEOUT:-60} \
|
||||
OS_LOG_CAPTURE=1 \
|
||||
${PYTHON:-python} -m subunit.run discover -t ./ . $LISTOPT $IDOPTION
|
||||
test_id_option=--load-list $IDFILE
|
||||
test_list_option=--list
|
@ -1,16 +0,0 @@
|
||||
If you would like to contribute to the development of OpenStack,
|
||||
you must follow the steps in this page:
|
||||
|
||||
http://docs.openstack.org/infra/manual/developers.html
|
||||
|
||||
Once those steps have been completed, changes to OpenStack
|
||||
should be submitted for review via the Gerrit tool, following
|
||||
the workflow documented at:
|
||||
|
||||
http://docs.openstack.org/infra/manual/developers.html#development-workflow
|
||||
|
||||
Pull requests submitted through GitHub will be ignored.
|
||||
|
||||
Bugs should be filed on Launchpad, not GitHub:
|
||||
|
||||
https://bugs.launchpad.net/networking-arista
|
@ -1,4 +0,0 @@
|
||||
networking-arista Style Commandments
|
||||
===============================================
|
||||
|
||||
Read the OpenStack Style Commandments http://docs.openstack.org/developer/hacking/
|
176
LICENSE
176
LICENSE
@ -1,176 +0,0 @@
|
||||
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. Definitions.
|
||||
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
@ -1,6 +0,0 @@
|
||||
include AUTHORS
|
||||
include ChangeLog
|
||||
exclude .gitignore
|
||||
exclude .gitreview
|
||||
|
||||
global-exclude *.pyc
|
14
README
Normal file
14
README
Normal file
@ -0,0 +1,14 @@
|
||||
This project is no longer maintained.
|
||||
|
||||
The contents of this repository are still available in the Git
|
||||
source code management system. To see the contents of this
|
||||
repository before it reached its end of life, please check out the
|
||||
previous commit with "git checkout HEAD^1".
|
||||
|
||||
For ongoing work on maintaining OpenStack packages in the Debian
|
||||
distribution, please see the Debian OpenStack packaging team at
|
||||
https://wiki.debian.org/OpenStack/.
|
||||
|
||||
For any further questions, please email
|
||||
openstack-dev@lists.openstack.org or join #openstack-dev on
|
||||
Freenode.
|
13
README.rst
13
README.rst
@ -1,13 +0,0 @@
|
||||
===============================
|
||||
networking-arista
|
||||
===============================
|
||||
|
||||
Arista Networking drivers
|
||||
|
||||
* Free software: Apache license
|
||||
* Source: http://git.openstack.org/cgit/openstack/networking-arista
|
||||
|
||||
Features
|
||||
--------
|
||||
|
||||
* TODO
|
@ -1,97 +0,0 @@
|
||||
# -*- mode: shell-script -*-
|
||||
|
||||
function install_lldp() {
|
||||
echo_summary "Installing LLDP"
|
||||
install_package lldpd
|
||||
restart_service lldpd
|
||||
}
|
||||
|
||||
function install_arista_driver() {
|
||||
echo_summary "Installing Arista Driver"
|
||||
setup_develop $ARISTA_DIR
|
||||
}
|
||||
|
||||
function configure_arista() {
|
||||
echo_summary "Configuring Neutron for Arista Driver"
|
||||
cp $ARISTA_ML2_CONF_SAMPLE $ARISTA_ML2_CONF_FILE
|
||||
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista eapi_host $ARISTA_EAPI_HOST
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista eapi_username $ARISTA_EAPI_USERNAME
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista eapi_password $ARISTA_EAPI_PASSWORD
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista api_type $ARISTA_API_TYPE
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista region_name $ARISTA_REGION_NAME
|
||||
|
||||
if [ -n "${ARISTA_USE_FQDN+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista use_fqdn $ARISTA_USE_FQDN
|
||||
fi
|
||||
|
||||
if [ -n "${ARISTA_ML2_SYNC_INTERVAL+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista sync_interval $ARISTA_ML2_SYNC_INTERVAL
|
||||
fi
|
||||
if [ -n "${ARISTA_SEC_GROUP_SUPPORT+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista sec_group_support $ARISTA_SEC_GROUP_SUPPORT
|
||||
fi
|
||||
if [ -n "${ARISTA_SWITCH_INFO+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE ml2_arista switch_info $ARISTA_SWITCH_INFO
|
||||
fi
|
||||
|
||||
if [ -n "${ARISTA_PRIMARY_L3_HOST+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista primary_l3_host $ARISTA_PRIMARY_L3_HOST
|
||||
fi
|
||||
if [ -n "${ARISTA_PRIMARY_L3_HOST_USERNAME+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista primary_l3_host_username $ARISTA_PRIMARY_L3_HOST_USERNAME
|
||||
fi
|
||||
if [ -n "${ARISTA_PRIMARY_L3_HOST_PASSWORD+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista primary_l3_host_password $ARISTA_PRIMARY_L3_HOST_PASSWORD
|
||||
fi
|
||||
if [ -n "${ARISTA_SECONDARY_L3_HOST+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista secondary_l3_host $ARISTA_SECONDARY_L3_HOST
|
||||
fi
|
||||
if [ -n "${ARISTA_SECONDARY_L3_HOST_USERNAME+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista secondary_l3_host_username $ARISTA_SECONDARY_L3_HOST_USERNAME
|
||||
fi
|
||||
if [ -n "${ARISTA_SECONDARY_L3_HOST_PASSWORD+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista secondary_l3_host_password $ARISTA_SECONDARY_L3_HOST_PASSWORD
|
||||
fi
|
||||
if [ -n "${ARISTA_MLAG_CONFIG+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista mlag_config $ARISTA_MLAG_CONFIG
|
||||
fi
|
||||
if [ -n "${ARISTA_USE_VRF+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista use_vrf $ARISTA_USE_VRF
|
||||
fi
|
||||
if [ -n "${ARISTA_L3_SYNC_INTERVAL+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE l3_arista l3_sync_interval $ARISTA_L3_SYNC_INTERVAL
|
||||
fi
|
||||
|
||||
if [ -n "${ARISTA_TYPE_DRIVER_SYNC_INTERVAL+x}" ]; then
|
||||
iniset $ARISTA_ML2_CONF_FILE arista_type_driver sync_interval $ARISTA_TYPE_DRIVER_SYNC_INTERVAL
|
||||
fi
|
||||
neutron_server_config_add $ARISTA_ML2_CONF_FILE
|
||||
}
|
||||
|
||||
if [[ "$1" == "stack" && "$2" == "pre-install" ]]; then
|
||||
if is_service_enabled "q-agt"; then
|
||||
install_lldp
|
||||
fi
|
||||
|
||||
elif [[ "$1" == "stack" && "$2" == "install" ]]; then
|
||||
install_arista_driver
|
||||
|
||||
elif [[ "$1" == "stack" && "$2" == "post-config" ]]; then
|
||||
configure_arista
|
||||
|
||||
elif [[ "$1" == "stack" && "$2" == "extra" ]]; then
|
||||
# no-op
|
||||
:
|
||||
fi
|
||||
|
||||
if [[ "$1" == "unstack" ]]; then
|
||||
# no-op
|
||||
:
|
||||
fi
|
||||
|
||||
if [[ "$1" == "clean" ]]; then
|
||||
# no-op
|
||||
:
|
||||
fi
|
||||
|
@ -1,11 +0,0 @@
|
||||
if ! [[ "$Q_ML2_PLUGIN_MECHANISM_DRIVERS" =~ "arista" ]]; then
|
||||
Q_ML2_PLUGIN_MECHANISM_DRIVERS="$Q_ML2_PLUGIN_MECHANISM_DRIVERS,arista"
|
||||
fi
|
||||
|
||||
ARISTA_DIR=${ARISTA_DIR:-$DEST/networking-arista}
|
||||
|
||||
ARISTA_ML2_CONF_SAMPLE=$ARISTA_DIR/etc/ml2_conf_arista.ini
|
||||
ARISTA_ML2_CONF_FILE=${ARISTA_ML2_CONF_FILE:-"$NEUTRON_CONF_DIR/ml2_conf_arista.ini"}
|
||||
ARISTA_API_TYPE=${ARISTA_API_TYPE:-"EAPI"}
|
||||
ARISTA_REGION_NAME=${ARISTA_REGION_NAME:-"$REGION_NAME"}
|
||||
|
@ -1,75 +0,0 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import os
|
||||
import sys
|
||||
|
||||
sys.path.insert(0, os.path.abspath('../..'))
|
||||
# -- General configuration ----------------------------------------------------
|
||||
|
||||
# Add any Sphinx extension module names here, as strings. They can be
|
||||
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom ones.
|
||||
extensions = [
|
||||
'sphinx.ext.autodoc',
|
||||
#'sphinx.ext.intersphinx',
|
||||
'oslosphinx'
|
||||
]
|
||||
|
||||
# autodoc generation is a bit aggressive and a nuisance when doing heavy
|
||||
# text edit cycles.
|
||||
# execute "export SPHINX_DEBUG=1" in your terminal to disable
|
||||
|
||||
# The suffix of source filenames.
|
||||
source_suffix = '.rst'
|
||||
|
||||
# The master toctree document.
|
||||
master_doc = 'index'
|
||||
|
||||
# General information about the project.
|
||||
project = u'networking-arista'
|
||||
copyright = u'2013, OpenStack Foundation'
|
||||
|
||||
# If true, '()' will be appended to :func: etc. cross-reference text.
|
||||
add_function_parentheses = True
|
||||
|
||||
# If true, the current module name will be prepended to all description
|
||||
# unit titles (such as .. function::).
|
||||
add_module_names = True
|
||||
|
||||
# The name of the Pygments (syntax highlighting) style to use.
|
||||
pygments_style = 'sphinx'
|
||||
|
||||
# -- Options for HTML output --------------------------------------------------
|
||||
|
||||
# The theme to use for HTML and HTML Help pages. Major themes that come with
|
||||
# Sphinx are currently 'default' and 'sphinxdoc'.
|
||||
# html_theme_path = ["."]
|
||||
# html_theme = '_theme'
|
||||
# html_static_path = ['static']
|
||||
|
||||
# Output file base name for HTML help builder.
|
||||
htmlhelp_basename = '%sdoc' % project
|
||||
|
||||
# Grouping the document tree into LaTeX files. List of tuples
|
||||
# (source start file, target name, title, author, documentclass
|
||||
# [howto/manual]).
|
||||
latex_documents = [
|
||||
('index',
|
||||
'%s.tex' % project,
|
||||
u'%s Documentation' % project,
|
||||
u'OpenStack Foundation', 'manual'),
|
||||
]
|
||||
|
||||
# Example configuration for intersphinx: refer to the Python standard library.
|
||||
#intersphinx_mapping = {'http://docs.python.org/': None}
|
@ -1,4 +0,0 @@
|
||||
============
|
||||
Contributing
|
||||
============
|
||||
.. include:: ../../CONTRIBUTING.rst
|
@ -1,25 +0,0 @@
|
||||
.. networking-arista documentation master file, created by
|
||||
sphinx-quickstart on Tue Jul 9 22:26:36 2013.
|
||||
You can adapt this file completely to your liking, but it should at least
|
||||
contain the root `toctree` directive.
|
||||
|
||||
Welcome to networking-arista's documentation!
|
||||
========================================================
|
||||
|
||||
Contents:
|
||||
|
||||
.. toctree::
|
||||
:maxdepth: 2
|
||||
|
||||
readme
|
||||
installation
|
||||
usage
|
||||
contributing
|
||||
|
||||
Indices and tables
|
||||
==================
|
||||
|
||||
* :ref:`genindex`
|
||||
* :ref:`modindex`
|
||||
* :ref:`search`
|
||||
|
@ -1,12 +0,0 @@
|
||||
============
|
||||
Installation
|
||||
============
|
||||
|
||||
At the command line::
|
||||
|
||||
$ pip install networking-arista
|
||||
|
||||
Or, if you have virtualenvwrapper installed::
|
||||
|
||||
$ mkvirtualenv networking-arista
|
||||
$ pip install networking-arista
|
@ -1 +0,0 @@
|
||||
.. include:: ../../README.rst
|
@ -1,7 +0,0 @@
|
||||
========
|
||||
Usage
|
||||
========
|
||||
|
||||
To use networking-arista in a project::
|
||||
|
||||
import networking-arista
|
@ -1,160 +0,0 @@
|
||||
# Defines configuration options specific for Arista ML2 Mechanism driver
|
||||
|
||||
[ml2_arista]
|
||||
# (StrOpt) Comma separated list of IP addresses for all CVX instances in
|
||||
# the high availabilty CVX cluster. This is a required field with
|
||||
# a minimum of one address (if CVX is deployed in a non-redundant
|
||||
# (standalone) manner). If not set, all communications to Arista
|
||||
# EOS will fail.
|
||||
#
|
||||
# eapi_host =
|
||||
# Example: eapi_host = 192.168.0.1, 192.168.11.1, 192.168.22.1
|
||||
#
|
||||
# (StrOpt) EOS command API username. This is required field.
|
||||
# if not set, all communications to Arista EOS will fail.
|
||||
#
|
||||
# eapi_username =
|
||||
# Example: eapi_username = admin
|
||||
#
|
||||
# (StrOpt) EOS command API password. This is required field.
|
||||
# if not set, all communications to Arista EOS will fail.
|
||||
#
|
||||
# eapi_password =
|
||||
# Example: eapi_password = my_password
|
||||
#
|
||||
# (StrOpt) Defines if hostnames are sent to Arista EOS as FQDNs
|
||||
# ("node1.domain.com") or as short names ("node1"). This is
|
||||
# optional. If not set, a value of "True" is assumed.
|
||||
#
|
||||
# use_fqdn =
|
||||
# Example: use_fqdn = True
|
||||
#
|
||||
# (IntOpt) Sync interval in seconds between Neutron plugin and EOS.
|
||||
# This field defines how often the synchronization is performed.
|
||||
# This is an optional field. If not set, a value of 30 seconds
|
||||
# is assumed.
|
||||
#
|
||||
# sync_interval =
|
||||
# Example: sync_interval = 30
|
||||
#
|
||||
# (StrOpt) Defines Region Name that is assigned to this OpenStack Controller.
|
||||
# This is useful when multiple OpenStack/Neutron controllers are
|
||||
# managing the same Arista HW clusters. Note that this name must
|
||||
# match with the region name registered (or known) to keystone
|
||||
# service. Authentication with Keysotne is performed by EOS.
|
||||
# This is optional. If not set, a value of "RegionOne" is assumed.
|
||||
#
|
||||
# region_name =
|
||||
# Example: region_name = RegionOne
|
||||
#
|
||||
# (BoolOpt) Specifies if the Security Groups need to be deployed for baremetal
|
||||
# deployments. If this flag is set to "True", this means switch_info
|
||||
# (see below) must be defined. If this flag is not defined, it is
|
||||
# assumed to be False.
|
||||
#
|
||||
# sec_group_support =
|
||||
# Example: sec_group_support = True
|
||||
#
|
||||
# (ListOpt) This is a comma separated list of Arista switches where
|
||||
# security groups (i.e. ACLs) need to be applied. Each string has
|
||||
# three values separated by ":" in the following format.
|
||||
# <switch IP>:<username>:<password>,<switch IP>:<username>:<password>
|
||||
# This is required if sec_group_support is set to "True"
|
||||
#
|
||||
# switch_info =
|
||||
# Example: switch_info = 172.13.23.55:admin:admin,172.13.23.56:admin:admin
|
||||
#
|
||||
# (StrOpt) Tells the plugin to use a sepcific API interfaces to communicate
|
||||
# with CVX. Valid options are:
|
||||
# EAPI - Use EOS' extensible API.
|
||||
# JSON - Use EOS' JSON/REST API.
|
||||
# api_type =
|
||||
# Example: api_type = EAPI
|
||||
#
|
||||
# (ListOpt) This is a comma separated list of physical networks which are
|
||||
# managed by Arista switches. This list will be used in
|
||||
# by the Arista ML2 plugin to make the decision if it can
|
||||
# participate on binding or updating a port.
|
||||
#
|
||||
# managed_physnets =
|
||||
# Example: managed_physnets = arista_network
|
||||
#
|
||||
# (BoolOpt) Specifies whether the Arista ML2 plugin should bind ports to vxlan
|
||||
# fabric segments and dynamically allocate vlan segments based on
|
||||
# the host to connect the port to the vxlan fabric.
|
||||
#
|
||||
# manage_fabric =
|
||||
# Example: manage_fabric = False
|
||||
|
||||
|
||||
[l3_arista]
|
||||
|
||||
# (StrOpt) primary host IP address. This is required field. If not set, all
|
||||
# communications to Arista EOS will fail. This is the host where
|
||||
# primary router is created.
|
||||
#
|
||||
# primary_l3_host =
|
||||
# Example: primary_l3_host = 192.168.10.10
|
||||
#
|
||||
# (StrOpt) Primary host username. This is required field.
|
||||
# if not set, all communications to Arista EOS will fail.
|
||||
#
|
||||
# primary_l3_host_username =
|
||||
# Example: primary_l3_username = admin
|
||||
#
|
||||
# (StrOpt) Primary host password. This is required field.
|
||||
# if not set, all communications to Arista EOS will fail.
|
||||
#
|
||||
# primary_l3_host_password =
|
||||
# Example: primary_l3_password = my_password
|
||||
#
|
||||
# (StrOpt) IP address of the second Arista switch paired as
|
||||
# MLAG (Multi-chassis Link Aggregation) with the first.
|
||||
# This is optional field, however, if mlag_config flag is set,
|
||||
# then this is a required field. If not set, all
|
||||
# communications to Arista EOS will fail. If mlag_config is set
|
||||
# to False, then this field is ignored
|
||||
#
|
||||
# secondary_l3_host =
|
||||
# Example: secondary_l3_host = 192.168.10.20
|
||||
#
|
||||
# (IntOpt) Connection timeout interval in seconds. This interval
|
||||
# defines how long an EAPI request from the driver to '
|
||||
# EOS waits before timing out. If not set, a value of 10
|
||||
# seconds is assumed.
|
||||
#
|
||||
# conn_timeout =
|
||||
# Example: conn_timeout = 10
|
||||
#
|
||||
# (BoolOpt) Defines if Arista switches are configured in MLAG mode
|
||||
# If yes, all L3 configuration is pushed to both switches
|
||||
# automatically. If this flag is set, ensure that secondary_l3_host
|
||||
# is set to the second switch's IP.
|
||||
# This flag is Optional. If not set, a value of "False" is assumed.
|
||||
#
|
||||
# mlag_config =
|
||||
# Example: mlag_config = True
|
||||
#
|
||||
# (BoolOpt) Defines if the router is created in default VRF or a
|
||||
# a specific VRF. This is optional.
|
||||
# If not set, a value of "False" is assumed.
|
||||
#
|
||||
# Example: use_vrf = True
|
||||
#
|
||||
# (IntOpt) Sync interval in seconds between Neutron plugin and EOS.
|
||||
# This field defines how often the synchronization is performed.
|
||||
# This is an optional field. If not set, a value of 180 seconds
|
||||
# is assumed.
|
||||
#
|
||||
# l3_sync_interval =
|
||||
# Example: l3_sync_interval = 60
|
||||
|
||||
[arista_type_driver]
|
||||
|
||||
# (IntOpt) VLAN Sync interval in seconds between the type driver and EOS.
|
||||
# This interval defines how often the VLAN synchronization is
|
||||
# performed. This is an optional field. If not set, a value of
|
||||
# 10 seconds is assumed.
|
||||
#
|
||||
# sync_interval =
|
||||
# Example: sync_interval = 10
|
143
etc/policy.json
143
etc/policy.json
@ -1,143 +0,0 @@
|
||||
{
|
||||
"context_is_admin": "role:admin",
|
||||
"admin_or_owner": "rule:context_is_admin or tenant_id:%(tenant_id)s",
|
||||
"context_is_advsvc": "role:advsvc",
|
||||
"admin_or_network_owner": "rule:context_is_admin or tenant_id:%(network:tenant_id)s",
|
||||
"admin_only": "rule:context_is_admin",
|
||||
"regular_user": "",
|
||||
"shared": "field:networks:shared=True",
|
||||
"shared_firewalls": "field:firewalls:shared=True",
|
||||
"external": "field:networks:router:external=True",
|
||||
"default": "rule:admin_or_owner",
|
||||
|
||||
"create_subnet": "rule:admin_or_network_owner",
|
||||
"get_subnet": "rule:admin_or_owner or rule:shared",
|
||||
"update_subnet": "rule:admin_or_network_owner",
|
||||
"delete_subnet": "rule:admin_or_network_owner",
|
||||
|
||||
"create_network": "",
|
||||
"get_network": "rule:admin_or_owner or rule:shared or rule:external or rule:context_is_advsvc",
|
||||
"get_network:router:external": "rule:regular_user",
|
||||
"get_network:segments": "rule:admin_only",
|
||||
"get_network:provider:network_type": "rule:admin_only",
|
||||
"get_network:provider:physical_network": "rule:admin_only",
|
||||
"get_network:provider:segmentation_id": "rule:admin_only",
|
||||
"get_network:queue_id": "rule:admin_only",
|
||||
"create_network:shared": "rule:admin_only",
|
||||
"create_network:router:external": "rule:admin_only",
|
||||
"create_network:segments": "rule:admin_only",
|
||||
"create_network:provider:network_type": "rule:admin_only",
|
||||
"create_network:provider:physical_network": "rule:admin_only",
|
||||
"create_network:provider:segmentation_id": "rule:admin_only",
|
||||
"update_network": "rule:admin_or_owner",
|
||||
"update_network:segments": "rule:admin_only",
|
||||
"update_network:shared": "rule:admin_only",
|
||||
"update_network:provider:network_type": "rule:admin_only",
|
||||
"update_network:provider:physical_network": "rule:admin_only",
|
||||
"update_network:provider:segmentation_id": "rule:admin_only",
|
||||
"update_network:router:external": "rule:admin_only",
|
||||
"delete_network": "rule:admin_or_owner",
|
||||
|
||||
"create_port": "",
|
||||
"create_port:mac_address": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"create_port:fixed_ips": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"create_port:port_security_enabled": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"create_port:binding:host_id": "rule:admin_only",
|
||||
"create_port:binding:profile": "rule:admin_only",
|
||||
"create_port:mac_learning_enabled": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"get_port": "rule:admin_or_owner or rule:context_is_advsvc",
|
||||
"get_port:queue_id": "rule:admin_only",
|
||||
"get_port:binding:vif_type": "rule:admin_only",
|
||||
"get_port:binding:vif_details": "rule:admin_only",
|
||||
"get_port:binding:host_id": "rule:admin_only",
|
||||
"get_port:binding:profile": "rule:admin_only",
|
||||
"update_port": "rule:admin_or_owner or rule:context_is_advsvc",
|
||||
"update_port:fixed_ips": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"update_port:port_security_enabled": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"update_port:binding:host_id": "rule:admin_only",
|
||||
"update_port:binding:profile": "rule:admin_only",
|
||||
"update_port:mac_learning_enabled": "rule:admin_or_network_owner or rule:context_is_advsvc",
|
||||
"delete_port": "rule:admin_or_owner or rule:context_is_advsvc",
|
||||
|
||||
"get_router:ha": "rule:admin_only",
|
||||
"create_router": "rule:regular_user",
|
||||
"create_router:external_gateway_info:enable_snat": "rule:admin_only",
|
||||
"create_router:distributed": "rule:admin_only",
|
||||
"create_router:ha": "rule:admin_only",
|
||||
"get_router": "rule:admin_or_owner",
|
||||
"get_router:distributed": "rule:admin_only",
|
||||
"update_router:external_gateway_info:enable_snat": "rule:admin_only",
|
||||
"update_router:distributed": "rule:admin_only",
|
||||
"update_router:ha": "rule:admin_only",
|
||||
"delete_router": "rule:admin_or_owner",
|
||||
|
||||
"add_router_interface": "rule:admin_or_owner",
|
||||
"remove_router_interface": "rule:admin_or_owner",
|
||||
|
||||
"create_router:external_gateway_info:external_fixed_ips": "rule:admin_only",
|
||||
"update_router:external_gateway_info:external_fixed_ips": "rule:admin_only",
|
||||
|
||||
"create_firewall": "",
|
||||
"get_firewall": "rule:admin_or_owner",
|
||||
"create_firewall:shared": "rule:admin_only",
|
||||
"get_firewall:shared": "rule:admin_only",
|
||||
"update_firewall": "rule:admin_or_owner",
|
||||
"update_firewall:shared": "rule:admin_only",
|
||||
"delete_firewall": "rule:admin_or_owner",
|
||||
|
||||
"create_firewall_policy": "",
|
||||
"get_firewall_policy": "rule:admin_or_owner or rule:shared_firewalls",
|
||||
"create_firewall_policy:shared": "rule:admin_or_owner",
|
||||
"update_firewall_policy": "rule:admin_or_owner",
|
||||
"delete_firewall_policy": "rule:admin_or_owner",
|
||||
|
||||
"create_firewall_rule": "",
|
||||
"get_firewall_rule": "rule:admin_or_owner or rule:shared_firewalls",
|
||||
"update_firewall_rule": "rule:admin_or_owner",
|
||||
"delete_firewall_rule": "rule:admin_or_owner",
|
||||
|
||||
"create_qos_queue": "rule:admin_only",
|
||||
"get_qos_queue": "rule:admin_only",
|
||||
|
||||
"update_agent": "rule:admin_only",
|
||||
"delete_agent": "rule:admin_only",
|
||||
"get_agent": "rule:admin_only",
|
||||
|
||||
"create_dhcp-network": "rule:admin_only",
|
||||
"delete_dhcp-network": "rule:admin_only",
|
||||
"get_dhcp-networks": "rule:admin_only",
|
||||
"create_l3-router": "rule:admin_only",
|
||||
"delete_l3-router": "rule:admin_only",
|
||||
"get_l3-routers": "rule:admin_only",
|
||||
"get_dhcp-agents": "rule:admin_only",
|
||||
"get_l3-agents": "rule:admin_only",
|
||||
"get_loadbalancer-agent": "rule:admin_only",
|
||||
"get_loadbalancer-pools": "rule:admin_only",
|
||||
|
||||
"create_floatingip": "rule:regular_user",
|
||||
"create_floatingip:floating_ip_address": "rule:admin_only",
|
||||
"update_floatingip": "rule:admin_or_owner",
|
||||
"delete_floatingip": "rule:admin_or_owner",
|
||||
"get_floatingip": "rule:admin_or_owner",
|
||||
|
||||
"create_network_profile": "rule:admin_only",
|
||||
"update_network_profile": "rule:admin_only",
|
||||
"delete_network_profile": "rule:admin_only",
|
||||
"get_network_profiles": "",
|
||||
"get_network_profile": "",
|
||||
"update_policy_profiles": "rule:admin_only",
|
||||
"get_policy_profiles": "",
|
||||
"get_policy_profile": "",
|
||||
|
||||
"create_metering_label": "rule:admin_only",
|
||||
"delete_metering_label": "rule:admin_only",
|
||||
"get_metering_label": "rule:admin_only",
|
||||
|
||||
"create_metering_label_rule": "rule:admin_only",
|
||||
"delete_metering_label_rule": "rule:admin_only",
|
||||
"get_metering_label_rule": "rule:admin_only",
|
||||
|
||||
"get_service_provider": "rule:regular_user",
|
||||
"get_lsn": "rule:admin_only",
|
||||
"create_lsn": "rule:admin_only"
|
||||
}
|
@ -1,27 +0,0 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import gettext
|
||||
|
||||
import pbr.version
|
||||
import six
|
||||
|
||||
|
||||
__version__ = pbr.version.VersionInfo(
|
||||
'networking_arista').version_string()
|
||||
|
||||
if six.PY2:
|
||||
gettext.install('networking_arista', unicode=1)
|
||||
else:
|
||||
gettext.install('networking_arista')
|
@ -1,42 +0,0 @@
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import oslo_i18n
|
||||
|
||||
DOMAIN = "networking_arista"
|
||||
|
||||
_translators = oslo_i18n.TranslatorFactory(domain=DOMAIN)
|
||||
|
||||
# The primary translation function using the well-known name "_"
|
||||
_ = _translators.primary
|
||||
|
||||
# The contextual translation function using the name "_C"
|
||||
_C = _translators.contextual_form
|
||||
|
||||
# The plural translation function using the name "_P"
|
||||
_P = _translators.plural_form
|
||||
|
||||
# Translators for log levels.
|
||||
#
|
||||
# The abbreviated names are meant to reflect the usual use of a short
|
||||
# name like '_'. The "L" is for "log" and the other letter comes from
|
||||
# the level.
|
||||
_LI = _translators.log_info
|
||||
_LW = _translators.log_warning
|
||||
_LE = _translators.log_error
|
||||
_LC = _translators.log_critical
|
||||
|
||||
|
||||
def get_available_languages():
|
||||
return oslo_i18n.get_available_languages(DOMAIN)
|
@ -1,133 +0,0 @@
|
||||
# Copyright (c) 2017 Arista Networks, Inc
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import json
|
||||
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import excutils
|
||||
import requests
|
||||
from requests import exceptions as requests_exc
|
||||
from six.moves.urllib import parse
|
||||
|
||||
from networking_arista._i18n import _LI, _LW
|
||||
from networking_arista.common import exceptions as arista_exc
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
# EAPI error message
|
||||
ERR_CVX_NOT_LEADER = 'only available on cluster leader'
|
||||
|
||||
|
||||
class EAPIClient(object):
|
||||
def __init__(self, host, username=None, password=None, verify=False,
|
||||
timeout=None):
|
||||
self.host = host
|
||||
self.timeout = timeout
|
||||
self.url = self._make_url(host)
|
||||
self.session = requests.Session()
|
||||
self.session.headers['Content-Type'] = 'application/json'
|
||||
self.session.headers['Accept'] = 'application/json'
|
||||
self.session.verify = verify
|
||||
if username and password:
|
||||
self.session.auth = (username, password)
|
||||
|
||||
@staticmethod
|
||||
def _make_url(host, scheme='https'):
|
||||
return parse.urlunsplit(
|
||||
(scheme, host, '/command-api', '', '')
|
||||
)
|
||||
|
||||
def execute(self, commands, commands_to_log=None):
|
||||
params = {
|
||||
'timestamps': False,
|
||||
'format': 'json',
|
||||
'version': 1,
|
||||
'cmds': commands
|
||||
}
|
||||
|
||||
data = {
|
||||
'id': 'Networking Arista Driver',
|
||||
'method': 'runCmds',
|
||||
'jsonrpc': '2.0',
|
||||
'params': params
|
||||
}
|
||||
|
||||
if commands_to_log:
|
||||
log_data = dict(data)
|
||||
log_data['params'] = dict(params)
|
||||
log_data['params']['cmds'] = commands_to_log
|
||||
else:
|
||||
log_data = data
|
||||
|
||||
LOG.info(
|
||||
_LI('EAPI request %(ip)s contains %(data)s'),
|
||||
{'ip': self.host, 'data': json.dumps(log_data)}
|
||||
)
|
||||
|
||||
# request handling
|
||||
try:
|
||||
error = None
|
||||
response = self.session.post(
|
||||
self.url,
|
||||
data=json.dumps(data),
|
||||
timeout=self.timeout
|
||||
)
|
||||
except requests_exc.ConnectionError:
|
||||
error = _LW('Error while trying to connect to %(ip)s')
|
||||
except requests_exc.ConnectTimeout:
|
||||
error = _LW('Timed out while trying to connect to %(ip)s')
|
||||
except requests_exc.Timeout:
|
||||
error = _LW('Timed out during an EAPI request to %(ip)s')
|
||||
except requests_exc.InvalidURL:
|
||||
error = _LW('Ingoring attempt to connect to invalid URL at %(ip)s')
|
||||
except Exception as e:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.warning(
|
||||
_LW('Error during processing the EAPI request %(error)s'),
|
||||
{'error': e}
|
||||
)
|
||||
finally:
|
||||
if error:
|
||||
msg = error % {'ip': self.host}
|
||||
# stop processing since we've encountered request error
|
||||
LOG.warning(msg)
|
||||
raise arista_exc.AristaRpcError(msg=msg)
|
||||
|
||||
# response handling
|
||||
try:
|
||||
resp_data = response.json()
|
||||
return resp_data['result']
|
||||
except ValueError as e:
|
||||
LOG.info(_LI('Ignoring invalid JSON response'))
|
||||
except KeyError:
|
||||
if 'error' in resp_data and resp_data['error']['code'] == 1002:
|
||||
for d in resp_data['error']['data']:
|
||||
if not isinstance(d, dict):
|
||||
continue
|
||||
elif ERR_CVX_NOT_LEADER in d.get('errors', {})[0]:
|
||||
LOG.info(
|
||||
_LI('%(ip)s is not the CVX leader'),
|
||||
{'ip': self.host}
|
||||
)
|
||||
return
|
||||
msg = _LI('Unexpected EAPI error')
|
||||
LOG.info(msg)
|
||||
raise arista_exc.AristaRpcError(msg=msg)
|
||||
except Exception as e:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.warning(
|
||||
_LW('Error during processing the EAPI response %(error)s'),
|
||||
{'error': e}
|
||||
)
|
@ -1,194 +0,0 @@
|
||||
# Copyright (c) 2013 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
from oslo_config import cfg
|
||||
|
||||
from networking_arista._i18n import _
|
||||
|
||||
|
||||
# Arista ML2 Mechanism driver specific configuration knobs.
|
||||
#
|
||||
# Following are user configurable options for Arista ML2 Mechanism
|
||||
# driver. The eapi_username, eapi_password, and eapi_host are
|
||||
# required options. Region Name must be the same that is used by
|
||||
# Keystone service. This option is available to support multiple
|
||||
# OpenStack/Neutron controllers.
|
||||
|
||||
ARISTA_DRIVER_OPTS = [
|
||||
cfg.StrOpt('eapi_username',
|
||||
default='',
|
||||
help=_('Username for Arista EOS. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail.')),
|
||||
cfg.StrOpt('eapi_password',
|
||||
default='',
|
||||
secret=True, # do not expose value in the logs
|
||||
help=_('Password for Arista EOS. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail.')),
|
||||
cfg.StrOpt('eapi_host',
|
||||
default='',
|
||||
help=_('Arista EOS IP address. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail.')),
|
||||
cfg.BoolOpt('use_fqdn',
|
||||
default=True,
|
||||
help=_('Defines if hostnames are sent to Arista EOS as FQDNs '
|
||||
'("node1.domain.com") or as short names ("node1"). '
|
||||
'This is optional. If not set, a value of "True" '
|
||||
'is assumed.')),
|
||||
cfg.IntOpt('sync_interval',
|
||||
default=30,
|
||||
help=_('Sync interval in seconds between Neutron plugin and '
|
||||
'EOS. This interval defines how often the '
|
||||
'synchronization is performed. This is an optional '
|
||||
'field. If not set, a value of 30 seconds is '
|
||||
'assumed.')),
|
||||
cfg.IntOpt('conn_timeout',
|
||||
default=10,
|
||||
help=_('Connection timeout interval in seconds. This interval '
|
||||
'defines how long an EAPI request from the driver to '
|
||||
'EOS waits before timing out. If not set, a value of 10 '
|
||||
'seconds is assumed.')),
|
||||
cfg.StrOpt('region_name',
|
||||
default='RegionOne',
|
||||
help=_('Defines Region Name that is assigned to this OpenStack '
|
||||
'Controller. This is useful when multiple '
|
||||
'OpenStack/Neutron controllers are managing the same '
|
||||
'Arista HW clusters. Note that this name must match '
|
||||
'with the region name registered (or known) to keystone '
|
||||
'service. Authentication with Keysotne is performed by '
|
||||
'EOS. This is optional. If not set, a value of '
|
||||
'"RegionOne" is assumed.')),
|
||||
cfg.BoolOpt('sec_group_support',
|
||||
default=False,
|
||||
help=_('Specifies if the Security Groups needs to deployed '
|
||||
'for baremetal deployments. If this flag is set to '
|
||||
'True, this means switch_info(see below) must be '
|
||||
'defined. If this flag is not defined, it is assumed '
|
||||
'to be False')),
|
||||
cfg.ListOpt('switch_info',
|
||||
default=[],
|
||||
help=_('This is a comma separated list of Arista switches '
|
||||
'where security groups (i.e. ACLs) need to be '
|
||||
'applied. Each string has three values separated '
|
||||
'by : in the follow format '
|
||||
'<IP of switch>:<username>:<password>, ...... '
|
||||
'For Example: 172.13.23.55:admin:admin, '
|
||||
'172.13.23.56:admin:admin, .... '
|
||||
'This is required if sec_group_support is set to '
|
||||
'"True"')),
|
||||
cfg.StrOpt('api_type',
|
||||
default='JSON',
|
||||
help=_('Tells the plugin to use a sepcific API interfaces '
|
||||
'to communicate with CVX. Valid options are:'
|
||||
'EAPI - Use EOS\' extensible API.'
|
||||
'JSON - Use EOS\' JSON/REST API.')),
|
||||
cfg.ListOpt('managed_physnets',
|
||||
default=[],
|
||||
help=_('This is a comma separated list of physical networks '
|
||||
'which are managed by Arista switches.'
|
||||
'This list will be used by the Arista ML2 plugin'
|
||||
'to make the decision if it can participate in binding'
|
||||
'or updating a port.'
|
||||
'For Example: '
|
||||
'managed_physnets = arista_network')),
|
||||
cfg.BoolOpt('manage_fabric',
|
||||
default=False,
|
||||
help=_('Specifies whether the Arista ML2 plugin should bind '
|
||||
'ports to vxlan fabric segments and dynamically '
|
||||
'allocate vlan segments based on the host to connect '
|
||||
'the port to the vxlan fabric')),
|
||||
]
|
||||
|
||||
|
||||
""" Arista L3 Service Plugin specific configuration knobs.
|
||||
|
||||
Following are user configurable options for Arista L3 plugin
|
||||
driver. The eapi_username, eapi_password, and eapi_host are
|
||||
required options.
|
||||
"""
|
||||
|
||||
ARISTA_L3_PLUGIN = [
|
||||
cfg.StrOpt('primary_l3_host_username',
|
||||
default='',
|
||||
help=_('Username for Arista EOS. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail')),
|
||||
cfg.StrOpt('primary_l3_host_password',
|
||||
default='',
|
||||
secret=True, # do not expose value in the logs
|
||||
help=_('Password for Arista EOS. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail')),
|
||||
cfg.StrOpt('primary_l3_host',
|
||||
default='',
|
||||
help=_('Arista EOS IP address. This is required field. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail')),
|
||||
cfg.StrOpt('secondary_l3_host',
|
||||
default='',
|
||||
help=_('Arista EOS IP address for second Switch MLAGed with '
|
||||
'the first one. This an optional field, however, if '
|
||||
'mlag_config flag is set, then this is required. '
|
||||
'If not set, all communications to Arista EOS '
|
||||
'will fail')),
|
||||
cfg.IntOpt('conn_timeout',
|
||||
default=10,
|
||||
help=_('Connection timeout interval in seconds. This interval '
|
||||
'defines how long an EAPI request from the driver to '
|
||||
'EOS waits before timing out. If not set, a value of 10 '
|
||||
'seconds is assumed.')),
|
||||
cfg.BoolOpt('mlag_config',
|
||||
default=False,
|
||||
help=_('This flag is used indicate if Arista Switches are '
|
||||
'configured in MLAG mode. If yes, all L3 config '
|
||||
'is pushed to both the switches automatically. '
|
||||
'If this flag is set to True, ensure to specify IP '
|
||||
'addresses of both switches. '
|
||||
'This is optional. If not set, a value of "False" '
|
||||
'is assumed.')),
|
||||
cfg.BoolOpt('use_vrf',
|
||||
default=False,
|
||||
help=_('A "True" value for this flag indicates to create a '
|
||||
'router in VRF. If not set, all routers are created '
|
||||
'in default VRF. '
|
||||
'This is optional. If not set, a value of "False" '
|
||||
'is assumed.')),
|
||||
cfg.IntOpt('l3_sync_interval',
|
||||
default=180,
|
||||
help=_('Sync interval in seconds between L3 Service plugin '
|
||||
'and EOS. This interval defines how often the '
|
||||
'synchronization is performed. This is an optional '
|
||||
'field. If not set, a value of 180 seconds is assumed'))
|
||||
]
|
||||
|
||||
|
||||
ARISTA_TYPE_DRIVER_OPTS = [
|
||||
cfg.IntOpt('sync_interval',
|
||||
default=10,
|
||||
help=_('VLAN Sync interval in seconds between Neutron plugin '
|
||||
'and EOS. This interval defines how often the VLAN '
|
||||
'synchronization is performed. This is an optional '
|
||||
'field. If not set, a value of 10 seconds is '
|
||||
'assumed.')),
|
||||
]
|
||||
|
||||
cfg.CONF.register_opts(ARISTA_L3_PLUGIN, "l3_arista")
|
||||
|
||||
cfg.CONF.register_opts(ARISTA_DRIVER_OPTS, "ml2_arista")
|
||||
|
||||
cfg.CONF.register_opts(ARISTA_TYPE_DRIVER_OPTS, "arista_type_driver")
|
@ -1,83 +0,0 @@
|
||||
# Copyright (c) 2013 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from neutron_lib.db import constants as db_const
|
||||
from neutron_lib.db import model_base
|
||||
import sqlalchemy as sa
|
||||
|
||||
UUID_LEN = 36
|
||||
STR_LEN = 255
|
||||
|
||||
|
||||
class HasTenant(object):
|
||||
"""Tenant mixin, add to subclasses that have a tenant."""
|
||||
|
||||
tenant_id = sa.Column(sa.String(db_const.PROJECT_ID_FIELD_SIZE),
|
||||
index=True)
|
||||
|
||||
|
||||
class AristaProvisionedNets(model_base.BASEV2, model_base.HasId,
|
||||
HasTenant):
|
||||
"""Stores networks provisioned on Arista EOS.
|
||||
|
||||
Saves the segmentation ID for each network that is provisioned
|
||||
on EOS. This information is used during synchronization between
|
||||
Neutron and EOS.
|
||||
"""
|
||||
__tablename__ = 'arista_provisioned_nets'
|
||||
|
||||
network_id = sa.Column(sa.String(UUID_LEN))
|
||||
segmentation_id = sa.Column(sa.Integer)
|
||||
|
||||
def eos_network_representation(self, segmentation_type):
|
||||
return {u'networkId': self.network_id,
|
||||
u'segmentationTypeId': self.segmentation_id,
|
||||
u'segmentationType': segmentation_type,
|
||||
u'tenantId': self.tenant_id,
|
||||
u'segmentId': self.id,
|
||||
}
|
||||
|
||||
|
||||
class AristaProvisionedVms(model_base.BASEV2, model_base.HasId,
|
||||
HasTenant):
|
||||
"""Stores VMs provisioned on Arista EOS.
|
||||
|
||||
All VMs launched on physical hosts connected to Arista
|
||||
Switches are remembered
|
||||
"""
|
||||
__tablename__ = 'arista_provisioned_vms'
|
||||
|
||||
vm_id = sa.Column(sa.String(STR_LEN))
|
||||
host_id = sa.Column(sa.String(STR_LEN))
|
||||
port_id = sa.Column(sa.String(UUID_LEN))
|
||||
network_id = sa.Column(sa.String(UUID_LEN))
|
||||
|
||||
def eos_port_representation(self):
|
||||
return {u'portId': self.port_id,
|
||||
u'deviceId': self.vm_id,
|
||||
u'hosts': [self.host_id],
|
||||
u'networkId': self.network_id}
|
||||
|
||||
|
||||
class AristaProvisionedTenants(model_base.BASEV2, model_base.HasId,
|
||||
HasTenant):
|
||||
"""Stores Tenants provisioned on Arista EOS.
|
||||
|
||||
Tenants list is maintained for sync between Neutron and EOS.
|
||||
"""
|
||||
__tablename__ = 'arista_provisioned_tenants'
|
||||
|
||||
def eos_tenant_representation(self):
|
||||
return {u'tenantId': self.tenant_id}
|
@ -1,595 +0,0 @@
|
||||
# Copyright (c) 2013 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from neutron_lib import constants as n_const
|
||||
from neutron_lib import context as nctx
|
||||
from neutron_lib.plugins.ml2 import api as driver_api
|
||||
|
||||
import neutron.db.api as db
|
||||
from neutron.db import db_base_plugin_v2
|
||||
from neutron.db import securitygroups_db as sec_db
|
||||
from neutron.db import segments_db
|
||||
from neutron.plugins.ml2 import models as ml2_models
|
||||
|
||||
from networking_arista.common import db as db_models
|
||||
|
||||
VLAN_SEGMENTATION = 'vlan'
|
||||
|
||||
|
||||
def remember_tenant(tenant_id):
|
||||
"""Stores a tenant information in repository.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
tenant = (session.query(db_models.AristaProvisionedTenants).
|
||||
filter_by(tenant_id=tenant_id).first())
|
||||
if not tenant:
|
||||
tenant = db_models.AristaProvisionedTenants(tenant_id=tenant_id)
|
||||
session.add(tenant)
|
||||
|
||||
|
||||
def forget_tenant(tenant_id):
|
||||
"""Removes a tenant information from repository.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
(session.query(db_models.AristaProvisionedTenants).
|
||||
filter_by(tenant_id=tenant_id).
|
||||
delete())
|
||||
|
||||
|
||||
def get_all_tenants():
|
||||
"""Returns a list of all tenants stored in repository."""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
return session.query(db_models.AristaProvisionedTenants).all()
|
||||
|
||||
|
||||
def num_provisioned_tenants():
|
||||
"""Returns number of tenants stored in repository."""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
return session.query(db_models.AristaProvisionedTenants).count()
|
||||
|
||||
|
||||
def remember_vm(vm_id, host_id, port_id, network_id, tenant_id):
|
||||
"""Stores all relevant information about a VM in repository.
|
||||
|
||||
:param vm_id: globally unique identifier for VM instance
|
||||
:param host_id: ID of the host where the VM is placed
|
||||
:param port_id: globally unique port ID that connects VM to network
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
vm = db_models.AristaProvisionedVms(
|
||||
vm_id=vm_id,
|
||||
host_id=host_id,
|
||||
port_id=port_id,
|
||||
network_id=network_id,
|
||||
tenant_id=tenant_id)
|
||||
session.add(vm)
|
||||
|
||||
|
||||
def forget_all_ports_for_network(net_id):
|
||||
"""Removes all ports for a given network fron repository.
|
||||
|
||||
:param net_id: globally unique network ID
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
(session.query(db_models.AristaProvisionedVms).
|
||||
filter_by(network_id=net_id).delete())
|
||||
|
||||
|
||||
def update_port(vm_id, host_id, port_id, network_id, tenant_id):
|
||||
"""Updates the port details in the database.
|
||||
|
||||
:param vm_id: globally unique identifier for VM instance
|
||||
:param host_id: ID of the new host where the VM is placed
|
||||
:param port_id: globally unique port ID that connects VM to network
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
port = session.query(db_models.AristaProvisionedVms).filter_by(
|
||||
port_id=port_id).first()
|
||||
if port:
|
||||
# Update the VM's host id
|
||||
port.host_id = host_id
|
||||
port.vm_id = vm_id
|
||||
port.network_id = network_id
|
||||
port.tenant_id = tenant_id
|
||||
|
||||
|
||||
def forget_port(port_id, host_id):
|
||||
"""Deletes the port from the database
|
||||
|
||||
:param port_id: globally unique port ID that connects VM to network
|
||||
:param host_id: host to which the port is bound to
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
session.query(db_models.AristaProvisionedVms).filter_by(
|
||||
port_id=port_id,
|
||||
host_id=host_id).delete()
|
||||
|
||||
|
||||
def remember_network_segment(tenant_id,
|
||||
network_id, segmentation_id, segment_id):
|
||||
"""Stores all relevant information about a Network in repository.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param segmentation_id: segmentation id that is assigned to the network
|
||||
:param segment_id: globally unique neutron network segment identifier
|
||||
"""
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
net = db_models.AristaProvisionedNets(
|
||||
tenant_id=tenant_id,
|
||||
id=segment_id,
|
||||
network_id=network_id,
|
||||
segmentation_id=segmentation_id)
|
||||
session.add(net)
|
||||
|
||||
|
||||
def forget_network_segment(tenant_id, network_id, segment_id=None):
|
||||
"""Deletes all relevant information about a Network from repository.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param segment_id: globally unique neutron network segment identifier
|
||||
"""
|
||||
filters = {
|
||||
'tenant_id': tenant_id,
|
||||
'network_id': network_id
|
||||
}
|
||||
if segment_id:
|
||||
filters['id'] = segment_id
|
||||
|
||||
session = db.get_writer_session()
|
||||
with session.begin():
|
||||
(session.query(db_models.AristaProvisionedNets).
|
||||
filter_by(**filters).delete())
|
||||
|
||||
|
||||
def get_segmentation_id(tenant_id, network_id):
|
||||
"""Returns Segmentation ID (VLAN) associated with a network.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
:param network_id: globally unique neutron network identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
net = (session.query(db_models.AristaProvisionedNets).
|
||||
filter_by(tenant_id=tenant_id,
|
||||
network_id=network_id).first())
|
||||
return net.segmentation_id if net else None
|
||||
|
||||
|
||||
def is_vm_provisioned(vm_id, host_id, port_id,
|
||||
network_id, tenant_id):
|
||||
"""Checks if a VM is already known to EOS
|
||||
|
||||
:returns: True, if yes; False otherwise.
|
||||
:param vm_id: globally unique identifier for VM instance
|
||||
:param host_id: ID of the host where the VM is placed
|
||||
:param port_id: globally unique port ID that connects VM to network
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
num_vm = (session.query(db_models.AristaProvisionedVms).
|
||||
filter_by(tenant_id=tenant_id,
|
||||
vm_id=vm_id,
|
||||
port_id=port_id,
|
||||
network_id=network_id,
|
||||
host_id=host_id).count())
|
||||
return num_vm > 0
|
||||
|
||||
|
||||
def is_port_provisioned(port_id, host_id=None):
|
||||
"""Checks if a port is already known to EOS
|
||||
|
||||
:returns: True, if yes; False otherwise.
|
||||
:param port_id: globally unique port ID that connects VM to network
|
||||
:param host_id: host to which the port is bound to
|
||||
"""
|
||||
|
||||
filters = {
|
||||
'port_id': port_id
|
||||
}
|
||||
if host_id:
|
||||
filters['host_id'] = host_id
|
||||
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
num_ports = (session.query(db_models.AristaProvisionedVms).
|
||||
filter_by(**filters).count())
|
||||
return num_ports > 0
|
||||
|
||||
|
||||
def is_network_provisioned(tenant_id, network_id, segmentation_id=None,
|
||||
segment_id=None):
|
||||
"""Checks if a networks is already known to EOS
|
||||
|
||||
:returns: True, if yes; False otherwise.
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
:param network_id: globally unique neutron network identifier
|
||||
:param segment_id: globally unique neutron network segment identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
filters = {'tenant_id': tenant_id,
|
||||
'network_id': network_id}
|
||||
if segmentation_id:
|
||||
filters['segmentation_id'] = segmentation_id
|
||||
if segment_id:
|
||||
filters['id'] = segment_id
|
||||
|
||||
num_nets = (session.query(db_models.AristaProvisionedNets).
|
||||
filter_by(**filters).count())
|
||||
|
||||
return num_nets > 0
|
||||
|
||||
|
||||
def is_tenant_provisioned(tenant_id):
|
||||
"""Checks if a tenant is already known to EOS
|
||||
|
||||
:returns: True, if yes; False otherwise.
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
num_tenants = (session.query(db_models.AristaProvisionedTenants).
|
||||
filter_by(tenant_id=tenant_id).count())
|
||||
return num_tenants > 0
|
||||
|
||||
|
||||
def num_nets_provisioned(tenant_id):
|
||||
"""Returns number of networks for a given tennat.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
return (session.query(db_models.AristaProvisionedNets).
|
||||
filter_by(tenant_id=tenant_id).count())
|
||||
|
||||
|
||||
def num_vms_provisioned(tenant_id):
|
||||
"""Returns number of VMs for a given tennat.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
return (session.query(db_models.AristaProvisionedVms).
|
||||
filter_by(tenant_id=tenant_id).count())
|
||||
|
||||
|
||||
def get_networks(tenant_id):
|
||||
"""Returns all networks for a given tenant in EOS-compatible format.
|
||||
|
||||
See AristaRPCWrapper.get_network_list() for return value format.
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
model = db_models.AristaProvisionedNets
|
||||
# hack for pep8 E711: comparison to None should be
|
||||
# 'if cond is not None'
|
||||
none = None
|
||||
all_nets = []
|
||||
if tenant_id != 'any':
|
||||
all_nets = (session.query(model).
|
||||
filter(model.tenant_id == tenant_id,
|
||||
model.segmentation_id != none))
|
||||
else:
|
||||
all_nets = (session.query(model).
|
||||
filter(model.segmentation_id != none))
|
||||
|
||||
res = dict(
|
||||
(net.network_id, net.eos_network_representation(
|
||||
VLAN_SEGMENTATION))
|
||||
for net in all_nets
|
||||
)
|
||||
return res
|
||||
|
||||
|
||||
def get_vms(tenant_id):
|
||||
"""Returns all VMs for a given tenant in EOS-compatible format.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
model = db_models.AristaProvisionedVms
|
||||
# hack for pep8 E711: comparison to None should be
|
||||
# 'if cond is not None'
|
||||
none = None
|
||||
all_ports = (session.query(model).
|
||||
filter(model.tenant_id == tenant_id,
|
||||
model.host_id != none,
|
||||
model.vm_id != none,
|
||||
model.network_id != none,
|
||||
model.port_id != none))
|
||||
ports = {}
|
||||
for port in all_ports:
|
||||
if port.port_id not in ports:
|
||||
ports[port.port_id] = port.eos_port_representation()
|
||||
else:
|
||||
ports[port.port_id]['hosts'].append(port.host_id)
|
||||
|
||||
vm_dict = dict()
|
||||
|
||||
def eos_vm_representation(port):
|
||||
return {u'vmId': port['deviceId'],
|
||||
u'baremetal_instance': False,
|
||||
u'ports': [port]}
|
||||
|
||||
for port in ports.values():
|
||||
deviceId = port['deviceId']
|
||||
if deviceId in vm_dict:
|
||||
vm_dict[deviceId]['ports'].append(port)
|
||||
else:
|
||||
vm_dict[deviceId] = eos_vm_representation(port)
|
||||
return vm_dict
|
||||
|
||||
|
||||
def are_ports_attached_to_network(net_id):
|
||||
"""Checks if a given network is used by any port, excluding dhcp port.
|
||||
|
||||
:param net_id: globally unique network ID
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
model = db_models.AristaProvisionedVms
|
||||
return session.query(model).filter_by(network_id=net_id).filter(
|
||||
~model.vm_id.startswith('dhcp')).count() > 0
|
||||
|
||||
|
||||
def get_ports(tenant_id=None):
|
||||
"""Returns all ports of VMs in EOS-compatible format.
|
||||
|
||||
:param tenant_id: globally unique neutron tenant identifier
|
||||
"""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
model = db_models.AristaProvisionedVms
|
||||
# hack for pep8 E711: comparison to None should be
|
||||
# 'if cond is not None'
|
||||
none = None
|
||||
if tenant_id:
|
||||
all_ports = (session.query(model).
|
||||
filter(model.tenant_id == tenant_id,
|
||||
model.host_id != none,
|
||||
model.vm_id != none,
|
||||
model.network_id != none,
|
||||
model.port_id != none))
|
||||
else:
|
||||
all_ports = (session.query(model).
|
||||
filter(model.tenant_id != none,
|
||||
model.host_id != none,
|
||||
model.vm_id != none,
|
||||
model.network_id != none,
|
||||
model.port_id != none))
|
||||
ports = {}
|
||||
for port in all_ports:
|
||||
if port.port_id not in ports:
|
||||
ports[port.port_id] = port.eos_port_representation()
|
||||
ports[port.port_id]['hosts'].append(port.host_id)
|
||||
|
||||
return ports
|
||||
|
||||
|
||||
def get_tenants():
|
||||
"""Returns list of all tenants in EOS-compatible format."""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
model = db_models.AristaProvisionedTenants
|
||||
all_tenants = session.query(model)
|
||||
res = dict(
|
||||
(tenant.tenant_id, tenant.eos_tenant_representation())
|
||||
for tenant in all_tenants
|
||||
)
|
||||
return res
|
||||
|
||||
|
||||
def _make_port_dict(record):
|
||||
"""Make a dict from the BM profile DB record."""
|
||||
return {'port_id': record.port_id,
|
||||
'host_id': record.host,
|
||||
'vnic_type': record.vnic_type,
|
||||
'profile': record.profile}
|
||||
|
||||
|
||||
def get_all_baremetal_ports():
|
||||
"""Returns a list of all ports that belong to baremetal hosts."""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
querry = session.query(ml2_models.PortBinding)
|
||||
bm_ports = querry.filter_by(vnic_type='baremetal').all()
|
||||
|
||||
return {bm_port.port_id: _make_port_dict(bm_port)
|
||||
for bm_port in bm_ports}
|
||||
|
||||
|
||||
def get_all_portbindings():
|
||||
"""Returns a list of all ports bindings."""
|
||||
session = db.get_session()
|
||||
with session.begin():
|
||||
query = session.query(ml2_models.PortBinding)
|
||||
ports = query.all()
|
||||
|
||||
return {port.port_id: _make_port_dict(port)
|
||||
for port in ports}
|
||||
|
||||
|
||||
def get_port_binding_level(filters):
|
||||
"""Returns entries from PortBindingLevel based on the specified filters."""
|
||||
session = db.get_reader_session()
|
||||
with session.begin():
|
||||
return (session.query(ml2_models.PortBindingLevel).
|
||||
filter_by(**filters).all())
|
||||
|
||||
|
||||
class NeutronNets(db_base_plugin_v2.NeutronDbPluginV2,
|
||||
sec_db.SecurityGroupDbMixin):
|
||||
"""Access to Neutron DB.
|
||||
|
||||
Provides access to the Neutron Data bases for all provisioned
|
||||
networks as well ports. This data is used during the synchronization
|
||||
of DB between ML2 Mechanism Driver and Arista EOS
|
||||
Names of the networks and ports are not stroed in Arista repository
|
||||
They are pulled from Neutron DB.
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
self.admin_ctx = nctx.get_admin_context()
|
||||
|
||||
def get_network_name(self, tenant_id, network_id):
|
||||
network = self._get_network(tenant_id, network_id)
|
||||
network_name = None
|
||||
if network:
|
||||
network_name = network[0]['name']
|
||||
return network_name
|
||||
|
||||
def get_all_networks_for_tenant(self, tenant_id):
|
||||
filters = {'tenant_id': [tenant_id]}
|
||||
return super(NeutronNets,
|
||||
self).get_networks(self.admin_ctx, filters=filters) or []
|
||||
|
||||
def get_all_networks(self):
|
||||
return super(NeutronNets, self).get_networks(self.admin_ctx) or []
|
||||
|
||||
def get_all_ports_for_tenant(self, tenant_id):
|
||||
filters = {'tenant_id': [tenant_id]}
|
||||
return super(NeutronNets,
|
||||
self).get_ports(self.admin_ctx, filters=filters) or []
|
||||
|
||||
def get_shared_network_owner_id(self, network_id):
|
||||
filters = {'id': [network_id]}
|
||||
nets = self.get_networks(self.admin_ctx, filters=filters) or []
|
||||
segments = segments_db.get_network_segments(self.admin_ctx,
|
||||
network_id)
|
||||
if not nets or not segments:
|
||||
return
|
||||
if (nets[0]['shared'] and
|
||||
segments[0][driver_api.NETWORK_TYPE] == n_const.TYPE_VLAN):
|
||||
return nets[0]['tenant_id']
|
||||
|
||||
def get_network_segments(self, network_id, dynamic=False, context=None):
|
||||
context = context if context is not None else self.admin_ctx
|
||||
segments = segments_db.get_network_segments(context, network_id,
|
||||
filter_dynamic=dynamic)
|
||||
if dynamic:
|
||||
for segment in segments:
|
||||
segment['is_dynamic'] = True
|
||||
return segments
|
||||
|
||||
def get_all_network_segments(self, network_id, context=None):
|
||||
segments = self.get_network_segments(network_id, context=context)
|
||||
segments += self.get_network_segments(network_id, dynamic=True,
|
||||
context=context)
|
||||
return segments
|
||||
|
||||
def get_segment_by_id(self, context, segment_id):
|
||||
return segments_db.get_segment_by_id(context,
|
||||
segment_id)
|
||||
|
||||
def get_network_from_net_id(self, network_id, context=None):
|
||||
filters = {'id': [network_id]}
|
||||
ctxt = context if context else self.admin_ctx
|
||||
return super(NeutronNets,
|
||||
self).get_networks(ctxt, filters=filters) or []
|
||||
|
||||
def _get_network(self, tenant_id, network_id):
|
||||
filters = {'tenant_id': [tenant_id],
|
||||
'id': [network_id]}
|
||||
return super(NeutronNets,
|
||||
self).get_networks(self.admin_ctx, filters=filters) or []
|
||||
|
||||
def get_subnet_info(self, subnet_id):
|
||||
return self.get_subnet(subnet_id)
|
||||
|
||||
def get_subnet_ip_version(self, subnet_id):
|
||||
subnet = self.get_subnet(subnet_id)
|
||||
return subnet['ip_version'] if 'ip_version' in subnet else None
|
||||
|
||||
def get_subnet_gateway_ip(self, subnet_id):
|
||||
subnet = self.get_subnet(subnet_id)
|
||||
return subnet['gateway_ip'] if 'gateway_ip' in subnet else None
|
||||
|
||||
def get_subnet_cidr(self, subnet_id):
|
||||
subnet = self.get_subnet(subnet_id)
|
||||
return subnet['cidr'] if 'cidr' in subnet else None
|
||||
|
||||
def get_network_id(self, subnet_id):
|
||||
subnet = self.get_subnet(subnet_id)
|
||||
return subnet['network_id'] if 'network_id' in subnet else None
|
||||
|
||||
def get_network_id_from_port_id(self, port_id):
|
||||
port = self.get_port(port_id)
|
||||
return port['network_id'] if 'network_id' in port else None
|
||||
|
||||
def get_subnet(self, subnet_id):
|
||||
return super(NeutronNets,
|
||||
self).get_subnet(self.admin_ctx, subnet_id) or {}
|
||||
|
||||
def get_port(self, port_id):
|
||||
return super(NeutronNets,
|
||||
self).get_port(self.admin_ctx, port_id) or {}
|
||||
|
||||
def get_all_security_gp_to_port_bindings(self):
|
||||
return super(NeutronNets, self)._get_port_security_group_bindings(
|
||||
self.admin_ctx) or []
|
||||
|
||||
def get_security_gp_to_port_bindings(self, sec_gp_id):
|
||||
filters = {'security_group_id': [sec_gp_id]}
|
||||
return super(NeutronNets, self)._get_port_security_group_bindings(
|
||||
self.admin_ctx, filters=filters) or []
|
||||
|
||||
def get_security_group(self, sec_gp_id):
|
||||
return super(NeutronNets,
|
||||
self).get_security_group(self.admin_ctx, sec_gp_id) or []
|
||||
|
||||
def get_security_groups(self):
|
||||
sgs = super(NeutronNets,
|
||||
self).get_security_groups(self.admin_ctx) or []
|
||||
sgs_all = {}
|
||||
if sgs:
|
||||
for s in sgs:
|
||||
sgs_all[s['id']] = s
|
||||
return sgs_all
|
||||
|
||||
def get_security_group_rule(self, sec_gpr_id):
|
||||
return super(NeutronNets,
|
||||
self).get_security_group_rule(self.admin_ctx,
|
||||
sec_gpr_id) or []
|
||||
|
||||
def validate_network_rbac_policy_change(self, resource, event, trigger,
|
||||
context, object_type, policy,
|
||||
**kwargs):
|
||||
return super(NeutronNets, self).validate_network_rbac_policy_change(
|
||||
resource, event, trigger, context, object_type, policy, kwargs)
|
@ -1,55 +0,0 @@
|
||||
# Copyright (c) 2013 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
"""Exceptions used by Arista ML2 Mechanism Driver."""
|
||||
|
||||
from neutron_lib import exceptions
|
||||
|
||||
from networking_arista._i18n import _
|
||||
|
||||
|
||||
class AristaRpcError(exceptions.NeutronException):
|
||||
message = _('%(msg)s')
|
||||
|
||||
|
||||
class AristaConfigError(exceptions.NeutronException):
|
||||
message = _('%(msg)s')
|
||||
|
||||
|
||||
class AristaServicePluginRpcError(exceptions.NeutronException):
|
||||
message = _('%(msg)s')
|
||||
|
||||
|
||||
class AristaServicePluginConfigError(exceptions.NeutronException):
|
||||
message = _('%(msg)s')
|
||||
|
||||
|
||||
class AristaSecurityGroupError(exceptions.NeutronException):
|
||||
message = _('%(msg)s')
|
||||
|
||||
|
||||
class VlanUnavailable(exceptions.NeutronException):
|
||||
"""An exception indicating VLAN creation failed because it's not available.
|
||||
|
||||
A specialization of the NeutronException indicating network creation failed
|
||||
because a specified VLAN is unavailable on the physical network.
|
||||
|
||||
:param vlan_id: The VLAN ID.
|
||||
:param physical_network: The physical network.
|
||||
"""
|
||||
message = _("Unable to create the network. "
|
||||
"The VLAN %(vlan_id)s on physical network "
|
||||
"%(physical_network)s is not available.")
|
@ -1 +0,0 @@
|
||||
Alembic database migration scripts for the networking-arista package.
|
@ -1 +0,0 @@
|
||||
Alembic database migration scripts for the networking-arista package.
|
@ -1 +0,0 @@
|
||||
Generic single-database configuration.
|
@ -1,123 +0,0 @@
|
||||
# Copyright (c) 2015 Arista Networks, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from logging.config import fileConfig
|
||||
|
||||
from alembic import context
|
||||
from neutron_lib.db import model_base
|
||||
from oslo_config import cfg
|
||||
from oslo_db.sqlalchemy import session
|
||||
import sqlalchemy as sa
|
||||
from sqlalchemy import event
|
||||
|
||||
from neutron.db.migration.alembic_migrations import external
|
||||
from neutron.db.migration.models import head # noqa
|
||||
|
||||
# this is the Alembic Config object, which provides
|
||||
# access to the values within the .ini file in use.
|
||||
config = context.config
|
||||
neutron_config = config.neutron_config
|
||||
|
||||
# Interpret the config file for Python logging.
|
||||
# This line sets up loggers basically.
|
||||
fileConfig(config.config_file_name)
|
||||
|
||||
# add your model's MetaData object here
|
||||
# for 'autogenerate' support
|
||||
# from myapp import mymodel
|
||||
# target_metadata = mymodel.Base.metadata
|
||||
target_metadata = model_base.BASEV2.metadata
|
||||
|
||||
MYSQL_ENGINE = None
|
||||
ARISTA_VERSION_TABLE = 'arista_alembic_version'
|
||||
|
||||
|
||||
def set_mysql_engine():
|
||||
try:
|
||||
mysql_engine = neutron_config.command.mysql_engine
|
||||
except cfg.NoSuchOptError:
|
||||
mysql_engine = None
|
||||
|
||||
global MYSQL_ENGINE
|
||||
MYSQL_ENGINE = (mysql_engine or
|
||||
model_base.BASEV2.__table_args__['mysql_engine'])
|
||||
|
||||
|
||||
def include_object(object, name, type_, reflected, compare_to):
|
||||
if type_ == 'table' and name in external.TABLES:
|
||||
return False
|
||||
else:
|
||||
return True
|
||||
|
||||
|
||||
def run_migrations_offline():
|
||||
"""Run migrations in 'offline' mode.
|
||||
|
||||
This configures the context with just a URL or an Engine.
|
||||
|
||||
Calls to context.execute() here emit the given string to the
|
||||
script output.
|
||||
|
||||
"""
|
||||
set_mysql_engine()
|
||||
|
||||
kwargs = dict()
|
||||
if neutron_config.database.connection:
|
||||
kwargs['url'] = neutron_config.database.connection
|
||||
else:
|
||||
kwargs['dialect_name'] = neutron_config.database.engine
|
||||
kwargs['include_object'] = include_object
|
||||
kwargs['version_table'] = ARISTA_VERSION_TABLE
|
||||
context.configure(**kwargs)
|
||||
|
||||
with context.begin_transaction():
|
||||
context.run_migrations()
|
||||
|
||||
|
||||
@event.listens_for(sa.Table, 'after_parent_attach')
|
||||
def set_storage_engine(target, parent):
|
||||
if MYSQL_ENGINE:
|
||||
target.kwargs['mysql_engine'] = MYSQL_ENGINE
|
||||
|
||||
|
||||
def run_migrations_online():
|
||||
"""Run migrations in 'online' mode.
|
||||
|
||||
In this scenario we need to create an Engine
|
||||
and associate a connection with the context.
|
||||
|
||||
"""
|
||||
set_mysql_engine()
|
||||
engine = session.create_engine(neutron_config.database.connection)
|
||||
|
||||
connection = engine.connect()
|
||||
context.configure(
|
||||
connection=connection,
|
||||
target_metadata=target_metadata,
|
||||
include_object=include_object,
|
||||
version_table=ARISTA_VERSION_TABLE,
|
||||
)
|
||||
|
||||
try:
|
||||
with context.begin_transaction():
|
||||
context.run_migrations()
|
||||
finally:
|
||||
connection.close()
|
||||
engine.dispose()
|
||||
|
||||
|
||||
if context.is_offline_mode():
|
||||
run_migrations_offline()
|
||||
else:
|
||||
run_migrations_online()
|
@ -1,36 +0,0 @@
|
||||
# Copyright ${create_date.year} Arista Networks, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
#
|
||||
|
||||
"""${message}
|
||||
|
||||
Revision ID: ${up_revision}
|
||||
Revises: ${down_revision | comma,n}
|
||||
Create Date: ${create_date}
|
||||
|
||||
"""
|
||||
|
||||
# revision identifiers, used by Alembic.
|
||||
revision = ${repr(up_revision)}
|
||||
down_revision = ${repr(down_revision)}
|
||||
branch_labels = ${repr(branch_labels)}
|
||||
depends_on = ${repr(depends_on)}
|
||||
|
||||
from alembic import op
|
||||
import sqlalchemy as sa
|
||||
${imports if imports else ""}
|
||||
|
||||
|
||||
def upgrade():
|
||||
${upgrades if upgrades else "pass"}
|
@ -1,29 +0,0 @@
|
||||
# Copyright (c) 2015 Arista Networks, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
#
|
||||
|
||||
"""Initial db version
|
||||
|
||||
Revision ID: 296b4e0236e0
|
||||
Create Date: 2015-10-23 14:37:49.594974
|
||||
|
||||
"""
|
||||
|
||||
# revision identifiers, used by Alembic.
|
||||
revision = '296b4e0236e0'
|
||||
down_revision = None
|
||||
|
||||
|
||||
def upgrade():
|
||||
pass
|
@ -1,32 +0,0 @@
|
||||
# Copyright (c) 2015 Arista Networks, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
#
|
||||
|
||||
"""Initial db version
|
||||
|
||||
Revision ID: 47036dc8697a
|
||||
Create Date: 2015-10-23 14:37:49.594974
|
||||
|
||||
"""
|
||||
|
||||
from neutron.db.migration import cli
|
||||
|
||||
# revision identifiers, used by Alembic.
|
||||
revision = '47036dc8697a'
|
||||
down_revision = '296b4e0236e0'
|
||||
branch_labels = (cli.CONTRACT_BRANCH,)
|
||||
|
||||
|
||||
def upgrade():
|
||||
pass
|
@ -1,32 +0,0 @@
|
||||
# Copyright (c) 2015 Arista Networks, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
#
|
||||
|
||||
"""Initial db version
|
||||
|
||||
Revision ID: 1c6993ce7db0
|
||||
Create Date: 2015-10-23 14:37:49.594974
|
||||
|
||||
"""
|
||||
|
||||
from neutron.db.migration import cli
|
||||
|
||||
# revision identifiers, used by Alembic.
|
||||
revision = '1c6993ce7db0'
|
||||
down_revision = '296b4e0236e0'
|
||||
branch_labels = (cli.EXPAND_BRANCH,)
|
||||
|
||||
|
||||
def upgrade():
|
||||
pass
|
@ -1,457 +0,0 @@
|
||||
# Copyright 2014 Arista Networks, Inc. All rights reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import hashlib
|
||||
import socket
|
||||
import struct
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log as logging
|
||||
|
||||
from networking_arista._i18n import _, _LI
|
||||
from networking_arista.common import api
|
||||
from networking_arista.common import exceptions as arista_exc
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
cfg.CONF.import_group('l3_arista', 'networking_arista.common.config')
|
||||
|
||||
EOS_UNREACHABLE_MSG = _('Unable to reach EOS')
|
||||
DEFAULT_VLAN = 1
|
||||
MLAG_SWITCHES = 2
|
||||
VIRTUAL_ROUTER_MAC = '00:11:22:33:44:55'
|
||||
IPV4_BITS = 32
|
||||
IPV6_BITS = 128
|
||||
|
||||
# This string-format-at-a-distance confuses pylint :(
|
||||
# pylint: disable=too-many-format-args
|
||||
router_in_vrf = {
|
||||
'router': {'create': ['vrf definition {0}',
|
||||
'rd {1}',
|
||||
'exit'],
|
||||
'delete': ['no vrf definition {0}']},
|
||||
|
||||
'interface': {'add': ['ip routing vrf {1}',
|
||||
'vlan {0}',
|
||||
'exit',
|
||||
'interface vlan {0}',
|
||||
'vrf forwarding {1}',
|
||||
'ip address {2}'],
|
||||
'remove': ['no interface vlan {0}']}}
|
||||
|
||||
router_in_default_vrf = {
|
||||
'router': {'create': [], # Place holder for now.
|
||||
'delete': []}, # Place holder for now.
|
||||
|
||||
'interface': {'add': ['ip routing',
|
||||
'vlan {0}',
|
||||
'exit',
|
||||
'interface vlan {0}',
|
||||
'ip address {2}'],
|
||||
'remove': ['no interface vlan {0}']}}
|
||||
|
||||
router_in_default_vrf_v6 = {
|
||||
'router': {'create': [],
|
||||
'delete': []},
|
||||
|
||||
'interface': {'add': ['ipv6 unicast-routing',
|
||||
'vlan {0}',
|
||||
'exit',
|
||||
'interface vlan {0}',
|
||||
'ipv6 enable',
|
||||
'ipv6 address {2}'],
|
||||
'remove': ['no interface vlan {0}']}}
|
||||
|
||||
additional_cmds_for_mlag = {
|
||||
'router': {'create': ['ip virtual-router mac-address {0}'],
|
||||
'delete': []},
|
||||
|
||||
'interface': {'add': ['ip virtual-router address {0}'],
|
||||
'remove': []}}
|
||||
|
||||
additional_cmds_for_mlag_v6 = {
|
||||
'router': {'create': [],
|
||||
'delete': []},
|
||||
|
||||
'interface': {'add': ['ipv6 virtual-router address {0}'],
|
||||
'remove': []}}
|
||||
|
||||
|
||||
class AristaL3Driver(object):
|
||||
"""Wraps Arista JSON RPC.
|
||||
|
||||
All communications between Neutron and EOS are over JSON RPC.
|
||||
EOS - operating system used on Arista hardware
|
||||
Command API - JSON RPC API provided by Arista EOS
|
||||
"""
|
||||
def __init__(self):
|
||||
self._servers = []
|
||||
self._hosts = []
|
||||
self._interfaceDict = None
|
||||
self._validate_config()
|
||||
host = cfg.CONF.l3_arista.primary_l3_host
|
||||
self._hosts.append(host)
|
||||
self._servers.append(self._make_eapi_client(host))
|
||||
self._mlag_configured = cfg.CONF.l3_arista.mlag_config
|
||||
self._use_vrf = cfg.CONF.l3_arista.use_vrf
|
||||
if self._mlag_configured:
|
||||
host = cfg.CONF.l3_arista.secondary_l3_host
|
||||
self._hosts.append(host)
|
||||
self._servers.append(self._make_eapi_client(host))
|
||||
self._additionalRouterCmdsDict = additional_cmds_for_mlag['router']
|
||||
self._additionalInterfaceCmdsDict = (
|
||||
additional_cmds_for_mlag['interface'])
|
||||
if self._use_vrf:
|
||||
self.routerDict = router_in_vrf['router']
|
||||
self._interfaceDict = router_in_vrf['interface']
|
||||
else:
|
||||
self.routerDict = router_in_default_vrf['router']
|
||||
self._interfaceDict = router_in_default_vrf['interface']
|
||||
|
||||
@staticmethod
|
||||
def _make_eapi_client(host):
|
||||
return api.EAPIClient(
|
||||
host,
|
||||
username=cfg.CONF.l3_arista.primary_l3_host_username,
|
||||
password=cfg.CONF.l3_arista.primary_l3_host_password,
|
||||
verify=False,
|
||||
timeout=cfg.CONF.l3_arista.conn_timeout
|
||||
)
|
||||
|
||||
def _validate_config(self):
|
||||
if cfg.CONF.l3_arista.get('primary_l3_host') == '':
|
||||
msg = _('Required option primary_l3_host is not set')
|
||||
LOG.error(msg)
|
||||
raise arista_exc.AristaServicePluginConfigError(msg=msg)
|
||||
if cfg.CONF.l3_arista.get('mlag_config'):
|
||||
if cfg.CONF.l3_arista.get('use_vrf'):
|
||||
# This is invalid/unsupported configuration
|
||||
msg = _('VRFs are not supported MLAG config mode')
|
||||
LOG.error(msg)
|
||||
raise arista_exc.AristaServicePluginConfigError(msg=msg)
|
||||
if cfg.CONF.l3_arista.get('secondary_l3_host') == '':
|
||||
msg = _('Required option secondary_l3_host is not set')
|
||||
LOG.error(msg)
|
||||
raise arista_exc.AristaServicePluginConfigError(msg=msg)
|
||||
if cfg.CONF.l3_arista.get('primary_l3_host_username') == '':
|
||||
msg = _('Required option primary_l3_host_username is not set')
|
||||
LOG.error(msg)
|
||||
raise arista_exc.AristaServicePluginConfigError(msg=msg)
|
||||
|
||||
def create_router_on_eos(self, router_name, rdm, server):
|
||||
"""Creates a router on Arista HW Device.
|
||||
|
||||
:param router_name: globally unique identifier for router/VRF
|
||||
:param rdm: A value generated by hashing router name
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
rd = "%s:%s" % (rdm, rdm)
|
||||
|
||||
for c in self.routerDict['create']:
|
||||
cmds.append(c.format(router_name, rd))
|
||||
|
||||
if self._mlag_configured:
|
||||
mac = VIRTUAL_ROUTER_MAC
|
||||
for c in self._additionalRouterCmdsDict['create']:
|
||||
cmds.append(c.format(mac))
|
||||
|
||||
self._run_openstack_l3_cmds(cmds, server)
|
||||
|
||||
def delete_router_from_eos(self, router_name, server):
|
||||
"""Deletes a router from Arista HW Device.
|
||||
|
||||
:param router_name: globally unique identifier for router/VRF
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
for c in self.routerDict['delete']:
|
||||
cmds.append(c.format(router_name))
|
||||
if self._mlag_configured:
|
||||
for c in self._additionalRouterCmdsDict['delete']:
|
||||
cmds.append(c)
|
||||
|
||||
self._run_openstack_l3_cmds(cmds, server)
|
||||
|
||||
def _select_dicts(self, ipv):
|
||||
if self._use_vrf:
|
||||
self._interfaceDict = router_in_vrf['interface']
|
||||
else:
|
||||
if ipv == 6:
|
||||
# for IPv6 use IPv6 commmands
|
||||
self._interfaceDict = router_in_default_vrf_v6['interface']
|
||||
self._additionalInterfaceCmdsDict = (
|
||||
additional_cmds_for_mlag_v6['interface'])
|
||||
else:
|
||||
self._interfaceDict = router_in_default_vrf['interface']
|
||||
self._additionalInterfaceCmdsDict = (
|
||||
additional_cmds_for_mlag['interface'])
|
||||
|
||||
def add_interface_to_router(self, segment_id,
|
||||
router_name, gip, router_ip, mask, server):
|
||||
"""Adds an interface to existing HW router on Arista HW device.
|
||||
|
||||
:param segment_id: VLAN Id associated with interface that is added
|
||||
:param router_name: globally unique identifier for router/VRF
|
||||
:param gip: Gateway IP associated with the subnet
|
||||
:param router_ip: IP address of the router
|
||||
:param mask: subnet mask to be used
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
|
||||
if not segment_id:
|
||||
segment_id = DEFAULT_VLAN
|
||||
cmds = []
|
||||
for c in self._interfaceDict['add']:
|
||||
if self._mlag_configured:
|
||||
# In VARP config, use router ID else, use gateway IP address.
|
||||
ip = router_ip
|
||||
else:
|
||||
ip = gip + '/' + mask
|
||||
cmds.append(c.format(segment_id, router_name, ip))
|
||||
if self._mlag_configured:
|
||||
for c in self._additionalInterfaceCmdsDict['add']:
|
||||
cmds.append(c.format(gip))
|
||||
|
||||
self._run_openstack_l3_cmds(cmds, server)
|
||||
|
||||
def delete_interface_from_router(self, segment_id, router_name, server):
|
||||
"""Deletes an interface from existing HW router on Arista HW device.
|
||||
|
||||
:param segment_id: VLAN Id associated with interface that is added
|
||||
:param router_name: globally unique identifier for router/VRF
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
|
||||
if not segment_id:
|
||||
segment_id = DEFAULT_VLAN
|
||||
cmds = []
|
||||
for c in self._interfaceDict['remove']:
|
||||
cmds.append(c.format(segment_id))
|
||||
|
||||
self._run_openstack_l3_cmds(cmds, server)
|
||||
|
||||
def create_router(self, context, tenant_id, router):
|
||||
"""Creates a router on Arista Switch.
|
||||
|
||||
Deals with multiple configurations - such as Router per VRF,
|
||||
a router in default VRF, Virtual Router in MLAG configurations
|
||||
"""
|
||||
if router:
|
||||
router_name = self._arista_router_name(tenant_id, router['name'])
|
||||
|
||||
hashed = hashlib.sha256(router_name.encode('utf-8'))
|
||||
rdm = str(int(hashed.hexdigest(), 16) % 65536)
|
||||
|
||||
mlag_peer_failed = False
|
||||
for s in self._servers:
|
||||
try:
|
||||
self.create_router_on_eos(router_name, rdm, s)
|
||||
mlag_peer_failed = False
|
||||
except Exception:
|
||||
if self._mlag_configured and not mlag_peer_failed:
|
||||
# In paied switch, it is OK to fail on one switch
|
||||
mlag_peer_failed = True
|
||||
else:
|
||||
msg = (_('Failed to create router %s on EOS') %
|
||||
router_name)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(msg=msg)
|
||||
|
||||
def delete_router(self, context, tenant_id, router_id, router):
|
||||
"""Deletes a router from Arista Switch."""
|
||||
|
||||
if router:
|
||||
router_name = self._arista_router_name(tenant_id, router['name'])
|
||||
mlag_peer_failed = False
|
||||
for s in self._servers:
|
||||
try:
|
||||
self.delete_router_from_eos(router_name, s)
|
||||
mlag_peer_failed = False
|
||||
except Exception:
|
||||
if self._mlag_configured and not mlag_peer_failed:
|
||||
# In paied switch, it is OK to fail on one switch
|
||||
mlag_peer_failed = True
|
||||
else:
|
||||
msg = (_('Failed to create router %s on EOS') %
|
||||
router_name)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(msg=msg)
|
||||
|
||||
def update_router(self, context, router_id, original_router, new_router):
|
||||
"""Updates a router which is already created on Arista Switch.
|
||||
|
||||
TODO: (Sukhdev) - to be implemented in next release.
|
||||
"""
|
||||
pass
|
||||
|
||||
def add_router_interface(self, context, router_info):
|
||||
"""Adds an interface to a router created on Arista HW router.
|
||||
|
||||
This deals with both IPv6 and IPv4 configurations.
|
||||
"""
|
||||
if router_info:
|
||||
self._select_dicts(router_info['ip_version'])
|
||||
cidr = router_info['cidr']
|
||||
subnet_mask = cidr.split('/')[1]
|
||||
router_name = self._arista_router_name(router_info['tenant_id'],
|
||||
router_info['name'])
|
||||
if self._mlag_configured:
|
||||
# For MLAG, we send a specific IP address as opposed to cidr
|
||||
# For now, we are using x.x.x.253 and x.x.x.254 as virtual IP
|
||||
mlag_peer_failed = False
|
||||
for i, server in enumerate(self._servers):
|
||||
# Get appropriate virtual IP address for this router
|
||||
router_ip = self._get_router_ip(cidr, i,
|
||||
router_info['ip_version'])
|
||||
try:
|
||||
self.add_interface_to_router(router_info['seg_id'],
|
||||
router_name,
|
||||
router_info['gip'],
|
||||
router_ip, subnet_mask,
|
||||
server)
|
||||
mlag_peer_failed = False
|
||||
except Exception:
|
||||
if not mlag_peer_failed:
|
||||
mlag_peer_failed = True
|
||||
else:
|
||||
msg = (_('Failed to add interface to router '
|
||||
'%s on EOS') % router_name)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(
|
||||
msg=msg)
|
||||
|
||||
else:
|
||||
for s in self._servers:
|
||||
self.add_interface_to_router(router_info['seg_id'],
|
||||
router_name,
|
||||
router_info['gip'],
|
||||
None, subnet_mask, s)
|
||||
|
||||
def remove_router_interface(self, context, router_info):
|
||||
"""Removes previously configured interface from router on Arista HW.
|
||||
|
||||
This deals with both IPv6 and IPv4 configurations.
|
||||
"""
|
||||
if router_info:
|
||||
router_name = self._arista_router_name(router_info['tenant_id'],
|
||||
router_info['name'])
|
||||
mlag_peer_failed = False
|
||||
for s in self._servers:
|
||||
try:
|
||||
self.delete_interface_from_router(router_info['seg_id'],
|
||||
router_name, s)
|
||||
if self._mlag_configured:
|
||||
mlag_peer_failed = False
|
||||
except Exception:
|
||||
if self._mlag_configured and not mlag_peer_failed:
|
||||
mlag_peer_failed = True
|
||||
else:
|
||||
msg = (_('Failed to add interface to router '
|
||||
'%s on EOS') % router_name)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(msg=msg)
|
||||
|
||||
def _run_openstack_l3_cmds(self, commands, server):
|
||||
"""Execute/sends a CAPI (Command API) command to EOS.
|
||||
|
||||
In this method, list of commands is appended with prefix and
|
||||
postfix commands - to make is understandble by EOS.
|
||||
|
||||
:param commands : List of command to be executed on EOS.
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
command_start = ['enable', 'configure']
|
||||
command_end = ['exit']
|
||||
full_command = command_start + commands + command_end
|
||||
|
||||
LOG.info(_LI('Executing command on Arista EOS: %s'), full_command)
|
||||
|
||||
try:
|
||||
# this returns array of return values for every command in
|
||||
# full_command list
|
||||
ret = server.execute(full_command)
|
||||
LOG.info(_LI('Results of execution on Arista EOS: %s'), ret)
|
||||
|
||||
except Exception:
|
||||
msg = (_('Error occurred while trying to execute '
|
||||
'commands %(cmd)s on EOS %(host)s') %
|
||||
{'cmd': full_command, 'host': server})
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(msg=msg)
|
||||
|
||||
def _arista_router_name(self, tenant_id, name):
|
||||
"""Generate an arista specific name for this router.
|
||||
|
||||
Use a unique name so that OpenStack created routers/SVIs
|
||||
can be distinguishged from the user created routers/SVIs
|
||||
on Arista HW.
|
||||
"""
|
||||
return 'OS' + '-' + tenant_id + '-' + name
|
||||
|
||||
def _get_binary_from_ipv4(self, ip_addr):
|
||||
"""Converts IPv4 address to binary form."""
|
||||
|
||||
return struct.unpack("!L", socket.inet_pton(socket.AF_INET,
|
||||
ip_addr))[0]
|
||||
|
||||
def _get_binary_from_ipv6(self, ip_addr):
|
||||
"""Converts IPv6 address to binary form."""
|
||||
|
||||
hi, lo = struct.unpack("!QQ", socket.inet_pton(socket.AF_INET6,
|
||||
ip_addr))
|
||||
return (hi << 64) | lo
|
||||
|
||||
def _get_ipv4_from_binary(self, bin_addr):
|
||||
"""Converts binary address to Ipv4 format."""
|
||||
|
||||
return socket.inet_ntop(socket.AF_INET, struct.pack("!L", bin_addr))
|
||||
|
||||
def _get_ipv6_from_binary(self, bin_addr):
|
||||
"""Converts binary address to Ipv6 format."""
|
||||
|
||||
hi = bin_addr >> 64
|
||||
lo = bin_addr & 0xFFFFFFFF
|
||||
return socket.inet_ntop(socket.AF_INET6, struct.pack("!QQ", hi, lo))
|
||||
|
||||
def _get_router_ip(self, cidr, ip_count, ip_ver):
|
||||
"""For a given IP subnet and IP version type, generate IP for router.
|
||||
|
||||
This method takes the network address (cidr) and selects an
|
||||
IP address that should be assigned to virtual router running
|
||||
on multiple switches. It uses upper addresses in a subnet address
|
||||
as IP for the router. Each instace of the router, on each switch,
|
||||
requires uniqe IP address. For example in IPv4 case, on a 255
|
||||
subnet, it will pick X.X.X.254 as first addess, X.X.X.253 for next,
|
||||
and so on.
|
||||
"""
|
||||
start_ip = MLAG_SWITCHES + ip_count
|
||||
network_addr, prefix = cidr.split('/')
|
||||
if ip_ver == 4:
|
||||
bits = IPV4_BITS
|
||||
ip = self._get_binary_from_ipv4(network_addr)
|
||||
elif ip_ver == 6:
|
||||
bits = IPV6_BITS
|
||||
ip = self._get_binary_from_ipv6(network_addr)
|
||||
|
||||
mask = (pow(2, bits) - 1) << (bits - int(prefix))
|
||||
|
||||
network_addr = ip & mask
|
||||
|
||||
router_ip = pow(2, bits - int(prefix)) - start_ip
|
||||
|
||||
router_ip = network_addr | router_ip
|
||||
if ip_ver == 4:
|
||||
return self._get_ipv4_from_binary(router_ip) + '/' + prefix
|
||||
else:
|
||||
return self._get_ipv6_from_binary(router_ip) + '/' + prefix
|
@ -1,277 +0,0 @@
|
||||
# Copyright 2014 Arista Networks, Inc. All rights reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import copy
|
||||
import threading
|
||||
|
||||
from neutron_lib import constants as n_const
|
||||
from neutron_lib import context as nctx
|
||||
from neutron_lib.plugins import constants as plugin_constants
|
||||
from oslo_config import cfg
|
||||
from oslo_log import helpers as log_helpers
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import excutils
|
||||
|
||||
from neutron.api.rpc.agentnotifiers import l3_rpc_agent_api
|
||||
from neutron.api.rpc.handlers import l3_rpc
|
||||
from neutron.common import rpc as n_rpc
|
||||
from neutron.common import topics
|
||||
from neutron.db import db_base_plugin_v2
|
||||
from neutron.db import extraroute_db
|
||||
from neutron.db import l3_agentschedulers_db
|
||||
from neutron.db import l3_gwmode_db
|
||||
from neutron.plugins.ml2.driver_context import NetworkContext # noqa
|
||||
|
||||
from networking_arista._i18n import _LE, _LI
|
||||
from networking_arista.common import db_lib
|
||||
from networking_arista.l3Plugin import arista_l3_driver
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class AristaL3ServicePlugin(db_base_plugin_v2.NeutronDbPluginV2,
|
||||
extraroute_db.ExtraRoute_db_mixin,
|
||||
l3_gwmode_db.L3_NAT_db_mixin,
|
||||
l3_agentschedulers_db.L3AgentSchedulerDbMixin):
|
||||
|
||||
"""Implements L3 Router service plugin for Arista hardware.
|
||||
|
||||
Creates routers in Arista hardware, manages them, adds/deletes interfaces
|
||||
to the routes.
|
||||
"""
|
||||
|
||||
supported_extension_aliases = ["router", "ext-gw-mode",
|
||||
"extraroute"]
|
||||
|
||||
def __init__(self, driver=None):
|
||||
|
||||
self.driver = driver or arista_l3_driver.AristaL3Driver()
|
||||
self.ndb = db_lib.NeutronNets()
|
||||
self.setup_rpc()
|
||||
self.sync_timeout = cfg.CONF.l3_arista.l3_sync_interval
|
||||
self.sync_lock = threading.Lock()
|
||||
self._synchronization_thread()
|
||||
|
||||
def setup_rpc(self):
|
||||
# RPC support
|
||||
self.topic = topics.L3PLUGIN
|
||||
self.conn = n_rpc.create_connection()
|
||||
self.agent_notifiers.update(
|
||||
{n_const.AGENT_TYPE_L3: l3_rpc_agent_api.L3AgentNotifyAPI()})
|
||||
self.endpoints = [l3_rpc.L3RpcCallback()]
|
||||
self.conn.create_consumer(self.topic, self.endpoints,
|
||||
fanout=False)
|
||||
self.conn.consume_in_threads()
|
||||
|
||||
def get_plugin_type(self):
|
||||
return plugin_constants.L3
|
||||
|
||||
def get_plugin_description(self):
|
||||
"""Returns string description of the plugin."""
|
||||
return ("Arista L3 Router Service Plugin for Arista Hardware "
|
||||
"based routing")
|
||||
|
||||
def _synchronization_thread(self):
|
||||
with self.sync_lock:
|
||||
self.synchronize()
|
||||
|
||||
self.timer = threading.Timer(self.sync_timeout,
|
||||
self._synchronization_thread)
|
||||
self.timer.start()
|
||||
|
||||
def stop_synchronization_thread(self):
|
||||
if self.timer:
|
||||
self.timer.cancel()
|
||||
self.timer = None
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def create_router(self, context, router):
|
||||
"""Create a new router entry in DB, and create it Arista HW."""
|
||||
|
||||
tenant_id = router['router']['tenant_id']
|
||||
|
||||
# Add router to the DB
|
||||
new_router = super(AristaL3ServicePlugin, self).create_router(
|
||||
context,
|
||||
router)
|
||||
# create router on the Arista Hw
|
||||
try:
|
||||
self.driver.create_router(context, tenant_id, new_router)
|
||||
return new_router
|
||||
except Exception:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.error(_LE("Error creating router on Arista HW router=%s "),
|
||||
new_router)
|
||||
super(AristaL3ServicePlugin, self).delete_router(
|
||||
context, new_router['id'])
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def update_router(self, context, router_id, router):
|
||||
"""Update an existing router in DB, and update it in Arista HW."""
|
||||
|
||||
# Read existing router record from DB
|
||||
original_router = super(AristaL3ServicePlugin, self).get_router(
|
||||
context, router_id)
|
||||
# Update router DB
|
||||
new_router = super(AristaL3ServicePlugin, self).update_router(
|
||||
context, router_id, router)
|
||||
|
||||
# Modify router on the Arista Hw
|
||||
try:
|
||||
self.driver.update_router(context, router_id,
|
||||
original_router, new_router)
|
||||
return new_router
|
||||
except Exception:
|
||||
LOG.error(_LE("Error updating router on Arista HW router=%s "),
|
||||
new_router)
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def delete_router(self, context, router_id):
|
||||
"""Delete an existing router from Arista HW as well as from the DB."""
|
||||
|
||||
router = super(AristaL3ServicePlugin, self).get_router(context,
|
||||
router_id)
|
||||
tenant_id = router['tenant_id']
|
||||
|
||||
# Delete router on the Arista Hw
|
||||
try:
|
||||
self.driver.delete_router(context, tenant_id, router_id, router)
|
||||
except Exception as e:
|
||||
LOG.error(_LE("Error deleting router on Arista HW "
|
||||
"router %(r)s exception=%(e)s"),
|
||||
{'r': router, 'e': e})
|
||||
|
||||
super(AristaL3ServicePlugin, self).delete_router(context, router_id)
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def add_router_interface(self, context, router_id, interface_info):
|
||||
"""Add a subnet of a network to an existing router."""
|
||||
|
||||
new_router = super(AristaL3ServicePlugin, self).add_router_interface(
|
||||
context, router_id, interface_info)
|
||||
|
||||
# Get network info for the subnet that is being added to the router.
|
||||
# Check if the interface information is by port-id or subnet-id
|
||||
add_by_port, add_by_sub = self._validate_interface_info(interface_info)
|
||||
if add_by_sub:
|
||||
subnet = self.get_subnet(context, interface_info['subnet_id'])
|
||||
elif add_by_port:
|
||||
port = self.get_port(context, interface_info['port_id'])
|
||||
subnet_id = port['fixed_ips'][0]['subnet_id']
|
||||
subnet = self.get_subnet(context, subnet_id)
|
||||
network_id = subnet['network_id']
|
||||
|
||||
# To create SVI's in Arista HW, the segmentation Id is required
|
||||
# for this network.
|
||||
ml2_db = NetworkContext(self, context, {'id': network_id})
|
||||
seg_id = ml2_db.network_segments[0]['segmentation_id']
|
||||
|
||||
# Package all the info needed for Hw programming
|
||||
router = super(AristaL3ServicePlugin, self).get_router(context,
|
||||
router_id)
|
||||
router_info = copy.deepcopy(new_router)
|
||||
router_info['seg_id'] = seg_id
|
||||
router_info['name'] = router['name']
|
||||
router_info['cidr'] = subnet['cidr']
|
||||
router_info['gip'] = subnet['gateway_ip']
|
||||
router_info['ip_version'] = subnet['ip_version']
|
||||
|
||||
try:
|
||||
self.driver.add_router_interface(context, router_info)
|
||||
return new_router
|
||||
except Exception:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.error(_LE("Error Adding subnet %(subnet)s to "
|
||||
"router %(router_id)s on Arista HW"),
|
||||
{'subnet': subnet, 'router_id': router_id})
|
||||
super(AristaL3ServicePlugin, self).remove_router_interface(
|
||||
context,
|
||||
router_id,
|
||||
interface_info)
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def remove_router_interface(self, context, router_id, interface_info):
|
||||
"""Remove a subnet of a network from an existing router."""
|
||||
|
||||
new_router = (
|
||||
super(AristaL3ServicePlugin, self).remove_router_interface(
|
||||
context, router_id, interface_info))
|
||||
|
||||
# Get network information of the subnet that is being removed
|
||||
subnet = self.get_subnet(context, new_router['subnet_id'])
|
||||
network_id = subnet['network_id']
|
||||
|
||||
# For SVI removal from Arista HW, segmentation ID is needed
|
||||
ml2_db = NetworkContext(self, context, {'id': network_id})
|
||||
seg_id = ml2_db.network_segments[0]['segmentation_id']
|
||||
|
||||
router = super(AristaL3ServicePlugin, self).get_router(context,
|
||||
router_id)
|
||||
router_info = copy.deepcopy(new_router)
|
||||
router_info['seg_id'] = seg_id
|
||||
router_info['name'] = router['name']
|
||||
|
||||
try:
|
||||
self.driver.remove_router_interface(context, router_info)
|
||||
return new_router
|
||||
except Exception as exc:
|
||||
LOG.error(_LE("Error removing interface %(interface)s from "
|
||||
"router %(router_id)s on Arista HW"
|
||||
"Exception =(exc)s"),
|
||||
{'interface': interface_info, 'router_id': router_id,
|
||||
'exc': exc})
|
||||
|
||||
def synchronize(self):
|
||||
"""Synchronizes Router DB from Neturon DB with EOS.
|
||||
|
||||
Walks through the Neturon Db and ensures that all the routers
|
||||
created in Netuton DB match with EOS. After creating appropriate
|
||||
routers, it ensures to add interfaces as well.
|
||||
Uses idempotent properties of EOS configuration, which means
|
||||
same commands can be repeated.
|
||||
"""
|
||||
LOG.info(_LI('Syncing Neutron Router DB <-> EOS'))
|
||||
ctx = nctx.get_admin_context()
|
||||
|
||||
routers = super(AristaL3ServicePlugin, self).get_routers(ctx)
|
||||
for r in routers:
|
||||
tenant_id = r['tenant_id']
|
||||
ports = self.ndb.get_all_ports_for_tenant(tenant_id)
|
||||
|
||||
try:
|
||||
self.driver.create_router(self, tenant_id, r)
|
||||
|
||||
except Exception:
|
||||
continue
|
||||
|
||||
# Figure out which interfaces are added to this router
|
||||
for p in ports:
|
||||
if p['device_id'] == r['id']:
|
||||
net_id = p['network_id']
|
||||
subnet_id = p['fixed_ips'][0]['subnet_id']
|
||||
subnet = self.ndb.get_subnet_info(subnet_id)
|
||||
ml2_db = NetworkContext(self, ctx, {'id': net_id})
|
||||
seg_id = ml2_db.network_segments[0]['segmentation_id']
|
||||
|
||||
r['seg_id'] = seg_id
|
||||
r['cidr'] = subnet['cidr']
|
||||
r['gip'] = subnet['gateway_ip']
|
||||
r['ip_version'] = subnet['ip_version']
|
||||
|
||||
try:
|
||||
self.driver.add_router_interface(self, r)
|
||||
except Exception:
|
||||
LOG.error(_LE("Error Adding interface %(subnet_id)s "
|
||||
"to router %(router_id)s on Arista HW"),
|
||||
{'subnet_id': subnet_id, 'router_id': r})
|
File diff suppressed because it is too large
Load Diff
@ -1,603 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log as logging
|
||||
|
||||
from networking_arista._i18n import _, _LI
|
||||
from networking_arista.common import api
|
||||
from networking_arista.common import db_lib
|
||||
from networking_arista.common import exceptions as arista_exc
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
EOS_UNREACHABLE_MSG = _('Unable to reach EOS')
|
||||
|
||||
# Note 'None,null' means default rule - i.e. deny everything
|
||||
SUPPORTED_SG_PROTOCOLS = [None, 'tcp', 'udp', 'icmp']
|
||||
|
||||
acl_cmd = {
|
||||
'acl': {'create': ['ip access-list {0}'],
|
||||
'in_rule': ['permit {0} {1} any range {2} {3}'],
|
||||
'out_rule': ['permit {0} any {1} range {2} {3}'],
|
||||
'in_icmp_custom1': ['permit icmp {0} any {1}'],
|
||||
'out_icmp_custom1': ['permit icmp any {0} {1}'],
|
||||
'in_icmp_custom2': ['permit icmp {0} any {1} {2}'],
|
||||
'out_icmp_custom2': ['permit icmp any {0} {1} {2}'],
|
||||
'default': [],
|
||||
'delete_acl': ['no ip access-list {0}'],
|
||||
'del_in_icmp_custom1': ['ip access-list {0}',
|
||||
'no permit icmp {1} any {2}',
|
||||
'exit'],
|
||||
'del_out_icmp_custom1': ['ip access-list {0}',
|
||||
'no permit icmp any {1} {2}',
|
||||
'exit'],
|
||||
'del_in_icmp_custom2': ['ip access-list {0}',
|
||||
'no permit icmp {1} any {2} {3}',
|
||||
'exit'],
|
||||
'del_out_icmp_custom2': ['ip access-list {0}',
|
||||
'no permit icmp any {1} {2} {3}',
|
||||
'exit'],
|
||||
'del_in_acl_rule': ['ip access-list {0}',
|
||||
'no permit {1} {2} any range {3} {4}',
|
||||
'exit'],
|
||||
'del_out_acl_rule': ['ip access-list {0}',
|
||||
'no permit {1} any {2} range {3} {4}',
|
||||
'exit']},
|
||||
|
||||
'apply': {'ingress': ['interface {0}',
|
||||
'ip access-group {1} in',
|
||||
'exit'],
|
||||
'egress': ['interface {0}',
|
||||
'ip access-group {1} out',
|
||||
'exit'],
|
||||
'rm_ingress': ['interface {0}',
|
||||
'no ip access-group {1} in',
|
||||
'exit'],
|
||||
'rm_egress': ['interface {0}',
|
||||
'no ip access-group {1} out',
|
||||
'exit']}}
|
||||
|
||||
|
||||
class AristaSecGroupSwitchDriver(object):
|
||||
"""Wraps Arista JSON RPC.
|
||||
|
||||
All communications between Neutron and EOS are over JSON RPC.
|
||||
EOS - operating system used on Arista hardware
|
||||
Command API - JSON RPC API provided by Arista EOS
|
||||
"""
|
||||
def __init__(self, neutron_db):
|
||||
self._ndb = neutron_db
|
||||
self._servers = []
|
||||
self._hosts = {}
|
||||
self.sg_enabled = cfg.CONF.ml2_arista.get('sec_group_support')
|
||||
self._validate_config()
|
||||
for s in cfg.CONF.ml2_arista.switch_info:
|
||||
switch_ip, switch_user, switch_pass = s.split(":")
|
||||
if switch_pass == "''":
|
||||
switch_pass = ''
|
||||
self._hosts[switch_ip] = (
|
||||
{'user': switch_user, 'password': switch_pass})
|
||||
self._servers.append(self._make_eapi_client(switch_ip))
|
||||
self.aclCreateDict = acl_cmd['acl']
|
||||
self.aclApplyDict = acl_cmd['apply']
|
||||
|
||||
def _make_eapi_client(self, host):
|
||||
return api.EAPIClient(
|
||||
host,
|
||||
username=self._hosts[host]['user'],
|
||||
password=self._hosts[host]['password'],
|
||||
verify=False,
|
||||
timeout=cfg.CONF.ml2_arista.conn_timeout
|
||||
)
|
||||
|
||||
def _validate_config(self):
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
if len(cfg.CONF.ml2_arista.get('switch_info')) < 1:
|
||||
msg = _('Required option - when "sec_group_support" is enabled, '
|
||||
'at least one switch must be specified ')
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaConfigError(msg=msg)
|
||||
|
||||
def _create_acl_on_eos(self, in_cmds, out_cmds, protocol, cidr,
|
||||
from_port, to_port, direction):
|
||||
"""Creates an ACL on Arista HW Device.
|
||||
|
||||
:param name: Name for the ACL
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
if protocol == 'icmp':
|
||||
# ICMP rules require special processing
|
||||
if ((from_port and to_port) or
|
||||
(not from_port and not to_port)):
|
||||
rule = 'icmp_custom2'
|
||||
elif from_port and not to_port:
|
||||
rule = 'icmp_custom1'
|
||||
else:
|
||||
msg = _('Invalid ICMP rule specified')
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
rule_type = 'in'
|
||||
cmds = in_cmds
|
||||
if direction == 'egress':
|
||||
rule_type = 'out'
|
||||
cmds = out_cmds
|
||||
final_rule = rule_type + '_' + rule
|
||||
acl_dict = self.aclCreateDict[final_rule]
|
||||
|
||||
# None port is probematic - should be replaced with 0
|
||||
if not from_port:
|
||||
from_port = 0
|
||||
if not to_port:
|
||||
to_port = 0
|
||||
|
||||
for c in acl_dict:
|
||||
if rule == 'icmp_custom2':
|
||||
cmds.append(c.format(cidr, from_port, to_port))
|
||||
else:
|
||||
cmds.append(c.format(cidr, from_port))
|
||||
return in_cmds, out_cmds
|
||||
else:
|
||||
# Non ICMP rules processing here
|
||||
acl_dict = self.aclCreateDict['in_rule']
|
||||
cmds = in_cmds
|
||||
if direction == 'egress':
|
||||
acl_dict = self.aclCreateDict['out_rule']
|
||||
cmds = out_cmds
|
||||
if not protocol:
|
||||
acl_dict = self.aclCreateDict['default']
|
||||
|
||||
for c in acl_dict:
|
||||
cmds.append(c.format(protocol, cidr,
|
||||
from_port, to_port))
|
||||
return in_cmds, out_cmds
|
||||
|
||||
def _delete_acl_from_eos(self, name, server):
|
||||
"""deletes an ACL from Arista HW Device.
|
||||
|
||||
:param name: Name for the ACL
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
|
||||
for c in self.aclCreateDict['delete_acl']:
|
||||
cmds.append(c.format(name))
|
||||
|
||||
self._run_openstack_sg_cmds(cmds, server)
|
||||
|
||||
def _delete_acl_rule_from_eos(self, name,
|
||||
protocol, cidr,
|
||||
from_port, to_port,
|
||||
direction, server):
|
||||
"""deletes an ACL from Arista HW Device.
|
||||
|
||||
:param name: Name for the ACL
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
|
||||
if protocol == 'icmp':
|
||||
# ICMP rules require special processing
|
||||
if ((from_port and to_port) or
|
||||
(not from_port and not to_port)):
|
||||
rule = 'icmp_custom2'
|
||||
elif from_port and not to_port:
|
||||
rule = 'icmp_custom1'
|
||||
else:
|
||||
msg = _('Invalid ICMP rule specified')
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
rule_type = 'del_in'
|
||||
if direction == 'egress':
|
||||
rule_type = 'del_out'
|
||||
final_rule = rule_type + '_' + rule
|
||||
acl_dict = self.aclCreateDict[final_rule]
|
||||
|
||||
# None port is probematic - should be replaced with 0
|
||||
if not from_port:
|
||||
from_port = 0
|
||||
if not to_port:
|
||||
to_port = 0
|
||||
|
||||
for c in acl_dict:
|
||||
if rule == 'icmp_custom2':
|
||||
cmds.append(c.format(name, cidr, from_port, to_port))
|
||||
else:
|
||||
cmds.append(c.format(name, cidr, from_port))
|
||||
|
||||
else:
|
||||
acl_dict = self.aclCreateDict['del_in_acl_rule']
|
||||
if direction == 'egress':
|
||||
acl_dict = self.aclCreateDict['del_out_acl_rule']
|
||||
|
||||
for c in acl_dict:
|
||||
cmds.append(c.format(name, protocol, cidr,
|
||||
from_port, to_port))
|
||||
|
||||
self._run_openstack_sg_cmds(cmds, server)
|
||||
|
||||
def _apply_acl_on_eos(self, port_id, name, direction, server):
|
||||
"""Creates an ACL on Arista HW Device.
|
||||
|
||||
:param port_id: The port where the ACL needs to be applied
|
||||
:param name: Name for the ACL
|
||||
:param direction: must contain "ingress" or "egress"
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
|
||||
for c in self.aclApplyDict[direction]:
|
||||
cmds.append(c.format(port_id, name))
|
||||
|
||||
self._run_openstack_sg_cmds(cmds, server)
|
||||
|
||||
def _remove_acl_from_eos(self, port_id, name, direction, server):
|
||||
"""Remove an ACL from a port on Arista HW Device.
|
||||
|
||||
:param port_id: The port where the ACL needs to be applied
|
||||
:param name: Name for the ACL
|
||||
:param direction: must contain "ingress" or "egress"
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
cmds = []
|
||||
|
||||
acl_cmd = self.aclApplyDict['rm_ingress']
|
||||
if direction == 'egress':
|
||||
acl_cmd = self.aclApplyDict['rm_egress']
|
||||
for c in acl_cmd:
|
||||
cmds.append(c.format(port_id, name))
|
||||
|
||||
self._run_openstack_sg_cmds(cmds, server)
|
||||
|
||||
def _create_acl_rule(self, in_cmds, out_cmds, sgr):
|
||||
"""Creates an ACL on Arista Switch.
|
||||
|
||||
For a given Security Group (ACL), it adds additional rule
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Only deal with valid protocols - skip the rest
|
||||
if not sgr or sgr['protocol'] not in SUPPORTED_SG_PROTOCOLS:
|
||||
return in_cmds, out_cmds
|
||||
|
||||
remote_ip = sgr['remote_ip_prefix']
|
||||
if not remote_ip:
|
||||
remote_ip = 'any'
|
||||
min_port = sgr['port_range_min']
|
||||
if not min_port:
|
||||
min_port = 0
|
||||
max_port = sgr['port_range_max']
|
||||
if not max_port and sgr['protocol'] != 'icmp':
|
||||
max_port = 65535
|
||||
in_cmds, out_cmds = self._create_acl_on_eos(in_cmds, out_cmds,
|
||||
sgr['protocol'],
|
||||
remote_ip,
|
||||
min_port,
|
||||
max_port,
|
||||
sgr['direction'])
|
||||
return in_cmds, out_cmds
|
||||
|
||||
def create_acl_rule(self, sgr):
|
||||
"""Creates an ACL on Arista Switch.
|
||||
|
||||
For a given Security Group (ACL), it adds additional rule
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
name = self._arista_acl_name(sgr['security_group_id'],
|
||||
sgr['direction'])
|
||||
cmds = []
|
||||
for c in self.aclCreateDict['create']:
|
||||
cmds.append(c.format(name))
|
||||
in_cmds, out_cmds = self._create_acl_rule(cmds, cmds, sgr)
|
||||
|
||||
cmds = in_cmds
|
||||
if sgr['direction'] == 'egress':
|
||||
cmds = out_cmds
|
||||
|
||||
cmds.append('exit')
|
||||
|
||||
for s in self._servers:
|
||||
try:
|
||||
self._run_openstack_sg_cmds(cmds, s)
|
||||
except Exception:
|
||||
msg = (_('Failed to create ACL rule on EOS %s') % s)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def delete_acl_rule(self, sgr):
|
||||
"""Deletes an ACL rule on Arista Switch.
|
||||
|
||||
For a given Security Group (ACL), it adds removes a rule
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
# Only deal with valid protocols - skip the rest
|
||||
if not sgr or sgr['protocol'] not in SUPPORTED_SG_PROTOCOLS:
|
||||
return
|
||||
|
||||
# Build seperate ACL for ingress and egress
|
||||
name = self._arista_acl_name(sgr['security_group_id'],
|
||||
sgr['direction'])
|
||||
remote_ip = sgr['remote_ip_prefix']
|
||||
if not remote_ip:
|
||||
remote_ip = 'any'
|
||||
min_port = sgr['port_range_min']
|
||||
if not min_port:
|
||||
min_port = 0
|
||||
max_port = sgr['port_range_max']
|
||||
if not max_port and sgr['protocol'] != 'icmp':
|
||||
max_port = 65535
|
||||
for s in self._servers:
|
||||
try:
|
||||
self._delete_acl_rule_from_eos(name,
|
||||
sgr['protocol'],
|
||||
remote_ip,
|
||||
min_port,
|
||||
max_port,
|
||||
sgr['direction'],
|
||||
s)
|
||||
except Exception:
|
||||
msg = (_('Failed to delete ACL on EOS %s') % s)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def _create_acl_shell(self, sg_id):
|
||||
"""Creates an ACL on Arista Switch.
|
||||
|
||||
For a given Security Group (ACL), it adds additional rule
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Build seperate ACL for ingress and egress
|
||||
direction = ['ingress', 'egress']
|
||||
cmds = []
|
||||
for d in range(len(direction)):
|
||||
name = self._arista_acl_name(sg_id, direction[d])
|
||||
cmds.append([])
|
||||
for c in self.aclCreateDict['create']:
|
||||
cmds[d].append(c.format(name))
|
||||
return cmds[0], cmds[1]
|
||||
|
||||
def create_acl(self, sg):
|
||||
"""Creates an ACL on Arista Switch.
|
||||
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
if not sg:
|
||||
msg = _('Invalid or Empty Security Group Specified')
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
in_cmds, out_cmds = self._create_acl_shell(sg['id'])
|
||||
for sgr in sg['security_group_rules']:
|
||||
in_cmds, out_cmds = self._create_acl_rule(in_cmds, out_cmds, sgr)
|
||||
in_cmds.append('exit')
|
||||
out_cmds.append('exit')
|
||||
|
||||
for s in self._servers:
|
||||
try:
|
||||
self._run_openstack_sg_cmds(in_cmds, s)
|
||||
self._run_openstack_sg_cmds(out_cmds, s)
|
||||
|
||||
except Exception:
|
||||
msg = (_('Failed to create ACL on EOS %s') % s)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def delete_acl(self, sg):
|
||||
"""Deletes an ACL from Arista Switch.
|
||||
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
if not sg:
|
||||
msg = _('Invalid or Empty Security Group Specified')
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
direction = ['ingress', 'egress']
|
||||
for d in range(len(direction)):
|
||||
name = self._arista_acl_name(sg['id'], direction[d])
|
||||
|
||||
for s in self._servers:
|
||||
try:
|
||||
self._delete_acl_from_eos(name, s)
|
||||
except Exception:
|
||||
msg = (_('Failed to create ACL on EOS %s') % s)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def apply_acl(self, sgs, switch_id, port_id, switch_info):
|
||||
"""Creates an ACL on Arista Switch.
|
||||
|
||||
Applies ACLs to the baremetal ports only. The port/switch
|
||||
details is passed through the parameters.
|
||||
Deals with multiple configurations - such as multiple switches
|
||||
param sgs: List of Security Groups
|
||||
param switch_id: Switch ID of TOR where ACL needs to be applied
|
||||
param port_id: Port ID of port where ACL needs to be applied
|
||||
param switch_info: IP address of the TOR
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
# We do not support more than one security group on a port
|
||||
if not sgs or len(sgs) > 1:
|
||||
msg = (_('Only one Security Group Supported on a port %s') % sgs)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
sg = self._ndb.get_security_group(sgs[0])
|
||||
|
||||
# We already have ACLs on the TORs.
|
||||
# Here we need to find out which ACL is applicable - i.e.
|
||||
# Ingress ACL, egress ACL or both
|
||||
direction = ['ingress', 'egress']
|
||||
|
||||
server = self._make_eapi_client(switch_info)
|
||||
|
||||
for d in range(len(direction)):
|
||||
name = self._arista_acl_name(sg['id'], direction[d])
|
||||
try:
|
||||
self._apply_acl_on_eos(port_id, name, direction[d], server)
|
||||
except Exception:
|
||||
msg = (_('Failed to apply ACL on port %s') % port_id)
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def remove_acl(self, sgs, switch_id, port_id, switch_info):
|
||||
"""Removes an ACL from Arista Switch.
|
||||
|
||||
Removes ACLs from the baremetal ports only. The port/switch
|
||||
details is passed throuhg the parameters.
|
||||
param sgs: List of Security Groups
|
||||
param switch_id: Switch ID of TOR where ACL needs to be removed
|
||||
param port_id: Port ID of port where ACL needs to be removed
|
||||
param switch_info: IP address of the TOR
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
# We do not support more than one security group on a port
|
||||
if not sgs or len(sgs) > 1:
|
||||
msg = (_('Only one Security Group Supported on a port %s') % sgs)
|
||||
raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
sg = self._ndb.get_security_group(sgs[0])
|
||||
|
||||
# We already have ACLs on the TORs.
|
||||
# Here we need to find out which ACL is applicable - i.e.
|
||||
# Ingress ACL, egress ACL or both
|
||||
direction = []
|
||||
for sgr in sg['security_group_rules']:
|
||||
# Only deal with valid protocols - skip the rest
|
||||
if not sgr or sgr['protocol'] not in SUPPORTED_SG_PROTOCOLS:
|
||||
continue
|
||||
|
||||
if sgr['direction'] not in direction:
|
||||
direction.append(sgr['direction'])
|
||||
|
||||
# THIS IS TOTAL HACK NOW - just for testing
|
||||
# Assumes the credential of all switches are same as specified
|
||||
# in the condig file
|
||||
server = self._make_eapi_client(switch_info)
|
||||
for d in range(len(direction)):
|
||||
name = self._arista_acl_name(sg['id'], direction[d])
|
||||
try:
|
||||
self._remove_acl_from_eos(port_id, name, direction[d], server)
|
||||
except Exception:
|
||||
msg = (_('Failed to remove ACL on port %s') % port_id)
|
||||
LOG.exception(msg)
|
||||
# No need to raise exception for ACL removal
|
||||
# raise arista_exc.AristaSecurityGroupError(msg=msg)
|
||||
|
||||
def _run_openstack_sg_cmds(self, commands, server):
|
||||
"""Execute/sends a CAPI (Command API) command to EOS.
|
||||
|
||||
In this method, list of commands is appended with prefix and
|
||||
postfix commands - to make is understandble by EOS.
|
||||
|
||||
:param commands : List of command to be executed on EOS.
|
||||
:param server: Server endpoint on the Arista switch to be configured
|
||||
"""
|
||||
command_start = ['enable', 'configure']
|
||||
command_end = ['exit']
|
||||
full_command = command_start + commands + command_end
|
||||
|
||||
LOG.info(_LI('Executing command on Arista EOS: %s'), full_command)
|
||||
|
||||
try:
|
||||
# this returns array of return values for every command in
|
||||
# full_command list
|
||||
ret = server.execute(full_command)
|
||||
LOG.info(_LI('Results of execution on Arista EOS: %s'), ret)
|
||||
|
||||
except Exception:
|
||||
msg = (_('Error occurred while trying to execute '
|
||||
'commands %(cmd)s on EOS %(host)s') %
|
||||
{'cmd': full_command, 'host': server})
|
||||
LOG.exception(msg)
|
||||
raise arista_exc.AristaServicePluginRpcError(msg=msg)
|
||||
|
||||
def _arista_acl_name(self, name, direction):
|
||||
"""Generate an arista specific name for this ACL.
|
||||
|
||||
Use a unique name so that OpenStack created ACLs
|
||||
can be distinguishged from the user created ACLs
|
||||
on Arista HW.
|
||||
"""
|
||||
in_out = 'IN'
|
||||
if direction == 'egress':
|
||||
in_out = 'OUT'
|
||||
return 'SG' + '-' + in_out + '-' + name
|
||||
|
||||
def perform_sync_of_sg(self):
|
||||
"""Perform sync of the security groups between ML2 and EOS.
|
||||
|
||||
This is unconditional sync to ensure that all security
|
||||
ACLs are pushed to all the switches, in case of switch
|
||||
or neutron reboot
|
||||
"""
|
||||
# Do nothing if Security Groups are not enabled
|
||||
if not self.sg_enabled:
|
||||
return
|
||||
|
||||
arista_ports = db_lib.get_ports()
|
||||
neutron_sgs = self._ndb.get_security_groups()
|
||||
sg_bindings = self._ndb.get_all_security_gp_to_port_bindings()
|
||||
sgs = []
|
||||
sgs_dict = {}
|
||||
arista_port_ids = arista_ports.keys()
|
||||
|
||||
# Get the list of Security Groups of interest to us
|
||||
for s in sg_bindings:
|
||||
if s['port_id'] in arista_port_ids:
|
||||
if not s['security_group_id'] in sgs:
|
||||
sgs_dict[s['port_id']] = (
|
||||
{'security_group_id': s['security_group_id']})
|
||||
sgs.append(s['security_group_id'])
|
||||
|
||||
# Create the ACLs on Arista Switches
|
||||
for idx in range(len(sgs)):
|
||||
self.create_acl(neutron_sgs[sgs[idx]])
|
||||
|
||||
# Get Baremetal port profiles, if any
|
||||
bm_port_profiles = db_lib.get_all_baremetal_ports()
|
||||
|
||||
if bm_port_profiles:
|
||||
for bm in bm_port_profiles.values():
|
||||
if bm['port_id'] in sgs_dict:
|
||||
sg = sgs_dict[bm['port_id']]['security_group_id']
|
||||
profile = json.loads(bm['profile'])
|
||||
link_info = profile['local_link_information']
|
||||
for l in link_info:
|
||||
if not l:
|
||||
# skip all empty entries
|
||||
continue
|
||||
self.apply_acl([sg], l['switch_id'],
|
||||
l['port_id'], l['switch_info'])
|
@ -1,148 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from neutron_lib.db import api as db_api
|
||||
from oslo_log import log
|
||||
from six import moves
|
||||
|
||||
from neutron.db.models.plugins.ml2 import vlanallocation
|
||||
|
||||
from networking_arista._i18n import _LI
|
||||
from networking_arista.common import exceptions as arista_exc
|
||||
from networking_arista.ml2.arista_ml2 import EOS_UNREACHABLE_MSG
|
||||
|
||||
LOG = log.getLogger(__name__)
|
||||
|
||||
|
||||
class VlanSyncService(object):
|
||||
"""Sync vlan assignment from CVX into the OpenStack db."""
|
||||
|
||||
def __init__(self, rpc_wrapper):
|
||||
self._rpc = rpc_wrapper
|
||||
self._force_sync = True
|
||||
self._vlan_assignment_uuid = None
|
||||
self._assigned_vlans = dict()
|
||||
|
||||
def force_sync(self):
|
||||
self._force_sync = True
|
||||
|
||||
def _parse_vlan_ranges(self, vlan_pool, return_as_ranges=False):
|
||||
vlan_ids = set()
|
||||
if return_as_ranges:
|
||||
vlan_ids = list()
|
||||
if not vlan_pool:
|
||||
return vlan_ids
|
||||
vlan_ranges = vlan_pool.split(',')
|
||||
for vlan_range in vlan_ranges:
|
||||
endpoints = vlan_range.split('-')
|
||||
if len(endpoints) == 2:
|
||||
vlan_min = int(endpoints[0])
|
||||
vlan_max = int(endpoints[1])
|
||||
if return_as_ranges:
|
||||
vlan_ids.append((vlan_min, vlan_max))
|
||||
else:
|
||||
vlan_ids |= set(moves.range(vlan_min, vlan_max + 1))
|
||||
elif len(endpoints) == 1:
|
||||
single_vlan = int(endpoints[0])
|
||||
if return_as_ranges:
|
||||
vlan_ids.append((single_vlan, single_vlan))
|
||||
else:
|
||||
vlan_ids.add(single_vlan)
|
||||
return vlan_ids
|
||||
|
||||
def get_network_vlan_ranges(self):
|
||||
return self._assigned_vlans
|
||||
|
||||
def _sync_required(self):
|
||||
try:
|
||||
if not self._force_sync and self._region_in_sync():
|
||||
LOG.info(_LI('VLANs are in sync!'))
|
||||
return False
|
||||
except arista_exc.AristaRpcError:
|
||||
LOG.warning(EOS_UNREACHABLE_MSG)
|
||||
self._force_sync = True
|
||||
return True
|
||||
|
||||
def _region_in_sync(self):
|
||||
eos_vlan_assignment_uuid = self._rpc.get_vlan_assignment_uuid()
|
||||
return (self._vlan_assignment_uuid and
|
||||
(self._vlan_assignment_uuid['uuid'] ==
|
||||
eos_vlan_assignment_uuid['uuid']))
|
||||
|
||||
def _set_vlan_assignment_uuid(self):
|
||||
try:
|
||||
self._vlan_assignment_uuid = self._rpc.get_vlan_assignment_uuid()
|
||||
except arista_exc.AristaRpcError:
|
||||
self._force_sync = True
|
||||
|
||||
def do_synchronize(self):
|
||||
if not self._sync_required():
|
||||
return
|
||||
|
||||
self.synchronize()
|
||||
self._set_vlan_assignment_uuid()
|
||||
|
||||
def synchronize(self):
|
||||
LOG.info(_LI('Syncing VLANs with EOS'))
|
||||
try:
|
||||
self._rpc.register_with_eos()
|
||||
vlan_pool = self._rpc.get_vlan_allocation()
|
||||
except arista_exc.AristaRpcError:
|
||||
LOG.warning(EOS_UNREACHABLE_MSG)
|
||||
self._force_sync = True
|
||||
return
|
||||
|
||||
self._assigned_vlans = {
|
||||
'default': self._parse_vlan_ranges(vlan_pool['assignedVlans'],
|
||||
return_as_ranges=True),
|
||||
}
|
||||
|
||||
assigned_vlans = (
|
||||
self._parse_vlan_ranges(vlan_pool['assignedVlans']))
|
||||
available_vlans = frozenset(
|
||||
self._parse_vlan_ranges(vlan_pool['availableVlans']))
|
||||
used_vlans = frozenset(
|
||||
self._parse_vlan_ranges(vlan_pool['allocatedVlans']))
|
||||
|
||||
self._force_sync = False
|
||||
|
||||
session = db_api.get_writer_session()
|
||||
with session.begin(subtransactions=True):
|
||||
allocs = (
|
||||
session.query(vlanallocation.VlanAllocation).with_lockmode(
|
||||
'update'))
|
||||
|
||||
for alloc in allocs:
|
||||
if alloc.physical_network != 'default':
|
||||
session.delete(alloc)
|
||||
|
||||
try:
|
||||
assigned_vlans.remove(alloc.vlan_id)
|
||||
except KeyError:
|
||||
session.delete(alloc)
|
||||
continue
|
||||
|
||||
if alloc.allocated and alloc.vlan_id in available_vlans:
|
||||
alloc.update({"allocated": False})
|
||||
elif not alloc.allocated and alloc.vlan_id in used_vlans:
|
||||
alloc.update({"allocated": True})
|
||||
|
||||
for vlan_id in sorted(assigned_vlans):
|
||||
allocated = vlan_id in used_vlans
|
||||
alloc = vlanallocation.VlanAllocation(
|
||||
physical_network='default',
|
||||
vlan_id=vlan_id,
|
||||
allocated=allocated)
|
||||
session.add(alloc)
|
@ -1,71 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import threading
|
||||
|
||||
from oslo_config import cfg
|
||||
from oslo_log import log
|
||||
|
||||
from neutron.plugins.ml2.drivers import type_vlan
|
||||
|
||||
from networking_arista._i18n import _LI
|
||||
from networking_arista.common import db_lib
|
||||
from networking_arista.common import exceptions as exc
|
||||
from networking_arista.ml2 import arista_ml2
|
||||
from networking_arista.ml2.drivers import driver_helpers
|
||||
|
||||
LOG = log.getLogger(__name__)
|
||||
cfg.CONF.import_group('arista_type_driver', 'networking_arista.common.config')
|
||||
|
||||
|
||||
class AristaVlanTypeDriver(type_vlan.VlanTypeDriver):
|
||||
"""Manage state for VLAN networks with ML2.
|
||||
|
||||
The VlanTypeDriver implements the 'vlan' network_type. VLAN
|
||||
network segments provide connectivity between VMs and other
|
||||
devices using any connected IEEE 802.1Q conformant
|
||||
physical_network segmented into virtual networks via IEEE 802.1Q
|
||||
headers. Up to 4094 VLAN network segments can exist on each
|
||||
available physical_network.
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
super(AristaVlanTypeDriver, self).__init__()
|
||||
ndb = db_lib.NeutronNets()
|
||||
self.rpc = arista_ml2.AristaRPCWrapperEapi(ndb)
|
||||
self.sync_service = driver_helpers.VlanSyncService(self.rpc)
|
||||
self.network_vlan_ranges = dict()
|
||||
self.sync_timeout = cfg.CONF.arista_type_driver['sync_interval']
|
||||
|
||||
def initialize(self):
|
||||
self.rpc.check_supported_features()
|
||||
self.rpc.check_vlan_type_driver_commands()
|
||||
self._synchronization_thread()
|
||||
LOG.info(_LI("AristaVlanTypeDriver initialization complete"))
|
||||
|
||||
def _synchronization_thread(self):
|
||||
self.sync_service.do_synchronize()
|
||||
self.network_vlan_ranges = self.sync_service.get_network_vlan_ranges()
|
||||
self.timer = threading.Timer(self.sync_timeout,
|
||||
self._synchronization_thread)
|
||||
self.timer.start()
|
||||
|
||||
def allocate_fully_specified_segment(self, session, **raw_segment):
|
||||
alloc = session.query(self.model).filter_by(**raw_segment).first()
|
||||
if not alloc:
|
||||
raise exc.VlanUnavailable(**raw_segment)
|
||||
return super(AristaVlanTypeDriver,
|
||||
self).allocate_fully_specified_segment(
|
||||
session, **raw_segment)
|
File diff suppressed because it is too large
Load Diff
@ -1,117 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
from neutron_lib.callbacks import events
|
||||
from neutron_lib.callbacks import registry
|
||||
from neutron_lib.callbacks import resources
|
||||
from oslo_log import helpers as log_helpers
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import excutils
|
||||
|
||||
from networking_arista._i18n import _LE
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class AristaSecurityGroupHandler(object):
|
||||
"""Security Group Handler for Arista networking hardware.
|
||||
|
||||
Registers for the notification of security group updates.
|
||||
Once a notification is recieved, it takes appropriate actions by updating
|
||||
Arista hardware appropriately.
|
||||
"""
|
||||
def __init__(self, client):
|
||||
self.client = client
|
||||
self.subscribe()
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def create_security_group(self, resource, event, trigger, **kwargs):
|
||||
sg = kwargs.get('security_group')
|
||||
try:
|
||||
self.client.create_security_group(sg)
|
||||
except Exception as e:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.error(_LE("Failed to create a security group %(sg_id)s "
|
||||
"in Arista Driver: %(err)s"),
|
||||
{"sg_id": sg["id"], "err": e})
|
||||
try:
|
||||
self.client.delete_security_group(sg)
|
||||
except Exception:
|
||||
LOG.exception(_LE("Failed to delete security group %s"),
|
||||
sg['id'])
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def delete_security_group(self, resource, event, trigger, **kwargs):
|
||||
sg = kwargs.get('security_group')
|
||||
try:
|
||||
self.client.delete_security_group(sg)
|
||||
except Exception as e:
|
||||
LOG.error(_LE("Failed to delete security group %(sg_id)s "
|
||||
"in Arista Driver: %(err)s"),
|
||||
{"sg_id": sg["id"], "err": e})
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def update_security_group(self, resource, event, trigger, **kwargs):
|
||||
sg = kwargs.get('security_group')
|
||||
try:
|
||||
self.client.update_security_group(sg)
|
||||
except Exception as e:
|
||||
LOG.error(_LE("Failed to update security group %(sg_id)s "
|
||||
"in Arista Driver: %(err)s"),
|
||||
{"sg_id": sg["id"], "err": e})
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def create_security_group_rule(self, resource, event, trigger, **kwargs):
|
||||
sgr = kwargs.get('security_group_rule')
|
||||
try:
|
||||
self.client.create_security_group_rule(sgr)
|
||||
except Exception as e:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.error(_LE("Failed to create a security group %(sgr_id)s "
|
||||
"rule in Arista Driver: %(err)s"),
|
||||
{"sgr_id": sgr["id"], "err": e})
|
||||
try:
|
||||
self.client.delete_security_group_rule(sgr)
|
||||
except Exception:
|
||||
LOG.exception(_LE("Failed to delete security group "
|
||||
"rule %s"), sgr['id'])
|
||||
|
||||
@log_helpers.log_method_call
|
||||
def delete_security_group_rule(self, resource, event, trigger, **kwargs):
|
||||
sgr_id = kwargs.get('security_group_rule_id')
|
||||
try:
|
||||
self.client.delete_security_group_rule(sgr_id)
|
||||
except Exception as e:
|
||||
LOG.error(_LE("Failed to delete security group %(sgr_id)s "
|
||||
"rule in Arista Driver: %(err)s"),
|
||||
{"sgr_id": sgr_id, "err": e})
|
||||
|
||||
def subscribe(self):
|
||||
# Subscribe to the events related to security groups and rules
|
||||
registry.subscribe(
|
||||
self.create_security_group, resources.SECURITY_GROUP,
|
||||
events.AFTER_CREATE)
|
||||
registry.subscribe(
|
||||
self.update_security_group, resources.SECURITY_GROUP,
|
||||
events.AFTER_UPDATE)
|
||||
registry.subscribe(
|
||||
self.delete_security_group, resources.SECURITY_GROUP,
|
||||
events.BEFORE_DELETE)
|
||||
registry.subscribe(
|
||||
self.create_security_group_rule, resources.SECURITY_GROUP_RULE,
|
||||
events.AFTER_CREATE)
|
||||
registry.subscribe(
|
||||
self.delete_security_group_rule, resources.SECURITY_GROUP_RULE,
|
||||
events.BEFORE_DELETE)
|
@ -1,23 +0,0 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Copyright 2010-2011 OpenStack Foundation
|
||||
# Copyright (c) 2013 Hewlett-Packard Development Company, L.P.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslotest import base
|
||||
|
||||
|
||||
class TestCase(base.BaseTestCase):
|
||||
|
||||
"""Test case base class for all unit tests."""
|
@ -1,28 +0,0 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
test_networking_arista
|
||||
----------------------------------
|
||||
|
||||
Tests for `networking_arista` module.
|
||||
"""
|
||||
|
||||
from networking_arista.tests import base
|
||||
|
||||
|
||||
class TestNetworking_arista(base.TestCase):
|
||||
|
||||
def test_something(self):
|
||||
pass
|
@ -1,19 +0,0 @@
|
||||
# Copyright 2011 OpenStack Foundation.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslo_config import cfg
|
||||
|
||||
|
||||
cfg.CONF.use_stderr = False
|
@ -1,257 +0,0 @@
|
||||
# Copyright (c) 2017 Arista Networks, Inc
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import mock
|
||||
import requests
|
||||
from requests import exceptions as requests_exc
|
||||
import testtools
|
||||
|
||||
from networking_arista.common import api
|
||||
|
||||
|
||||
class TestEAPIClientInit(testtools.TestCase):
|
||||
def test_basic_init(self):
|
||||
host_ip = '10.20.30.40'
|
||||
client = api.EAPIClient(host_ip)
|
||||
self.assertEqual(client.host, host_ip)
|
||||
self.assertEqual(client.url, 'https://10.20.30.40/command-api')
|
||||
self.assertDictContainsSubset(
|
||||
{'Content-Type': 'application/json', 'Accept': 'application/json'},
|
||||
client.session.headers
|
||||
)
|
||||
|
||||
def test_init_enable_verify(self):
|
||||
client = api.EAPIClient('10.0.0.1', verify=True)
|
||||
self.assertTrue(client.session.verify)
|
||||
|
||||
def test_init_auth(self):
|
||||
client = api.EAPIClient('10.0.0.1', username='user', password='pass')
|
||||
self.assertEqual(client.session.auth, ('user', 'pass'))
|
||||
|
||||
def test_init_timeout(self):
|
||||
client = api.EAPIClient('10.0.0.1', timeout=99)
|
||||
self.assertEqual(client.timeout, 99)
|
||||
|
||||
def test_make_url(self):
|
||||
url = api.EAPIClient._make_url('1.2.3.4')
|
||||
self.assertEqual(url, 'https://1.2.3.4/command-api')
|
||||
|
||||
def test_make_url_http(self):
|
||||
url = api.EAPIClient._make_url('5.6.7.8', 'http')
|
||||
self.assertEqual(url, 'http://5.6.7.8/command-api')
|
||||
|
||||
|
||||
class TestEAPIClientExecute(testtools.TestCase):
|
||||
def setUp(self):
|
||||
super(TestEAPIClientExecute, self).setUp()
|
||||
|
||||
mock.patch('requests.Session.post').start()
|
||||
self.mock_log = mock.patch.object(api, 'LOG').start()
|
||||
self.mock_json_dumps = mock.patch.object(api.json, 'dumps').start()
|
||||
|
||||
self.addCleanup(mock.patch.stopall)
|
||||
|
||||
self.client = api.EAPIClient('10.0.0.1', timeout=99)
|
||||
|
||||
def _test_execute_helper(self, commands, commands_to_log=None):
|
||||
expected_data = {
|
||||
'id': 'Networking Arista Driver',
|
||||
'method': 'runCmds',
|
||||
'jsonrpc': '2.0',
|
||||
'params': {
|
||||
'timestamps': False,
|
||||
'format': 'json',
|
||||
'version': 1,
|
||||
'cmds': commands
|
||||
}
|
||||
}
|
||||
|
||||
self.client.session.post.assert_called_once_with(
|
||||
'https://10.0.0.1/command-api',
|
||||
data=self.mock_json_dumps.return_value,
|
||||
timeout=99
|
||||
)
|
||||
|
||||
self.mock_log.info.assert_has_calls(
|
||||
[
|
||||
mock.call(
|
||||
mock.ANY,
|
||||
{
|
||||
'ip': '10.0.0.1',
|
||||
'data': self.mock_json_dumps.return_value
|
||||
}
|
||||
)
|
||||
]
|
||||
)
|
||||
|
||||
log_data = dict(expected_data)
|
||||
log_data['params'] = dict(expected_data['params'])
|
||||
log_data['params']['cmds'] = commands_to_log or commands
|
||||
|
||||
self.mock_json_dumps.assert_has_calls(
|
||||
[
|
||||
mock.call(log_data),
|
||||
mock.call(expected_data)
|
||||
]
|
||||
)
|
||||
|
||||
def test_command_prep(self):
|
||||
commands = ['enable']
|
||||
self.client.execute(commands)
|
||||
self._test_execute_helper(commands)
|
||||
|
||||
def test_commands_to_log(self):
|
||||
commands = ['config', 'secret']
|
||||
commands_to_log = ['config', '******']
|
||||
self.client.execute(commands, commands_to_log)
|
||||
self._test_execute_helper(commands, commands_to_log)
|
||||
|
||||
def _test_execute_error_helper(self, raise_exception, expected_exception,
|
||||
warning_has_params=False):
|
||||
commands = ['config']
|
||||
|
||||
self.client.session.post.side_effect = raise_exception
|
||||
|
||||
self.assertRaises(
|
||||
expected_exception,
|
||||
self.client.execute,
|
||||
commands
|
||||
)
|
||||
|
||||
self._test_execute_helper(commands)
|
||||
|
||||
if warning_has_params:
|
||||
args = (mock.ANY, mock.ANY)
|
||||
else:
|
||||
args = (mock.ANY,)
|
||||
self.mock_log.warning.assert_called_once_with(*args)
|
||||
|
||||
def test_request_connection_error(self):
|
||||
self._test_execute_error_helper(
|
||||
requests_exc.ConnectionError,
|
||||
api.arista_exc.AristaRpcError
|
||||
)
|
||||
|
||||
def test_request_connect_timeout(self):
|
||||
self._test_execute_error_helper(
|
||||
requests_exc.ConnectTimeout,
|
||||
api.arista_exc.AristaRpcError
|
||||
)
|
||||
|
||||
def test_request_timeout(self):
|
||||
self._test_execute_error_helper(
|
||||
requests_exc.Timeout,
|
||||
api.arista_exc.AristaRpcError
|
||||
)
|
||||
|
||||
def test_request_connect_InvalidURL(self):
|
||||
self._test_execute_error_helper(
|
||||
requests_exc.InvalidURL,
|
||||
api.arista_exc.AristaRpcError
|
||||
)
|
||||
|
||||
def test_request_other_exception(self):
|
||||
class OtherException(Exception):
|
||||
pass
|
||||
|
||||
self._test_execute_error_helper(
|
||||
OtherException,
|
||||
OtherException,
|
||||
warning_has_params=True
|
||||
)
|
||||
|
||||
def _test_response_helper(self, response_data):
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.return_value = response_data
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
def test_response_success(self):
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.return_value = {'result': mock.sentinel}
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
retval = self.client.execute(['enable'])
|
||||
self.assertEqual(retval, mock.sentinel)
|
||||
|
||||
def test_response_json_error(self):
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.side_effect = ValueError
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
retval = self.client.execute(['enable'])
|
||||
self.assertIsNone(retval)
|
||||
self.mock_log.info.assert_has_calls([mock.call(mock.ANY)])
|
||||
|
||||
def _test_response_format_error_helper(self, bad_response):
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.return_value = bad_response
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
self.assertRaises(
|
||||
api.arista_exc.AristaRpcError,
|
||||
self.client.execute,
|
||||
['enable']
|
||||
)
|
||||
self.mock_log.info.assert_has_calls([mock.call(mock.ANY)])
|
||||
|
||||
def test_response_format_error(self):
|
||||
self._test_response_format_error_helper({})
|
||||
|
||||
def test_response_unknown_error_code(self):
|
||||
self._test_response_format_error_helper(
|
||||
{'error': {'code': 999}}
|
||||
)
|
||||
|
||||
def test_response_known_error_code(self):
|
||||
self._test_response_format_error_helper(
|
||||
{'error': {'code': 1002, 'data': []}}
|
||||
)
|
||||
|
||||
def test_response_known_error_code_data_is_not_dict(self):
|
||||
self._test_response_format_error_helper(
|
||||
{'error': {'code': 1002, 'data': ['some text']}}
|
||||
)
|
||||
|
||||
def test_response_not_cvx_leader(self):
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.return_value = {
|
||||
'error': {
|
||||
'code': 1002,
|
||||
'data': [{'errors': [api.ERR_CVX_NOT_LEADER]}]
|
||||
}
|
||||
}
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
retval = self.client.execute(['enable'])
|
||||
self.assertIsNone(retval)
|
||||
|
||||
def test_response_other_exception(self):
|
||||
class OtherException(Exception):
|
||||
pass
|
||||
|
||||
mock_response = mock.MagicMock(requests.Response)
|
||||
mock_response.json.return_value = 'text'
|
||||
self.client.session.post.return_value = mock_response
|
||||
|
||||
self.assertRaises(
|
||||
TypeError,
|
||||
self.client.execute,
|
||||
['enable']
|
||||
)
|
||||
self.mock_log.warning.assert_has_calls(
|
||||
[
|
||||
mock.call(mock.ANY, {'error': mock.ANY})
|
||||
]
|
||||
)
|
@ -1,19 +0,0 @@
|
||||
# Copyright 2011 OpenStack Foundation.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslo_config import cfg
|
||||
|
||||
|
||||
cfg.CONF.use_stderr = False
|
@ -1,456 +0,0 @@
|
||||
# Copyright (c) 2013 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
import mock
|
||||
from oslo_config import cfg
|
||||
|
||||
from neutron.tests import base
|
||||
|
||||
from networking_arista.l3Plugin import arista_l3_driver as arista
|
||||
|
||||
|
||||
def setup_arista_config(value='', vrf=False, mlag=False):
|
||||
cfg.CONF.set_override('primary_l3_host', value, "l3_arista")
|
||||
cfg.CONF.set_override('primary_l3_host_username', value, "l3_arista")
|
||||
if vrf:
|
||||
cfg.CONF.set_override('use_vrf', vrf, "l3_arista")
|
||||
if mlag:
|
||||
cfg.CONF.set_override('secondary_l3_host', value, "l3_arista")
|
||||
cfg.CONF.set_override('mlag_config', mlag, "l3_arista")
|
||||
|
||||
|
||||
class AristaL3DriverTestCasesDefaultVrf(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions in Default VRF.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCasesDefaultVrf, self).setUp()
|
||||
setup_arista_config('value')
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_create_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
route_domain = '123:123'
|
||||
|
||||
self.drv.create_router_on_eos(router_name, route_domain,
|
||||
self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_delete_router_from_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
|
||||
self.drv.delete_router_from_eos(router_name, self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_add_interface_to_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
router_ip = '10.10.10.10'
|
||||
gw_ip = '10.10.10.1'
|
||||
mask = '255.255.255.0'
|
||||
|
||||
self.drv.add_interface_to_router(segment_id, router_name, gw_ip,
|
||||
router_ip, mask, self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'ip routing',
|
||||
'vlan %s' % segment_id, 'exit',
|
||||
'interface vlan %s' % segment_id,
|
||||
'ip address %s/%s' % (gw_ip, mask), 'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_delete_interface_from_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
|
||||
self.drv.delete_interface_from_router(segment_id, router_name,
|
||||
self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'no interface vlan %s' % segment_id,
|
||||
'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
|
||||
class AristaL3DriverTestCasesUsingVRFs(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions using multiple VRFs.
|
||||
Note that the configuration commands are different when VRFs are used.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCasesUsingVRFs, self).setUp()
|
||||
setup_arista_config('value', vrf=True)
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_create_router_on_eos(self):
|
||||
max_vrfs = 5
|
||||
routers = ['testRouter-%s' % n for n in range(max_vrfs)]
|
||||
domains = ['10%s' % n for n in range(max_vrfs)]
|
||||
|
||||
for (r, d) in zip(routers, domains):
|
||||
self.drv.create_router_on_eos(r, d, self.drv._servers[0])
|
||||
|
||||
cmds = ['enable', 'configure',
|
||||
'vrf definition %s' % r,
|
||||
'rd %(rd)s:%(rd)s' % {'rd': d}, 'exit', 'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_with(cmds)
|
||||
|
||||
def test_delete_router_from_eos(self):
|
||||
max_vrfs = 5
|
||||
routers = ['testRouter-%s' % n for n in range(max_vrfs)]
|
||||
|
||||
for r in routers:
|
||||
self.drv.delete_router_from_eos(r, self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'no vrf definition %s' % r,
|
||||
'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_with(cmds)
|
||||
|
||||
def test_add_interface_to_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
router_ip = '10.10.10.10'
|
||||
gw_ip = '10.10.10.1'
|
||||
mask = '255.255.255.0'
|
||||
|
||||
self.drv.add_interface_to_router(segment_id, router_name, gw_ip,
|
||||
router_ip, mask, self.drv._servers[0])
|
||||
cmds = ['enable', 'configure',
|
||||
'ip routing vrf %s' % router_name,
|
||||
'vlan %s' % segment_id, 'exit',
|
||||
'interface vlan %s' % segment_id,
|
||||
'vrf forwarding %s' % router_name,
|
||||
'ip address %s/%s' % (gw_ip, mask), 'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_delete_interface_from_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
|
||||
self.drv.delete_interface_from_router(segment_id, router_name,
|
||||
self.drv._servers[0])
|
||||
cmds = ['enable', 'configure', 'no interface vlan %s' % segment_id,
|
||||
'exit']
|
||||
|
||||
self.drv._servers[0].execute.assert_called_once_with(cmds)
|
||||
|
||||
|
||||
class AristaL3DriverTestCasesMlagConfig(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions in Default VRF using MLAG configuration.
|
||||
MLAG configuration means that the commands will be sent to both
|
||||
primary and secondary Arista Switches.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCasesMlagConfig, self).setUp()
|
||||
setup_arista_config('value', mlag=True)
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_create_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
route_domain = '123:123'
|
||||
router_mac = '00:11:22:33:44:55'
|
||||
|
||||
for s in self.drv._servers:
|
||||
self.drv.create_router_on_eos(router_name, route_domain, s)
|
||||
cmds = ['enable', 'configure',
|
||||
'ip virtual-router mac-address %s' % router_mac, 'exit']
|
||||
|
||||
s.execute.assert_called_with(cmds)
|
||||
|
||||
def test_delete_router_from_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
|
||||
for s in self.drv._servers:
|
||||
self.drv.delete_router_from_eos(router_name, s)
|
||||
cmds = ['enable', 'configure', 'exit']
|
||||
|
||||
s.execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_add_interface_to_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
router_ip = '10.10.10.10'
|
||||
gw_ip = '10.10.10.1'
|
||||
mask = '255.255.255.0'
|
||||
|
||||
for s in self.drv._servers:
|
||||
self.drv.add_interface_to_router(segment_id, router_name, gw_ip,
|
||||
router_ip, mask, s)
|
||||
cmds = ['enable', 'configure', 'ip routing',
|
||||
'vlan %s' % segment_id, 'exit',
|
||||
'interface vlan %s' % segment_id,
|
||||
'ip address %s' % router_ip,
|
||||
'ip virtual-router address %s' % gw_ip, 'exit']
|
||||
|
||||
s.execute.assert_called_once_with(cmds)
|
||||
|
||||
def test_delete_interface_from_router_on_eos(self):
|
||||
router_name = 'test-router-1'
|
||||
segment_id = '123'
|
||||
|
||||
for s in self.drv._servers:
|
||||
self.drv.delete_interface_from_router(segment_id, router_name, s)
|
||||
|
||||
cmds = ['enable', 'configure', 'no interface vlan %s' % segment_id,
|
||||
'exit']
|
||||
|
||||
s.execute.assert_called_once_with(cmds)
|
||||
|
||||
|
||||
class AristaL3DriverTestCases_v4(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions in Default VRF using IPv4.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCases_v4, self).setUp()
|
||||
setup_arista_config('value')
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_add_v4_interface_to_router(self):
|
||||
gateway_ip = '10.10.10.1'
|
||||
cidrs = ['10.10.10.0/24', '10.11.11.0/24']
|
||||
|
||||
# Add couple of IPv4 subnets to router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 4}
|
||||
|
||||
self.assertFalse(self.drv.add_router_interface(None, router))
|
||||
|
||||
def test_delete_v4_interface_from_router(self):
|
||||
gateway_ip = '10.10.10.1'
|
||||
cidrs = ['10.10.10.0/24', '10.11.11.0/24']
|
||||
|
||||
# remove couple of IPv4 subnets from router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 4}
|
||||
|
||||
self.assertFalse(self.drv.remove_router_interface(None, router))
|
||||
|
||||
|
||||
class AristaL3DriverTestCases_v6(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions in Default VRF using IPv6.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCases_v6, self).setUp()
|
||||
setup_arista_config('value')
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_add_v6_interface_to_router(self):
|
||||
gateway_ip = '3FFE::1'
|
||||
cidrs = ['3FFE::/16', '2001::/16']
|
||||
|
||||
# Add couple of IPv6 subnets to router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 6}
|
||||
|
||||
self.assertFalse(self.drv.add_router_interface(None, router))
|
||||
|
||||
def test_delete_v6_interface_from_router(self):
|
||||
gateway_ip = '3FFE::1'
|
||||
cidrs = ['3FFE::/16', '2001::/16']
|
||||
|
||||
# remove couple of IPv6 subnets from router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 6}
|
||||
|
||||
self.assertFalse(self.drv.remove_router_interface(None, router))
|
||||
|
||||
|
||||
class AristaL3DriverTestCases_MLAG_v6(base.BaseTestCase):
|
||||
"""Test cases to test the RPC between Arista Driver and EOS.
|
||||
|
||||
Tests all methods used to send commands between Arista L3 Driver and EOS
|
||||
to program routing functions in Default VRF on MLAG'ed switches using IPv6.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCases_MLAG_v6, self).setUp()
|
||||
setup_arista_config('value', mlag=True)
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_no_exception_on_correct_configuration(self):
|
||||
self.assertIsNotNone(self.drv)
|
||||
|
||||
def test_add_v6_interface_to_router(self):
|
||||
gateway_ip = '3FFE::1'
|
||||
cidrs = ['3FFE::/16', '2001::/16']
|
||||
|
||||
# Add couple of IPv6 subnets to router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 6}
|
||||
|
||||
self.assertFalse(self.drv.add_router_interface(None, router))
|
||||
|
||||
def test_delete_v6_interface_from_router(self):
|
||||
gateway_ip = '3FFE::1'
|
||||
cidrs = ['3FFE::/16', '2001::/16']
|
||||
|
||||
# remove couple of IPv6 subnets from router
|
||||
for cidr in cidrs:
|
||||
router = {'name': 'test-router-1',
|
||||
'tenant_id': 'ten-a',
|
||||
'seg_id': '123',
|
||||
'cidr': "%s" % cidr,
|
||||
'gip': "%s" % gateway_ip,
|
||||
'ip_version': 6}
|
||||
|
||||
self.assertFalse(self.drv.remove_router_interface(None, router))
|
||||
|
||||
|
||||
class AristaL3DriverTestCasesMlag_one_switch_failed(base.BaseTestCase):
|
||||
"""Test cases to test with non redundant hardare in redundancy mode.
|
||||
|
||||
In the following test cases, the driver is configured in MLAG (redundancy
|
||||
mode) but, one of the switches is mocked to throw exceptoin to mimic
|
||||
failure of the switch. Ensure that the the operation does not fail when
|
||||
one of the switches fails.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(AristaL3DriverTestCasesMlag_one_switch_failed, self).setUp()
|
||||
setup_arista_config('value', mlag=True)
|
||||
self.drv = arista.AristaL3Driver()
|
||||
self.drv._servers = []
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
self.drv._servers.append(mock.MagicMock())
|
||||
|
||||
def test_create_router_when_one_switch_fails(self):
|
||||
router = {}
|
||||
router['name'] = 'test-router-1'
|
||||
tenant = '123'
|
||||
|
||||
# Make one of the switches throw an exception - i.e. fail
|
||||
self.drv._servers[0].execute = mock.Mock(side_effect=Exception)
|
||||
with mock.patch.object(arista.LOG, 'exception') as log_exception:
|
||||
self.drv.create_router(None, tenant, router)
|
||||
log_exception.assert_called_once_with(mock.ANY)
|
||||
|
||||
def test_delete_router_when_one_switch_fails(self):
|
||||
router = {}
|
||||
router['name'] = 'test-router-1'
|
||||
tenant = '123'
|
||||
router_id = '345'
|
||||
|
||||
# Make one of the switches throw an exception - i.e. fail
|
||||
self.drv._servers[1].execute = mock.Mock(side_effect=Exception)
|
||||
with mock.patch.object(arista.LOG, 'exception') as log_exception:
|
||||
self.drv.delete_router(None, tenant, router_id, router)
|
||||
log_exception.assert_called_once_with(mock.ANY)
|
||||
|
||||
def test_add_router_interface_when_one_switch_fails(self):
|
||||
router = {}
|
||||
router['name'] = 'test-router-1'
|
||||
router['tenant_id'] = 'ten-1'
|
||||
router['seg_id'] = '100'
|
||||
router['ip_version'] = 4
|
||||
router['cidr'] = '10.10.10.0/24'
|
||||
router['gip'] = '10.10.10.1'
|
||||
|
||||
# Make one of the switches throw an exception - i.e. fail
|
||||
self.drv._servers[1].execute = mock.Mock(side_effect=Exception)
|
||||
with mock.patch.object(arista.LOG, 'exception') as log_exception:
|
||||
self.drv.add_router_interface(None, router)
|
||||
log_exception.assert_called_once_with(mock.ANY)
|
||||
|
||||
def test_remove_router_interface_when_one_switch_fails(self):
|
||||
router = {}
|
||||
router['name'] = 'test-router-1'
|
||||
router['tenant_id'] = 'ten-1'
|
||||
router['seg_id'] = '100'
|
||||
router['ip_version'] = 4
|
||||
router['cidr'] = '10.10.10.0/24'
|
||||
router['gip'] = '10.10.10.1'
|
||||
|
||||
# Make one of the switches throw an exception - i.e. fail
|
||||
self.drv._servers[0].execute = mock.Mock(side_effect=Exception)
|
||||
with mock.patch.object(arista.LOG, 'exception') as log_exception:
|
||||
self.drv.remove_router_interface(None, router)
|
||||
log_exception.assert_called_once_with(mock.ANY)
|
@ -1,19 +0,0 @@
|
||||
# Copyright 2011 OpenStack Foundation.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslo_config import cfg
|
||||
|
||||
|
||||
cfg.CONF.use_stderr = False
|
File diff suppressed because it is too large
Load Diff
@ -1,111 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import itertools
|
||||
|
||||
import mock
|
||||
from mock import patch
|
||||
from neutron_lib.db import api as db_api
|
||||
from oslo_config import cfg
|
||||
|
||||
from neutron.db.models.plugins.ml2 import vlanallocation
|
||||
from neutron.tests.unit import testlib_api
|
||||
|
||||
from networking_arista.ml2.drivers.driver_helpers import VlanSyncService
|
||||
from networking_arista.ml2.drivers.type_arista_vlan import AristaVlanTypeDriver
|
||||
import networking_arista.tests.unit.ml2.utils as utils
|
||||
|
||||
|
||||
EAPI_SEND_FUNC = ('networking_arista.ml2.arista_ml2.AristaRPCWrapperEapi'
|
||||
'._send_eapi_req')
|
||||
|
||||
|
||||
class AristaTypeDriverTest(testlib_api.SqlTestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(AristaTypeDriverTest, self).setUp()
|
||||
utils.setup_arista_wrapper_config(cfg)
|
||||
|
||||
@patch(EAPI_SEND_FUNC)
|
||||
def test_initialize_type_driver(self, mock_send_eapi_req):
|
||||
type_driver = AristaVlanTypeDriver()
|
||||
type_driver.sync_service._force_sync = False
|
||||
type_driver.sync_service._vlan_assignment_uuid = {'uuid': 1}
|
||||
type_driver.sync_service._rpc = mock.MagicMock()
|
||||
rpc = type_driver.sync_service._rpc
|
||||
rpc.get_vlan_assignment_uuid.return_value = {'uuid': 1}
|
||||
type_driver.initialize()
|
||||
|
||||
cmds = ['show openstack agent uuid',
|
||||
'show openstack instances',
|
||||
'show openstack agent uuid',
|
||||
'show openstack features']
|
||||
|
||||
calls = [mock.call(cmds=[cmd], commands_to_log=[cmd])
|
||||
for cmd in cmds]
|
||||
mock_send_eapi_req.assert_has_calls(calls)
|
||||
type_driver.timer.cancel()
|
||||
|
||||
|
||||
class VlanSyncServiceTest(testlib_api.SqlTestCase):
|
||||
"""Test that VLANs are synchronized between EOS and Neutron."""
|
||||
|
||||
def _ensure_in_db(self, assigned, allocated, available):
|
||||
session = db_api.get_reader_session()
|
||||
with session.begin():
|
||||
vlans = session.query(vlanallocation.VlanAllocation).all()
|
||||
for vlan in vlans:
|
||||
self.assertIn(vlan.vlan_id, assigned)
|
||||
|
||||
if vlan.vlan_id in available:
|
||||
self.assertFalse(vlan.allocated)
|
||||
elif vlan.vlan_id in allocated:
|
||||
self.assertTrue(vlan.allocated)
|
||||
|
||||
def test_synchronization_test(self):
|
||||
rpc = mock.MagicMock()
|
||||
|
||||
rpc.get_vlan_allocation.return_value = {
|
||||
'assignedVlans': '1-10,21-30',
|
||||
'availableVlans': '1-5,21,23,25,27,29',
|
||||
'allocatedVlans': '6-10,22,24,26,28,30'
|
||||
}
|
||||
|
||||
assigned = list(itertools.chain(range(1, 11), range(21, 31)))
|
||||
|
||||
available = [1, 2, 3, 4, 5, 21, 23, 25, 27, 29]
|
||||
allocated = list(set(assigned) - set(available))
|
||||
|
||||
sync_service = VlanSyncService(rpc)
|
||||
sync_service.synchronize()
|
||||
|
||||
self._ensure_in_db(assigned, allocated, available)
|
||||
|
||||
# Call synchronize again which returns different data
|
||||
rpc.get_vlan_allocation.return_value = {
|
||||
'assignedVlans': '51-60,71-80',
|
||||
'availableVlans': '51-55,71,73,75,77,79',
|
||||
'allocatedVlans': '56-60,72,74,76,78,80'
|
||||
}
|
||||
|
||||
assigned = list(itertools.chain(range(51, 61), range(71, 81)))
|
||||
|
||||
available = [51, 52, 53, 54, 55, 71, 73, 75, 77, 79]
|
||||
allocated = list(set(assigned) - set(available))
|
||||
|
||||
sync_service = VlanSyncService(rpc)
|
||||
sync_service.synchronize()
|
||||
|
||||
self._ensure_in_db(assigned, allocated, available)
|
File diff suppressed because it is too large
Load Diff
@ -1,39 +0,0 @@
|
||||
# Copyright (c) 2016 OpenStack Foundation
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
def setup_arista_wrapper_config(cfg, host='host', user='user'):
|
||||
cfg.CONF.keystone_authtoken = fake_keystone_info_class()
|
||||
cfg.CONF.set_override('eapi_host', host, "ml2_arista")
|
||||
cfg.CONF.set_override('eapi_username', user, "ml2_arista")
|
||||
cfg.CONF.set_override('sync_interval', 10, "ml2_arista")
|
||||
cfg.CONF.set_override('conn_timeout', 20, "ml2_arista")
|
||||
cfg.CONF.set_override('switch_info', ['switch1:user:pass'], "ml2_arista")
|
||||
cfg.CONF.set_override('sec_group_support', False, "ml2_arista")
|
||||
|
||||
|
||||
class fake_keystone_info_class(object):
|
||||
"""To generate fake Keystone Authentication token information
|
||||
|
||||
Arista Driver expects Keystone auth info. This fake information
|
||||
is for testing only
|
||||
"""
|
||||
auth_uri = False
|
||||
auth_protocol = 'abc'
|
||||
auth_host = 'host'
|
||||
auth_port = 5000
|
||||
admin_user = 'neutron'
|
||||
admin_password = 'fun'
|
||||
admin_tenant_name = 'tenant_name'
|
@ -1,15 +0,0 @@
|
||||
# The order of packages is significant, because pip processes them in the order
|
||||
# of appearance. Changing the order has an impact on the overall integration
|
||||
# process, which may cause wedges in the gate later.
|
||||
pbr!=2.1.0,>=2.0.0 # Apache-2.0
|
||||
|
||||
alembic>=0.8.10 # MIT
|
||||
neutron-lib>=1.9.0 # Apache-2.0
|
||||
oslo.i18n!=3.15.2,>=2.1.0 # Apache-2.0
|
||||
oslo.config!=4.3.0,!=4.4.0,>=4.0.0 # Apache-2.0
|
||||
oslo.log>=3.22.0 # Apache-2.0
|
||||
oslo.service>=1.10.0 # Apache-2.0
|
||||
oslo.utils>=3.20.0 # Apache-2.0
|
||||
requests>=2.14.2 # Apache-2.0
|
||||
six>=1.9.0 # MIT
|
||||
SQLAlchemy!=1.1.5,!=1.1.6,!=1.1.7,!=1.1.8,>=1.0.10 # MIT
|
66
setup.cfg
66
setup.cfg
@ -1,66 +0,0 @@
|
||||
[metadata]
|
||||
name = networking_arista
|
||||
summary = Arista Networking drivers
|
||||
description-file =
|
||||
README.rst
|
||||
author = Arista Networks
|
||||
author-email = openstack-dev@arista.com
|
||||
home-page = https://github.com/openstack/networking-arista/
|
||||
classifier =
|
||||
Environment :: OpenStack
|
||||
Intended Audience :: Information Technology
|
||||
Intended Audience :: System Administrators
|
||||
License :: OSI Approved :: Apache Software License
|
||||
Operating System :: POSIX :: Linux
|
||||
Programming Language :: Python
|
||||
Programming Language :: Python :: 2
|
||||
Programming Language :: Python :: 2.7
|
||||
Programming Language :: Python :: 3
|
||||
Programming Language :: Python :: 3.5
|
||||
|
||||
[files]
|
||||
packages =
|
||||
networking_arista
|
||||
data_files =
|
||||
/etc/neutron/plugins/ml2 =
|
||||
etc/ml2_conf_arista.ini
|
||||
|
||||
[global]
|
||||
setup-hooks =
|
||||
pbr.hooks.setup_hook
|
||||
|
||||
[entry_points]
|
||||
neutron.ml2.mechanism_drivers =
|
||||
arista = networking_arista.ml2.mechanism_arista:AristaDriver
|
||||
arista_ml2 = networking_arista.ml2.mechanism_arista:AristaDriver
|
||||
neutron.service_plugins =
|
||||
arista_l3 = networking_arista.l3Plugin.l3_arista:AristaL3ServicePlugin
|
||||
neutron.db.alembic_migrations =
|
||||
networking-arista = networking_arista.db.migration:alembic_migrations
|
||||
neutron.ml2.type_drivers =
|
||||
arista_vlan = networking_arista.ml2.drivers.type_arista_vlan:AristaVlanTypeDriver
|
||||
|
||||
[build_sphinx]
|
||||
source-dir = doc/source
|
||||
build-dir = doc/build
|
||||
all_files = 1
|
||||
|
||||
[upload_sphinx]
|
||||
upload-dir = doc/build/html
|
||||
|
||||
[compile_catalog]
|
||||
directory = networking_arista/locale
|
||||
domain = networking-arista
|
||||
|
||||
[update_catalog]
|
||||
domain = networking-arista
|
||||
output_dir = networking_arista/locale
|
||||
input_file = networking_arista/locale/networking-arista.pot
|
||||
|
||||
[extract_messages]
|
||||
keywords = _ gettext ngettext l_ lazy_gettext
|
||||
mapping_file = babel.cfg
|
||||
output_file = networking_arista/locale/networking-arista.pot
|
||||
|
||||
[wheel]
|
||||
universal = 1
|
29
setup.py
29
setup.py
@ -1,29 +0,0 @@
|
||||
# Copyright (c) 2013 Hewlett-Packard Development Company, L.P.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
# THIS FILE IS MANAGED BY THE GLOBAL REQUIREMENTS REPO - DO NOT EDIT
|
||||
import setuptools
|
||||
|
||||
# In python < 2.7.4, a lazy loading of package `pbr` will break
|
||||
# setuptools if some other modules registered functions in `atexit`.
|
||||
# solution from: http://bugs.python.org/issue15881#msg170215
|
||||
try:
|
||||
import multiprocessing # noqa
|
||||
except ImportError:
|
||||
pass
|
||||
|
||||
setuptools.setup(
|
||||
setup_requires=['pbr>=2.0.0'],
|
||||
pbr=True)
|
@ -1,15 +0,0 @@
|
||||
# The order of packages is significant, because pip processes them in the order
|
||||
# of appearance. Changing the order has an impact on the overall integration
|
||||
# process, which may cause wedges in the gate later.
|
||||
hacking!=0.13.0,<0.14,>=0.12.0 # Apache-2.0
|
||||
|
||||
coverage!=4.4,>=4.0 # Apache-2.0
|
||||
mock>=2.0 # BSD
|
||||
python-subunit>=0.0.18 # Apache-2.0/BSD
|
||||
sphinx>=1.6.2 # BSD
|
||||
oslosphinx>=4.7.0 # Apache-2.0
|
||||
oslotest>=1.10.0 # Apache-2.0
|
||||
testrepository>=0.0.18 # Apache-2.0/BSD
|
||||
testtools>=1.4.0 # MIT
|
||||
testresources>=0.2.4 # Apache-2.0/BSD
|
||||
testscenarios>=0.4 # Apache-2.0/BSD
|
40
tox.ini
40
tox.ini
@ -1,40 +0,0 @@
|
||||
[tox]
|
||||
envlist = py27,py35,pep8
|
||||
minversion = 1.6
|
||||
skipsdist = True
|
||||
|
||||
[testenv]
|
||||
usedevelop = True
|
||||
install_command = pip install -c {env:UPPER_CONSTRAINTS_FILE:https://git.openstack.org/cgit/openstack/requirements/plain/upper-constraints.txt} -r requirements.txt -U {opts} {packages}
|
||||
setenv = VIRTUAL_ENV={envdir}
|
||||
PYTHONWARNINGS=default::DeprecationWarning
|
||||
deps = -r{toxinidir}/test-requirements.txt
|
||||
-egit+https://git.openstack.org/openstack/neutron.git#egg=neutron
|
||||
whitelist_externals = sh
|
||||
commands = python setup.py testr --slowest --testr-args='{posargs}'
|
||||
|
||||
[testenv:pep8]
|
||||
commands =
|
||||
flake8
|
||||
neutron-db-manage --subproject networking-arista check_migration
|
||||
|
||||
[testenv:venv]
|
||||
commands = {posargs}
|
||||
|
||||
[testenv:cover]
|
||||
commands = python setup.py testr --coverage --testr-args='{posargs}'
|
||||
|
||||
[testenv:docs]
|
||||
commands = python setup.py build_sphinx
|
||||
|
||||
[flake8]
|
||||
# H803 skipped on purpose per list discussion.
|
||||
# E123, E125 skipped as they are invalid PEP-8.
|
||||
|
||||
show-source = True
|
||||
ignore = E123,E125,H803
|
||||
builtins = _
|
||||
exclude=.venv,.git,.tox,dist,doc,*lib/python*,*egg,build
|
||||
|
||||
[hacking]
|
||||
import_exceptions = networking_arista._i18n
|
Loading…
Reference in New Issue
Block a user