Align project files structure with cookiecutter template
1.What is the problem An OpenStack project should be created based on the template generated by cookiecutter: http://docs.openstack.org/infra/manual/creators.html#preparing-a-new-git-repository-using-cookiecutter There are some files missing in the Tricircle compared to the files generated by cookiecutter template. 2.What's need to be fixed: This patch sets is to amend this missing files. Moreover, README.md was renamed to README.rst, and the installation part with devstack was moved to doc/source/installation.rst according to the cookiecutter generated template. The README.rst in the Tricircle root folder was updated accordingly, and only the description of the Tricircle was remained. 3.What is the purpose of this patch set: To make the Tricircle follow other OpenStack project folder structure template which could be generated by cookiecutter to keep consistency. Change-Id: I65d4c376a87eccded12bd3f08bcd9ee89def95d4 Signed-off-by: Chaoyi Huang <joehuang@huawei.com>
This commit is contained in:
parent
98364eda20
commit
5a6239391f
7
.coveragerc
Normal file
7
.coveragerc
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
[run]
|
||||||
|
branch = True
|
||||||
|
source = tricircle
|
||||||
|
omit = tricircle/tests/*, tricircle/tempestplugin/*
|
||||||
|
|
||||||
|
[report]
|
||||||
|
ignore_errors = True
|
17
CONTRIBUTING.rst
Normal file
17
CONTRIBUTING.rst
Normal file
@ -0,0 +1,17 @@
|
|||||||
|
If you would like to contribute to the development of OpenStack, you should
|
||||||
|
follow the steps in this page:
|
||||||
|
|
||||||
|
http://docs.openstack.org/infra/manual/developers.html
|
||||||
|
|
||||||
|
If you already knew how the OpenStack CI system works and your
|
||||||
|
OpenStack accounts is setup properly, you can start from the development
|
||||||
|
workflow section in that documentation to know how you should commit your
|
||||||
|
patch set for review via the Gerrit tool:
|
||||||
|
|
||||||
|
http://docs.openstack.org/infra/manual/developers.html#development-workflow
|
||||||
|
|
||||||
|
Any pull requests submitted through GitHub will be ignored.
|
||||||
|
|
||||||
|
Any bug should be filed on Launchpad, not GitHub:
|
||||||
|
|
||||||
|
https://bugs.launchpad.net/tricircle
|
6
HACKING.rst
Normal file
6
HACKING.rst
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
================================
|
||||||
|
The Tricircle Style Commandments
|
||||||
|
================================
|
||||||
|
|
||||||
|
Please read the OpenStack Style Commandments
|
||||||
|
http://docs.openstack.org/developer/hacking/
|
6
MANIFEST.in
Normal file
6
MANIFEST.in
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
include AUTHORS
|
||||||
|
include ChangeLog
|
||||||
|
exclude .gitignore
|
||||||
|
exclude .gitreview
|
||||||
|
|
||||||
|
global-exclude *.pyc
|
447
README.md
447
README.md
@ -1,447 +0,0 @@
|
|||||||
# Tricircle
|
|
||||||
|
|
||||||
(The original PoC source code, please switch to
|
|
||||||
["poc"](https://github.com/openstack/tricircle/tree/poc) tag, or
|
|
||||||
["stable/fortest"](https://github.com/openstack/tricircle/tree/stable/fortest)
|
|
||||||
branch)
|
|
||||||
|
|
||||||
Tricircle is an OpenStack project that aims to deal with multiple OpenStack
|
|
||||||
deployment across multiple data centers. It provides users a single management
|
|
||||||
view by having only one Tricircle instance on behalf of all the involved
|
|
||||||
OpenStack instances.
|
|
||||||
|
|
||||||
Tricircle presents one big region to the end user in KeyStone. And each
|
|
||||||
OpenStack instance, which is called a pod, is a sub-region of Tricircle in
|
|
||||||
KeyStone, and not visible to end user directly.
|
|
||||||
|
|
||||||
Tricircle acts as OpenStack API gateway, can accept all OpenStack API calls
|
|
||||||
and forward the API calls to regarding OpenStack instance(pod), and deal with
|
|
||||||
cross pod networking automaticly.
|
|
||||||
|
|
||||||
The end user can see avaialbility zone (AZ in short) and use AZ to provision
|
|
||||||
VM, Volume, even Network through Tricircle.
|
|
||||||
|
|
||||||
Similar as AWS, one AZ can includes many pods, and a tenant's resources will
|
|
||||||
be bound to specific pods automaticly.
|
|
||||||
|
|
||||||
## Project Resources
|
|
||||||
License: Apache 2.0
|
|
||||||
|
|
||||||
- Design documentation: [Tricircle Design Blueprint](https://docs.google.com/document/d/18kZZ1snMOCD9IQvUKI5NVDzSASpw-QKj7l2zNqMEd3g/)
|
|
||||||
- Wiki: https://wiki.openstack.org/wiki/tricircle
|
|
||||||
- Documentation: http://docs.openstack.org/developer/tricircle
|
|
||||||
- Source: https://github.com/openstack/tricircle
|
|
||||||
- Bugs: http://bugs.launchpad.net/tricircle
|
|
||||||
- Blueprints: https://launchpad.net/tricircle
|
|
||||||
|
|
||||||
## Play with DevStack
|
|
||||||
Now stateless design can be played with DevStack.
|
|
||||||
|
|
||||||
- 1 Git clone DevStack.
|
|
||||||
- 2 Git clone Tricircle, or just download devstack/local.conf.sample.
|
|
||||||
- 3 Copy devstack/local.conf.sample to DevStack folder and rename it to
|
|
||||||
local.conf, change password in the file if needed.
|
|
||||||
- 4 Run DevStack.
|
|
||||||
- 5 After DevStack successfully starts, check if services have been correctly
|
|
||||||
registered. Run "openstack endpoint list" and you should get similar output
|
|
||||||
as following:
|
|
||||||
```
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
| ID | Region | Service Name | Service Type |
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
| 230059e8533e4d389e034fd68257034b | RegionOne | glance | image |
|
|
||||||
| 25180a0a08cb41f69de52a7773452b28 | RegionOne | nova | compute |
|
|
||||||
| bd1ed1d6f0cc42398688a77bcc3bda91 | Pod1 | neutron | network |
|
|
||||||
| 673736f54ec147b79e97c395afe832f9 | RegionOne | ec2 | ec2 |
|
|
||||||
| fd7f188e2ba04ebd856d582828cdc50c | RegionOne | neutron | network |
|
|
||||||
| ffb56fd8b24a4a27bf6a707a7f78157f | RegionOne | keystone | identity |
|
|
||||||
| 88da40693bfa43b9b02e1478b1fa0bc6 | Pod1 | nova | compute |
|
|
||||||
| f35d64c2ddc44c16a4f9dfcd76e23d9f | RegionOne | nova_legacy | compute_legacy |
|
|
||||||
| 8759b2941fe7469e9651de3f6a123998 | RegionOne | tricircle | Cascading |
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
```
|
|
||||||
"RegionOne" is the region you set in local.conf via REGION_NAME, whose default
|
|
||||||
value is "RegionOne", we use it as the region for top OpenStack(Tricircle);
|
|
||||||
"Pod1" is the region set via "POD_REGION_NAME", new configuration option
|
|
||||||
introduced by Tricircle, we use it as the bottom OpenStack.
|
|
||||||
- 6 Create pod instances for Tricircle and bottom OpenStack
|
|
||||||
```
|
|
||||||
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "RegionOne"}}'
|
|
||||||
|
|
||||||
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod1", "az_name": "az1"}}'
|
|
||||||
```
|
|
||||||
Pay attention to "pod_name" parameter we specify when creating pod. Pod name
|
|
||||||
should exactly match the region name registered in Keystone since it is used
|
|
||||||
by Tricircle to route API request. In the above commands, we create pods named
|
|
||||||
"RegionOne" and "Pod1" for top OpenStack(Tricircle) and bottom OpenStack.
|
|
||||||
Tricircle API service will automatically create a aggregate when user creates
|
|
||||||
a bottom pod, so command "nova aggregate-list" will show the following result:
|
|
||||||
```
|
|
||||||
+----+----------+-------------------+
|
|
||||||
| Id | Name | Availability Zone |
|
|
||||||
+----+----------+-------------------+
|
|
||||||
| 1 | ag_Pod1 | az1 |
|
|
||||||
+----+----------+-------------------+
|
|
||||||
```
|
|
||||||
- 7 Create necessary resources to boot a virtual machine.
|
|
||||||
```
|
|
||||||
nova flavor-create test 1 1024 10 1
|
|
||||||
neutron net-create net1
|
|
||||||
neutron subnet-create net1 10.0.0.0/24
|
|
||||||
glance image-list
|
|
||||||
```
|
|
||||||
Note that flavor mapping has not been implemented yet so the created flavor is
|
|
||||||
just a database record and actually flavor in bottom OpenStack with the same id
|
|
||||||
will be used.
|
|
||||||
- 8 Boot a virtual machine.
|
|
||||||
```
|
|
||||||
nova boot --flavor 1 --image $image_id --nic net-id=$net_id --availability-zone az1 vm1
|
|
||||||
```
|
|
||||||
- 9 Create, list, show and delete volume.
|
|
||||||
```
|
|
||||||
cinder --debug create --availability-zone=az1 1
|
|
||||||
cinder --debug list
|
|
||||||
cinder --debug show $volume_id
|
|
||||||
cinder --debug delete $volume_id
|
|
||||||
cinder --debug list
|
|
||||||
```
|
|
||||||
### Quick Verify
|
|
||||||
|
|
||||||
A sample of admin-openrc.sh and an installation verification script can be found
|
|
||||||
in devstack/ directory.
|
|
||||||
|
|
||||||
#### admin-openrc.sh
|
|
||||||
|
|
||||||
Create client environment variables for the admin user as the following:
|
|
||||||
|
|
||||||
```
|
|
||||||
export OS_PROJECT_DOMAIN_ID=default
|
|
||||||
export OS_USER_DOMAIN_ID=default
|
|
||||||
export OS_PROJECT_NAME=admin
|
|
||||||
export OS_TENANT_NAME=admin
|
|
||||||
export OS_USERNAME=admin
|
|
||||||
export OS_PASSWORD=password #change password as you set in your own environment
|
|
||||||
export OS_AUTH_URL=http://127.0.0.1:5000
|
|
||||||
export OS_IDENTITY_API_VERSION=3 #It's very important to set region name of top openstack,
|
|
||||||
#because tricircle has different API urls.
|
|
||||||
export OS_IMAGE_API_VERSION=2
|
|
||||||
export OS_REGION_NAME=RegionOne
|
|
||||||
|
|
||||||
```
|
|
||||||
The command to use the admin-openrc.sh is:
|
|
||||||
|
|
||||||
```
|
|
||||||
source tricircle/devstack/admin-openrc.sh
|
|
||||||
```
|
|
||||||
|
|
||||||
#### verify_top_install.sh
|
|
||||||
|
|
||||||
This script is to quickly verify the installation of Tricircle in Top OpenStack
|
|
||||||
as the step 5-9 above and save the output to logs.
|
|
||||||
Before verifying the installation, you should modify the script to your own
|
|
||||||
environment.
|
|
||||||
|
|
||||||
- 1 The default post URL is 127.0.0.1, change it if needed,
|
|
||||||
- 2 The default create net1 is 10.0.0.0/24, change it if needed.
|
|
||||||
|
|
||||||
Then you do the following steps in Top OpenStack to verify:
|
|
||||||
|
|
||||||
```
|
|
||||||
cd tricircle/devstack/
|
|
||||||
./verify_top_install.sh 2>&1 | tee logs
|
|
||||||
```
|
|
||||||
|
|
||||||
## Cross-pod L3 networking with DevStack
|
|
||||||
Now stateless design supports cross-pod l3 networking.
|
|
||||||
|
|
||||||
### Introduction
|
|
||||||
|
|
||||||
To achieve cross-pod l3 networking, Tricircle utilizes a shared provider VLAN
|
|
||||||
network at first phase. We are considering later using DCI controller to create
|
|
||||||
a multi-segment VLAN network, VxLAN network for L3 networking purpose. When a
|
|
||||||
subnet is attached to a router in top pod, Tricircle not only creates corresponding
|
|
||||||
subnet and router in bottom pod, but also creates a VLAN type "bridge" network.
|
|
||||||
Both tenant network and "bridge" network are attached to bottom router. Each
|
|
||||||
tenant will have one allocated VLAN, which is shared by the tenant's "bridge"
|
|
||||||
networks across bottom pods. The CIDRs of "bridge" networks for one tenant are
|
|
||||||
also the same, so the router interfaces in "bridge" networks across different
|
|
||||||
bottom pods can communicate with each other via the provider VLAN network. By
|
|
||||||
adding an extra route as following:
|
|
||||||
```
|
|
||||||
destination: CIDR of tenant network in another bottom pod
|
|
||||||
nexthop: "bridge" network interface ip in another bottom pod
|
|
||||||
```
|
|
||||||
when a server sends a packet whose receiver is in another network and in
|
|
||||||
another bottom pod, the packet first goes to router namespace, then is
|
|
||||||
forwarded to the router namespace in another bottom pod according to the extra
|
|
||||||
route, at last the packet is sent to the target server. This configuration job
|
|
||||||
is triggered when user attaches a subnet to a router in top pod and finished
|
|
||||||
asynchronously.
|
|
||||||
|
|
||||||
Currently cross-pod L2 networking is not supported yet, so tenant networks
|
|
||||||
cannot cross pods, that is to say, one network in top pod can only locate in
|
|
||||||
one bottom pod, tenant network is bound to bottom pod. Otherwise we cannot
|
|
||||||
correctly configure extra route since for one destination CIDR, we have more
|
|
||||||
than one possible nexthop addresses.
|
|
||||||
|
|
||||||
> When cross-pod L2 networking is introduced, L2GW will be used to connect L2
|
|
||||||
> network in different pods. No extra route is required to connect L2 network.
|
|
||||||
> All L3 traffic will be forwarded to the local L2 network, then go to the
|
|
||||||
> server in another pod via the L2GW.
|
|
||||||
|
|
||||||
We use "availability_zone_hints" attribute for user to specify the bottom pod
|
|
||||||
he wants to create the bottom network. Currently we do not support attaching
|
|
||||||
a network to a router without setting "availability_zone_hints" attribute of
|
|
||||||
the network.
|
|
||||||
|
|
||||||
### Prerequisite
|
|
||||||
|
|
||||||
To play cross-pod L3 networking, two nodes are needed. One to run Tricircle
|
|
||||||
and one bottom pod, the other one to run another bottom pod. Both nodes have
|
|
||||||
two network interfaces, for management and provider VLAN network. For VLAN
|
|
||||||
network, the physical network infrastructure should support VLAN tagging. If
|
|
||||||
you would like to try north-south networking, too, you should prepare one more
|
|
||||||
network interface in the second node for external network. In this guide, the
|
|
||||||
external network is also vlan type, so the local.conf sample is based on vlan
|
|
||||||
type external network setup.
|
|
||||||
|
|
||||||
### Setup
|
|
||||||
|
|
||||||
In node1,
|
|
||||||
|
|
||||||
- 1 Git clone DevStack.
|
|
||||||
- 2 Git clone Tricircle, or just download devstack/local.conf.node_1.sample.
|
|
||||||
- 3 Copy devstack/local.conf.node_1.sample to DevStack folder and rename it to
|
|
||||||
local.conf, change password in the file if needed.
|
|
||||||
- 4 Change the following options according to your environment:
|
|
||||||
```
|
|
||||||
HOST_IP=10.250.201.24
|
|
||||||
- change to your management interface ip.
|
|
||||||
Q_ML2_PLUGIN_VLAN_TYPE_OPTIONS=(network_vlan_ranges=bridge:2001:3000)
|
|
||||||
- the format is (network_vlan_ranges=<physical network name>:<min vlan>:<max vlan>),
|
|
||||||
you can change physical network name, but remember to adapt your change to the
|
|
||||||
commands showed in this guide; also, change min vlan and max vlan to adapt the
|
|
||||||
vlan range your physical network supports.
|
|
||||||
OVS_BRIDGE_MAPPINGS=bridge:br-bridge
|
|
||||||
- the format is <physical network name>:<ovs bridge name>, you can change these names,
|
|
||||||
but remember to adapt your change to the commands showed in this guide.
|
|
||||||
Q_USE_PROVIDERNET_FOR_PUBLIC=True
|
|
||||||
- use this option if you would like to try L3 north-south networking.
|
|
||||||
```
|
|
||||||
- 5 Create OVS bridge and attach the VLAN network interface to it
|
|
||||||
```
|
|
||||||
sudo ovs-vsctl add-br br-bridge
|
|
||||||
sudo ovs-vsctl add-port br-bridge eth1
|
|
||||||
```
|
|
||||||
br-bridge is the OVS bridge name you configure on OVS_PHYSICAL_BRIDGE, eth1 is
|
|
||||||
the device name of your VLAN network interface
|
|
||||||
- 6 Run DevStack.
|
|
||||||
- 7 After DevStack successfully starts, begin to setup node2.
|
|
||||||
|
|
||||||
In node2,
|
|
||||||
|
|
||||||
- 1 Git clone DevStack.
|
|
||||||
- 2 Git clone Tricircle, or just download devstack/local.conf.node_2.sample.
|
|
||||||
- 3 Copy devstack/local.conf.node_2.sample to DevStack folder and rename it to
|
|
||||||
local.conf, change password in the file if needed.
|
|
||||||
- 4 Change the following options according to your environment:
|
|
||||||
```
|
|
||||||
HOST_IP=10.250.201.25
|
|
||||||
- change to your management interface ip.
|
|
||||||
KEYSTONE_SERVICE_HOST=10.250.201.24
|
|
||||||
- change to management interface ip of node1.
|
|
||||||
KEYSTONE_AUTH_HOST=10.250.201.24
|
|
||||||
- change to management interface ip of node1.
|
|
||||||
GLANCE_SERVICE_HOST=10.250.201.24
|
|
||||||
- change to management interface ip of node1.
|
|
||||||
Q_ML2_PLUGIN_VLAN_TYPE_OPTIONS=(network_vlan_ranges=bridge:2001:3000,extern:3001:4000)
|
|
||||||
- the format is (network_vlan_ranges=<physical network name>:<min vlan>:<max vlan>),
|
|
||||||
you can change physical network name, but remember to adapt your change to the
|
|
||||||
commands showed in this guide; also, change min vlan and max vlan to adapt the
|
|
||||||
vlan range your physical network supports.
|
|
||||||
OVS_BRIDGE_MAPPINGS=bridge:br-bridge,extern:br-ext
|
|
||||||
- the format is <physical network name>:<ovs bridge name>, you can change these names,
|
|
||||||
but remember to adapt your change to the commands showed in this guide.
|
|
||||||
Q_USE_PROVIDERNET_FOR_PUBLIC=True
|
|
||||||
- use this option if you would like to try L3 north-south networking.
|
|
||||||
```
|
|
||||||
In this guide, we define two physical networks in node2, one is "bridge" for
|
|
||||||
bridge network, the other one is "extern" for external network. If you do not
|
|
||||||
want to try L3 north-south networking, you can simply remove the "extern" part.
|
|
||||||
The external network type we use in the guide is vlan, if you want to use other
|
|
||||||
network type like flat, please refer to
|
|
||||||
[DevStack document](http://docs.openstack.org/developer/devstack/).
|
|
||||||
|
|
||||||
- 5 Create OVS bridge and attach the VLAN network interface to it
|
|
||||||
```
|
|
||||||
sudo ovs-vsctl add-br br-bridge
|
|
||||||
sudo ovs-vsctl add-port br-bridge eth1
|
|
||||||
```
|
|
||||||
br-bridge is the OVS bridge name you configure on OVS_PHYSICAL_BRIDGE, eth1 is
|
|
||||||
the device name of your VLAN network interface
|
|
||||||
- 6 Run DevStack.
|
|
||||||
- 7 After DevStack successfully starts, the setup is finished.
|
|
||||||
|
|
||||||
### How to play
|
|
||||||
All the following operations are performed in node1
|
|
||||||
|
|
||||||
- 1 Check if services have been correctly registered. Run "openstack endpoint
|
|
||||||
list" and you should get similar output as following:
|
|
||||||
```
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
| ID | Region | Service Name | Service Type |
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
| 1fadbddef9074f81b986131569c3741e | RegionOne | tricircle | Cascading |
|
|
||||||
| a5c5c37613244cbab96230d9051af1a5 | RegionOne | ec2 | ec2 |
|
|
||||||
| 809a3f7282f94c8e86f051e15988e6f5 | Pod2 | neutron | network |
|
|
||||||
| e6ad9acc51074f1290fc9d128d236bca | Pod1 | neutron | network |
|
|
||||||
| aee8a185fa6944b6860415a438c42c32 | RegionOne | keystone | identity |
|
|
||||||
| 280ebc45bf9842b4b4156eb5f8f9eaa4 | RegionOne | glance | image |
|
|
||||||
| aa54df57d7b942a1a327ed0722dba96e | Pod2 | nova_legacy | compute_legacy |
|
|
||||||
| aa25ae2a3f5a4e4d8bc0cae2f5fbb603 | Pod2 | nova | compute |
|
|
||||||
| 932550311ae84539987bfe9eb874dea3 | RegionOne | nova_legacy | compute_legacy |
|
|
||||||
| f89fbeffd7e446d0a552e2a6cf7be2ec | Pod1 | nova | compute |
|
|
||||||
| e2e19c164060456f8a1e75f8d3331f47 | Pod2 | ec2 | ec2 |
|
|
||||||
| de698ad5c6794edd91e69f0e57113e97 | RegionOne | nova | compute |
|
|
||||||
| 8a4b2332d2a4460ca3f740875236a967 | Pod2 | keystone | identity |
|
|
||||||
| b3ad80035f8742f29d12df67bdc2f70c | RegionOne | neutron | network |
|
|
||||||
+----------------------------------+-----------+--------------+----------------+
|
|
||||||
```
|
|
||||||
"RegionOne" is the region you set in local.conf via REGION_NAME in node1, whose
|
|
||||||
default value is "RegionOne", we use it as the region for Tricircle; "Pod1" is
|
|
||||||
the region set via POD_REGION_NAME, new configuration option introduced by
|
|
||||||
Tricircle, we use it as the bottom OpenStack; "Pod2" is the region you set via
|
|
||||||
REGION_NAME in node2, we use it as another bottom OpenStack. In node2, you also
|
|
||||||
need to set KEYSTONE_REGION_NAME the same as REGION_NAME in node1, which is
|
|
||||||
"RegionOne" in this example. So services in node2 can interact with Keystone
|
|
||||||
service in RegionOne.
|
|
||||||
- 2 Create pod instances for Tricircle and bottom OpenStack
|
|
||||||
```
|
|
||||||
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "RegionOne"}}'
|
|
||||||
|
|
||||||
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod1", "az_name": "az1"}}'
|
|
||||||
|
|
||||||
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod2", "az_name": "az2"}}'
|
|
||||||
```
|
|
||||||
- 3 Create network with AZ scheduler hints specified
|
|
||||||
```
|
|
||||||
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" \
|
|
||||||
-d '{"network": {"name": "net1", "admin_state_up": true, "availability_zone_hints": ["az1"]}}'
|
|
||||||
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" \
|
|
||||||
-d '{"network": {"name": "net2", "admin_state_up": true, "availability_zone_hints": ["az2"]}}'
|
|
||||||
```
|
|
||||||
Here we create two networks separately bound to Pod1 and Pod2
|
|
||||||
- 4 Create necessary resources to boot virtual machines.
|
|
||||||
```
|
|
||||||
nova flavor-create test 1 1024 10 1
|
|
||||||
neutron subnet-create net1 10.0.1.0/24
|
|
||||||
neutron subnet-create net2 10.0.2.0/24
|
|
||||||
glance image-list
|
|
||||||
```
|
|
||||||
- 5 Boot virtual machines.
|
|
||||||
```
|
|
||||||
nova boot --flavor 1 --image $image_id --nic net-id=$net1_id --availability-zone az1 vm1
|
|
||||||
nova boot --flavor 1 --image $image_id --nic net-id=$net2_id --availability-zone az2 vm2
|
|
||||||
```
|
|
||||||
- 6 Create router and attach interface
|
|
||||||
```
|
|
||||||
neutron router-create router
|
|
||||||
neutron router-interface-add router $subnet1_id
|
|
||||||
neutron router-interface-add router $subnet2_id
|
|
||||||
```
|
|
||||||
- 7 Launch VNC console anc check connectivity
|
|
||||||
By now, two networks are connected by the router, the two virtual machines
|
|
||||||
should be able to communicate with each other, we can launch a VNC console to
|
|
||||||
check. Currently Tricircle doesn't support VNC proxy, we need to go to bottom
|
|
||||||
OpenStack to obtain a VNC console.
|
|
||||||
```
|
|
||||||
nova --os-region-name Pod1 get-vnc-console vm1 novnc
|
|
||||||
nova --os-region-name Pod2 get-vnc-console vm2 novnc
|
|
||||||
```
|
|
||||||
Login one virtual machine via VNC and you should find it can "ping" the other
|
|
||||||
virtual machine. Default security group is applied so no need to configure
|
|
||||||
security group rule.
|
|
||||||
|
|
||||||
### North-South Networking
|
|
||||||
|
|
||||||
Before running DevStack in node2, you need to create another ovs bridge for
|
|
||||||
external network and then attach port.
|
|
||||||
```
|
|
||||||
sudo ovs-vsctl add-br br-ext
|
|
||||||
sudo ovs-vsctl add-port br-ext eth2
|
|
||||||
```
|
|
||||||
|
|
||||||
Below listed the operations related to north-south networking:
|
|
||||||
- 1 Create external network
|
|
||||||
```
|
|
||||||
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
|
||||||
-H "X-Auth-Token: $token" \
|
|
||||||
-d '{"network": {"name": "ext-net", "admin_state_up": true, "router:external": true, "provider:network_type": "vlan", "provider:physical_network": "extern", "availability_zone_hints": ["Pod2"]}}'
|
|
||||||
```
|
|
||||||
Pay attention that when creating external network, we still need to pass
|
|
||||||
"availability_zone_hints" parameter, but the value we pass is the name of pod,
|
|
||||||
not the name of availability zone.
|
|
||||||
> Currently external network needs to be created before attaching subnet to the
|
|
||||||
> router, because plugin needs to utilize external network information to setup
|
|
||||||
> bridge network when handling interface adding operation. This limitation will
|
|
||||||
> be removed later.
|
|
||||||
|
|
||||||
- 2 Create external subnet
|
|
||||||
```
|
|
||||||
neutron subnet-create --name ext-subnet --disable-dhcp ext-net 163.3.124.0/24
|
|
||||||
```
|
|
||||||
- 3 Set router external gateway
|
|
||||||
```
|
|
||||||
neutron router-gateway-set router ext-net
|
|
||||||
```
|
|
||||||
Now virtual machine in the subnet attached to the router should be able to
|
|
||||||
"ping" machines in the external network. In our test, we use hypervisor tool
|
|
||||||
to directly start a virtual machine in the external network to check the
|
|
||||||
network connectivity.
|
|
||||||
- 4 Create floating ip
|
|
||||||
```
|
|
||||||
neutron floatingip-create ext-net
|
|
||||||
```
|
|
||||||
- 5 Associate floating ip
|
|
||||||
```
|
|
||||||
neutron floatingip-list
|
|
||||||
neutron port-list
|
|
||||||
neutron floatingip-associate $floatingip_id $port_id
|
|
||||||
```
|
|
||||||
Now you should be able to access virtual machine with floating ip bound from
|
|
||||||
the external network.
|
|
||||||
|
|
||||||
### Quick verify
|
|
||||||
|
|
||||||
A sample of admin-openrc.sh and an installation verification script can be found
|
|
||||||
in devstack/ directory.
|
|
||||||
|
|
||||||
And a demo blog with virtualbox can be found in [this](http://shipengfei92.cn/play_tricircle_with_virtualbox).
|
|
||||||
|
|
||||||
#### verify_cross_pod_install.sh
|
|
||||||
|
|
||||||
This script is to quickly verify the installation of Tricircle in Cross Pod
|
|
||||||
OpenStack as the contents above and save the output to logs.
|
|
||||||
Before verifying the installation, some parameters should be modified to your own
|
|
||||||
environment.
|
|
||||||
|
|
||||||
- 1 The default URL is 127.0.0.1, change it if needed,
|
|
||||||
- 2 This script create a external network 10.50.11.0/26 according to the work environment,
|
|
||||||
change it if needed.
|
|
||||||
- 3 This script create 2 subnets 10.0.1.0/24 and 10.0.2.0/24, Change these if needed.
|
|
||||||
- 4 The default created floating-ip is attached to the VM with port 10.0.2.3 created by
|
|
||||||
the subnets, modify it according to your environment.
|
|
||||||
|
|
||||||
Then do the following steps in Node1 OpenStack to verify network functions:
|
|
||||||
|
|
||||||
```
|
|
||||||
cd tricircle/devstack/
|
|
||||||
./verify_cross_pod_install.sh 2>&1 | tee logs
|
|
||||||
```
|
|
37
README.rst
Normal file
37
README.rst
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
=========
|
||||||
|
Tricircle
|
||||||
|
=========
|
||||||
|
|
||||||
|
The Tricircle provides an OpenStack API gateway and networking automation
|
||||||
|
funtionality to allow multiple OpenStack instances, spanning in one site or
|
||||||
|
multiple sites or in hybrid cloud, to be managed as a single OpenStack cloud.
|
||||||
|
|
||||||
|
The Tricircle and these managed OpenStack instances will use shared KeyStone
|
||||||
|
(with centralized or distributed deployment) or federated KeyStones for
|
||||||
|
identity management.
|
||||||
|
|
||||||
|
The Tricircle presents one big region to the end user in KeyStone. And each
|
||||||
|
OpenStack instance called a pod is a sub-region of the Tricircle in
|
||||||
|
KeyStone, and usually not visible to end user directly.
|
||||||
|
|
||||||
|
The Tricircle acts as OpenStack API gateway, can handle OpenStack API calls,
|
||||||
|
schedule one proper OpenStack instance if needed during the API calls handling,
|
||||||
|
forward the API calls to the appropriate OpenStack instance, and deal with
|
||||||
|
tenant level L2/L3 networking across OpenStack instances automatically. So it
|
||||||
|
doesn't matter on which bottom OpenStack instance the VMs for the tenant are
|
||||||
|
running, they can communicate with each other via L2 or L3.
|
||||||
|
|
||||||
|
The end user can see avaialbility zone(AZ) and use AZ to provision
|
||||||
|
VM, Volume, even Network through the Tricircle. One AZ can include many
|
||||||
|
OpenStack instances, the Tricircle can schedule and bind OpenStack instance
|
||||||
|
for the tenant inside one AZ. A tenant's resources could be bound to multiple
|
||||||
|
specific bottom OpenStack instances in one or multiple AZs automatically.
|
||||||
|
|
||||||
|
* Free software: Apache license
|
||||||
|
* Design documentation: `Tricircle Design Blueprint <https://docs.google.com/document/d/18kZZ1snMOCD9IQvUKI5NVDzSASpw-QKj7l2zNqMEd3g/>`_
|
||||||
|
* Wiki: https://wiki.openstack.org/wiki/tricircle
|
||||||
|
* Installation with DevStack: https://github.com/openstack/tricircle/doc/source/installation.rst
|
||||||
|
* Documentation: http://docs.openstack.org/developer/tricircle
|
||||||
|
* Source: https://github.com/openstack/tricircle
|
||||||
|
* Bugs: http://bugs.launchpad.net/tricircle
|
||||||
|
* Blueprints: https://launchpad.net/tricircle
|
4
doc/source/contributing.rst
Normal file
4
doc/source/contributing.rst
Normal file
@ -0,0 +1,4 @@
|
|||||||
|
=============
|
||||||
|
Contributing
|
||||||
|
=============
|
||||||
|
.. include:: ../../CONTRIBUTING.rst
|
@ -11,7 +11,11 @@ Contents:
|
|||||||
.. toctree::
|
.. toctree::
|
||||||
:maxdepth: 2
|
:maxdepth: 2
|
||||||
|
|
||||||
|
readme
|
||||||
|
installation
|
||||||
|
usage
|
||||||
api_v1
|
api_v1
|
||||||
|
contributing
|
||||||
|
|
||||||
Indices and tables
|
Indices and tables
|
||||||
==================
|
==================
|
||||||
|
467
doc/source/installation.rst
Normal file
467
doc/source/installation.rst
Normal file
@ -0,0 +1,467 @@
|
|||||||
|
=====================
|
||||||
|
Installation with pip
|
||||||
|
=====================
|
||||||
|
|
||||||
|
At the command line::
|
||||||
|
|
||||||
|
$ pip install tricircle
|
||||||
|
|
||||||
|
Or, if you have virtualenvwrapper installed::
|
||||||
|
|
||||||
|
$ mkvirtualenv tricircle
|
||||||
|
$ pip install tricircle
|
||||||
|
|
||||||
|
|
||||||
|
======================================
|
||||||
|
Single node installation with DevStack
|
||||||
|
======================================
|
||||||
|
|
||||||
|
Now the Tricircle can be played with DevStack.
|
||||||
|
|
||||||
|
- 1 Install DevStack. Please refer to
|
||||||
|
http://docs.openstack.org/developer/devstack/
|
||||||
|
on how to install DevStack into single VM or physcial machine
|
||||||
|
- 2 In DevStack folder, create a file local.conf, and copy the content of
|
||||||
|
https://github.com/openstack/tricircle/blob/master/devstack/local.conf.sample
|
||||||
|
to local.conf, change password in the file if needed.
|
||||||
|
- 3 Run DevStack. In DevStack folder, run::
|
||||||
|
|
||||||
|
./stack.sh
|
||||||
|
|
||||||
|
- 4 In DevStack folder, create a file adminrc, and copy the content of
|
||||||
|
https://github.com/openstack/tricircle/blob/master/devstack/admin-openrc.sh
|
||||||
|
to the adminrc, change the password in the file if needed.
|
||||||
|
And run the following command to set the environment variables::
|
||||||
|
|
||||||
|
source adminrc
|
||||||
|
|
||||||
|
- 5 After DevStack successfully starts, check if services have been correctly
|
||||||
|
registered. Run "openstack endpoint list" and you should get output look
|
||||||
|
like as following::
|
||||||
|
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
| ID | Region | Service Name | Service Type |
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
| 230059e8533e4d389e034fd68257034b | RegionOne | glance | image |
|
||||||
|
| 25180a0a08cb41f69de52a7773452b28 | RegionOne | nova | compute |
|
||||||
|
| bd1ed1d6f0cc42398688a77bcc3bda91 | Pod1 | neutron | network |
|
||||||
|
| 673736f54ec147b79e97c395afe832f9 | RegionOne | ec2 | ec2 |
|
||||||
|
| fd7f188e2ba04ebd856d582828cdc50c | RegionOne | neutron | network |
|
||||||
|
| ffb56fd8b24a4a27bf6a707a7f78157f | RegionOne | keystone | identity |
|
||||||
|
| 88da40693bfa43b9b02e1478b1fa0bc6 | Pod1 | nova | compute |
|
||||||
|
| f35d64c2ddc44c16a4f9dfcd76e23d9f | RegionOne | nova_legacy | compute_legacy |
|
||||||
|
| 8759b2941fe7469e9651de3f6a123998 | RegionOne | tricircle | Cascading |
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
|
||||||
|
|
||||||
|
"RegionOne" is the region you set in local.conf via REGION_NAME, whose default
|
||||||
|
value is "RegionOne", we use it as the region for the Tricircle instance;
|
||||||
|
"Pod1" is the region set via "POD_REGION_NAME", new configuration option
|
||||||
|
introduced by the Tricircle, we use it as the bottom OpenStack instance.
|
||||||
|
- 6 Create pod instances for Tricircle and bottom OpenStack::
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "RegionOne"}}'
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod1", "az_name": "az1"}}'
|
||||||
|
|
||||||
|
Pay attention to "pod_name" parameter we specify when creating pod. Pod name
|
||||||
|
should exactly match the region name registered in Keystone since it is used
|
||||||
|
by the Tricircle to route API request. In the above commands, we create pods
|
||||||
|
named "RegionOne" and "Pod1" for the Tricircle instance and bottom OpenStack
|
||||||
|
instance. The Tricircle API service will automatically create an aggregate
|
||||||
|
when user creates a bottom pod, so command "nova aggregate-list" will show
|
||||||
|
the following result::
|
||||||
|
|
||||||
|
+----+----------+-------------------+
|
||||||
|
| Id | Name | Availability Zone |
|
||||||
|
+----+----------+-------------------+
|
||||||
|
| 1 | ag_Pod1 | az1 |
|
||||||
|
+----+----------+-------------------+
|
||||||
|
|
||||||
|
- 7 Create necessary resources to boot a virtual machine::
|
||||||
|
|
||||||
|
nova flavor-create test 1 1024 10 1
|
||||||
|
neutron net-create net1
|
||||||
|
neutron subnet-create net1 10.0.0.0/24
|
||||||
|
glance image-list
|
||||||
|
|
||||||
|
Note that flavor mapping has not been implemented yet so the created flavor
|
||||||
|
is just record saved in database as metadata. Actual flavor is saved in
|
||||||
|
bottom OpenStack instance.
|
||||||
|
- 8 Boot a virtual machine::
|
||||||
|
|
||||||
|
nova boot --flavor 1 --image $image_id --nic net-id=$net_id --availability-zone az1 vm1
|
||||||
|
|
||||||
|
- 9 Create, list, show and delete volume::
|
||||||
|
|
||||||
|
cinder --debug create --availability-zone=az1 1
|
||||||
|
cinder --debug list
|
||||||
|
cinder --debug show $volume_id
|
||||||
|
cinder --debug delete $volume_id
|
||||||
|
cinder --debug list
|
||||||
|
|
||||||
|
Verification with script
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
|
||||||
|
A sample of admin-openrc.sh and an installation verification script can be found
|
||||||
|
in devstack/ in the Tricircle root folder. 'admin-openrc.sh' is used to create
|
||||||
|
environment variables for the admin user as the following::
|
||||||
|
|
||||||
|
export OS_PROJECT_DOMAIN_ID=default
|
||||||
|
export OS_USER_DOMAIN_ID=default
|
||||||
|
export OS_PROJECT_NAME=admin
|
||||||
|
export OS_TENANT_NAME=admin
|
||||||
|
export OS_USERNAME=admin
|
||||||
|
export OS_PASSWORD=password #change password as you set in your own environment
|
||||||
|
export OS_AUTH_URL=http://127.0.0.1:5000
|
||||||
|
export OS_IDENTITY_API_VERSION=3
|
||||||
|
export OS_IMAGE_API_VERSION=2
|
||||||
|
export OS_REGION_NAME=RegionOne
|
||||||
|
|
||||||
|
The command to use the admin-openrc.sh is::
|
||||||
|
|
||||||
|
source tricircle/devstack/admin-openrc.sh
|
||||||
|
|
||||||
|
'verify_top_install.sh' script is to quickly verify the installation of
|
||||||
|
the Tricircle in Top OpenStack as the step 5-9 above and save the output
|
||||||
|
to logs.
|
||||||
|
|
||||||
|
Before verifying the installation, you should modify the script based on your
|
||||||
|
own environment.
|
||||||
|
|
||||||
|
- 1 The default post URL is 127.0.0.1, change it if needed,
|
||||||
|
- 2 The default create net1's networ address is 10.0.0.0/24, change it if
|
||||||
|
needed.
|
||||||
|
|
||||||
|
Then you do the following steps to verify::
|
||||||
|
|
||||||
|
cd tricircle/devstack/
|
||||||
|
./verify_top_install.sh 2>&1 | tee logs
|
||||||
|
|
||||||
|
|
||||||
|
======================================================================
|
||||||
|
Two nodes installation with DevStack for Cross-OpenStack L3 networking
|
||||||
|
======================================================================
|
||||||
|
|
||||||
|
Introduction
|
||||||
|
^^^^^^^^^^^^
|
||||||
|
|
||||||
|
Now the Tricircle supports cross-pod l3 networking.
|
||||||
|
|
||||||
|
To achieve cross-pod l3 networking, Tricircle utilizes a shared provider VLAN
|
||||||
|
network at first phase. We are considering later using DCI controller to create
|
||||||
|
a multi-segment VLAN network, VxLAN network for L3 networking purpose. When a
|
||||||
|
subnet is attached to a router in top pod, Tricircle not only creates
|
||||||
|
corresponding subnet and router in bottom pod, but also creates a VLAN type
|
||||||
|
"bridge" network. Both tenant network and "bridge" network are attached to
|
||||||
|
bottom router. Each tenant will have one allocated VLAN, which is shared by
|
||||||
|
the tenant's "bridge" networks across bottom pods. The CIDRs of "bridge"
|
||||||
|
networks for one tenant are also the same, so the router interfaces in
|
||||||
|
"bridge" networks across different bottom pods can communicate with each
|
||||||
|
other via the provider VLAN network. By adding an extra route as following::
|
||||||
|
|
||||||
|
destination: CIDR of tenant network in another bottom pod
|
||||||
|
nexthop: "bridge" network interface ip in another bottom pod
|
||||||
|
|
||||||
|
when a server sends a packet whose receiver is in another network and in
|
||||||
|
another bottom pod, the packet first goes to router namespace, then is
|
||||||
|
forwarded to the router namespace in another bottom pod according to the extra
|
||||||
|
route, at last the packet is sent to the target server. This configuration job
|
||||||
|
is triggered when user attaches a subnet to a router in top pod and finished
|
||||||
|
asynchronously.
|
||||||
|
|
||||||
|
Currently cross-pod L2 networking is not supported yet, so tenant networks
|
||||||
|
cannot cross pods, that is to say, one network in top pod can only locate in
|
||||||
|
one bottom pod, tenant network is bound to bottom pod. Otherwise we cannot
|
||||||
|
correctly configure extra route since for one destination CIDR, we have more
|
||||||
|
than one possible nexthop addresses.
|
||||||
|
|
||||||
|
*When cross-pod L2 networking is introduced, L2GW will be used to connect L2
|
||||||
|
network in different pods. No extra route is required to connect L2 network
|
||||||
|
All L3 traffic will be forwarded to the local L2 network, then go to the
|
||||||
|
server in another pod via the L2GW.*
|
||||||
|
|
||||||
|
We use "availability_zone_hints" attribute for user to specify the bottom pod
|
||||||
|
he wants to create the bottom network. Currently we do not support attaching
|
||||||
|
a network to a router without setting "availability_zone_hints" attribute of
|
||||||
|
the network.
|
||||||
|
|
||||||
|
Prerequisite
|
||||||
|
^^^^^^^^^^^^
|
||||||
|
|
||||||
|
To play cross-pod L3 networking, two nodes are needed. One to run Tricircle
|
||||||
|
and one bottom pod, the other one to run another bottom pod. Both nodes have
|
||||||
|
two network interfaces, for management and provider VLAN network. For VLAN
|
||||||
|
network, the physical network infrastructure should support VLAN tagging. If
|
||||||
|
you would like to try north-south networking, too, you should prepare one more
|
||||||
|
network interface in the second node for external network. In this guide, the
|
||||||
|
external network is also vlan type, so the local.conf sample is based on vlan
|
||||||
|
type external network setup.
|
||||||
|
|
||||||
|
Setup
|
||||||
|
^^^^^
|
||||||
|
In node1,
|
||||||
|
|
||||||
|
- 1 Git clone DevStack.
|
||||||
|
- 2 Git clone Tricircle, or just download devstack/local.conf.node_1.sample.
|
||||||
|
- 3 Copy devstack/local.conf.node_1.sample to DevStack folder and rename it to
|
||||||
|
local.conf, change password in the file if needed.
|
||||||
|
- 4 Change the following options according to your environment::
|
||||||
|
|
||||||
|
HOST_IP=10.250.201.24
|
||||||
|
|
||||||
|
change to your management interface ip::
|
||||||
|
|
||||||
|
Q_ML2_PLUGIN_VLAN_TYPE_OPTIONS=(network_vlan_ranges=bridge:2001:3000)
|
||||||
|
|
||||||
|
the format is (network_vlan_ranges=<physical network name>:<min vlan>:<max vlan>),
|
||||||
|
you can change physical network name, but remember to adapt your change
|
||||||
|
to the commands showed in this guide; also, change min vlan and max vlan
|
||||||
|
to adapt the vlan range your physical network supports::
|
||||||
|
|
||||||
|
OVS_BRIDGE_MAPPINGS=bridge:br-bridge
|
||||||
|
|
||||||
|
the format is <physical network name>:<ovs bridge name>, you can change
|
||||||
|
these names, but remember to adapt your change to the commands showed in
|
||||||
|
this guide::
|
||||||
|
|
||||||
|
Q_USE_PROVIDERNET_FOR_PUBLIC=True
|
||||||
|
|
||||||
|
use this option if you would like to try L3 north-south networking.
|
||||||
|
|
||||||
|
|
||||||
|
- 5 Create OVS bridge and attach the VLAN network interface to it::
|
||||||
|
|
||||||
|
sudo ovs-vsctl add-br br-bridge
|
||||||
|
sudo ovs-vsctl add-port br-bridge eth1
|
||||||
|
|
||||||
|
br-bridge is the OVS bridge name you configure on OVS_PHYSICAL_BRIDGE, eth1 is
|
||||||
|
the device name of your VLAN network interface
|
||||||
|
- 6 Run DevStack.
|
||||||
|
- 7 After DevStack successfully starts, begin to setup node2.
|
||||||
|
|
||||||
|
In node2,
|
||||||
|
|
||||||
|
- 1 Git clone DevStack.
|
||||||
|
- 2 Git clone Tricircle, or just download devstack/local.conf.node_2.sample.
|
||||||
|
- 3 Copy devstack/local.conf.node_2.sample to DevStack folder and rename it to
|
||||||
|
local.conf, change password in the file if needed.
|
||||||
|
- 4 Change the following options according to your environment::
|
||||||
|
|
||||||
|
HOST_IP=10.250.201.25
|
||||||
|
|
||||||
|
change to your management interface ip::
|
||||||
|
|
||||||
|
KEYSTONE_SERVICE_HOST=10.250.201.24
|
||||||
|
|
||||||
|
change to management interface ip of node1::
|
||||||
|
|
||||||
|
KEYSTONE_AUTH_HOST=10.250.201.24
|
||||||
|
|
||||||
|
change to management interface ip of node1::
|
||||||
|
|
||||||
|
GLANCE_SERVICE_HOST=10.250.201.24
|
||||||
|
|
||||||
|
change to management interface ip of node1::
|
||||||
|
|
||||||
|
Q_ML2_PLUGIN_VLAN_TYPE_OPTIONS=(network_vlan_ranges=bridge:2001:3000,extern:3001:4000)
|
||||||
|
|
||||||
|
the format is (network_vlan_ranges=<physical network name>:<min vlan>:<max vlan>),
|
||||||
|
you can change physical network name, but remember to adapt your change
|
||||||
|
to the commands showed in this guide; also, change min vlan and max vlan
|
||||||
|
to adapt the vlan range your physical network supports::
|
||||||
|
|
||||||
|
OVS_BRIDGE_MAPPINGS=bridge:br-bridge,extern:br-ext
|
||||||
|
|
||||||
|
the format is <physical network name>:<ovs bridge name>, you can change
|
||||||
|
these names, but remember to adapt your change to the commands showed in
|
||||||
|
this guide::
|
||||||
|
|
||||||
|
Q_USE_PROVIDERNET_FOR_PUBLIC=True
|
||||||
|
|
||||||
|
use this option if you would like to try L3 north-south networking.
|
||||||
|
|
||||||
|
In this guide, we define two physical networks in node2, one is "bridge" for
|
||||||
|
bridge network, the other one is "extern" for external network. If you do not
|
||||||
|
want to try L3 north-south networking, you can simply remove the "extern" part.
|
||||||
|
The external network type we use in the guide is vlan, if you want to use other
|
||||||
|
network type like flat, please refer to
|
||||||
|
[DevStack document](http://docs.openstack.org/developer/devstack/).
|
||||||
|
|
||||||
|
- 5 Create OVS bridge and attach the VLAN network interface to it::
|
||||||
|
|
||||||
|
sudo ovs-vsctl add-br br-bridge
|
||||||
|
sudo ovs-vsctl add-port br-bridge eth1
|
||||||
|
|
||||||
|
br-bridge is the OVS bridge name you configure on OVS_PHYSICAL_BRIDGE, eth1 is
|
||||||
|
the device name of your VLAN network interface
|
||||||
|
- 6 Run DevStack.
|
||||||
|
- 7 After DevStack successfully starts, the setup is finished.
|
||||||
|
|
||||||
|
How to play
|
||||||
|
^^^^^^^^^^^
|
||||||
|
|
||||||
|
All the following operations are performed in node1
|
||||||
|
|
||||||
|
- 1 Check if services have been correctly registered. Run "openstack endpoint
|
||||||
|
list" and you should get similar output as following::
|
||||||
|
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
| ID | Region | Service Name | Service Type |
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
| 1fadbddef9074f81b986131569c3741e | RegionOne | tricircle | Cascading |
|
||||||
|
| a5c5c37613244cbab96230d9051af1a5 | RegionOne | ec2 | ec2 |
|
||||||
|
| 809a3f7282f94c8e86f051e15988e6f5 | Pod2 | neutron | network |
|
||||||
|
| e6ad9acc51074f1290fc9d128d236bca | Pod1 | neutron | network |
|
||||||
|
| aee8a185fa6944b6860415a438c42c32 | RegionOne | keystone | identity |
|
||||||
|
| 280ebc45bf9842b4b4156eb5f8f9eaa4 | RegionOne | glance | image |
|
||||||
|
| aa54df57d7b942a1a327ed0722dba96e | Pod2 | nova_legacy | compute_legacy |
|
||||||
|
| aa25ae2a3f5a4e4d8bc0cae2f5fbb603 | Pod2 | nova | compute |
|
||||||
|
| 932550311ae84539987bfe9eb874dea3 | RegionOne | nova_legacy | compute_legacy |
|
||||||
|
| f89fbeffd7e446d0a552e2a6cf7be2ec | Pod1 | nova | compute |
|
||||||
|
| e2e19c164060456f8a1e75f8d3331f47 | Pod2 | ec2 | ec2 |
|
||||||
|
| de698ad5c6794edd91e69f0e57113e97 | RegionOne | nova | compute |
|
||||||
|
| 8a4b2332d2a4460ca3f740875236a967 | Pod2 | keystone | identity |
|
||||||
|
| b3ad80035f8742f29d12df67bdc2f70c | RegionOne | neutron | network |
|
||||||
|
+----------------------------------+-----------+--------------+----------------+
|
||||||
|
|
||||||
|
"RegionOne" is the region you set in local.conf via REGION_NAME in node1, whose
|
||||||
|
default value is "RegionOne", we use it as the region for Tricircle; "Pod1" is
|
||||||
|
the region set via POD_REGION_NAME, new configuration option introduced by
|
||||||
|
Tricircle, we use it as the bottom OpenStack; "Pod2" is the region you set via
|
||||||
|
REGION_NAME in node2, we use it as another bottom OpenStack. In node2, you also
|
||||||
|
need to set KEYSTONE_REGION_NAME the same as REGION_NAME in node1, which is
|
||||||
|
"RegionOne" in this example. So services in node2 can interact with Keystone
|
||||||
|
service in RegionOne.
|
||||||
|
- 2 Create pod instances for Tricircle and bottom OpenStack::
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "RegionOne"}}'
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod1", "az_name": "az1"}}'
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:19999/v1.0/pods -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" -d '{"pod": {"pod_name": "Pod2", "az_name": "az2"}}'
|
||||||
|
|
||||||
|
- 3 Create network with AZ scheduler hints specified::
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" \
|
||||||
|
-d '{"network": {"name": "net1", "admin_state_up": true, "availability_zone_hints": ["az1"]}}'
|
||||||
|
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" \
|
||||||
|
-d '{"network": {"name": "net2", "admin_state_up": true, "availability_zone_hints": ["az2"]}}'
|
||||||
|
|
||||||
|
Here we create two networks separately bound to Pod1 and Pod2
|
||||||
|
- 4 Create necessary resources to boot virtual machines::
|
||||||
|
|
||||||
|
nova flavor-create test 1 1024 10 1
|
||||||
|
neutron subnet-create net1 10.0.1.0/24
|
||||||
|
neutron subnet-create net2 10.0.2.0/24
|
||||||
|
glance image-list
|
||||||
|
|
||||||
|
- 5 Boot virtual machines::
|
||||||
|
|
||||||
|
nova boot --flavor 1 --image $image_id --nic net-id=$net1_id --availability-zone az1 vm1
|
||||||
|
nova boot --flavor 1 --image $image_id --nic net-id=$net2_id --availability-zone az2 vm2
|
||||||
|
|
||||||
|
- 6 Create router and attach interface::
|
||||||
|
|
||||||
|
neutron router-create router
|
||||||
|
neutron router-interface-add router $subnet1_id
|
||||||
|
neutron router-interface-add router $subnet2_id
|
||||||
|
|
||||||
|
- 7 Launch VNC console anc check connectivity
|
||||||
|
By now, two networks are connected by the router, the two virtual machines
|
||||||
|
should be able to communicate with each other, we can launch a VNC console to
|
||||||
|
check. Currently Tricircle doesn't support VNC proxy, we need to go to bottom
|
||||||
|
OpenStack to obtain a VNC console::
|
||||||
|
|
||||||
|
nova --os-region-name Pod1 get-vnc-console vm1 novnc
|
||||||
|
nova --os-region-name Pod2 get-vnc-console vm2 novnc
|
||||||
|
|
||||||
|
Login one virtual machine via VNC and you should find it can "ping" the other
|
||||||
|
virtual machine. Default security group is applied so no need to configure
|
||||||
|
security group rule.
|
||||||
|
|
||||||
|
North-South Networking
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
Before running DevStack in node2, you need to create another ovs bridge for
|
||||||
|
external network and then attach port::
|
||||||
|
|
||||||
|
sudo ovs-vsctl add-br br-ext
|
||||||
|
sudo ovs-vsctl add-port br-ext eth2
|
||||||
|
|
||||||
|
Below listed the operations related to north-south networking.
|
||||||
|
|
||||||
|
- 1 Create external network::
|
||||||
|
|
||||||
|
curl -X POST http://127.0.0.1:9696/v2.0/networks -H "Content-Type: application/json" \
|
||||||
|
-H "X-Auth-Token: $token" \
|
||||||
|
-d '{"network": {"name": "ext-net", "admin_state_up": true, "router:external": true, "provider:network_type": "vlan", "provider:physical_network": "extern", "availability_zone_hints": ["Pod2"]}}'
|
||||||
|
|
||||||
|
Pay attention that when creating external network, we still need to pass
|
||||||
|
"availability_zone_hints" parameter, but the value we pass is the name of pod,
|
||||||
|
not the name of availability zone.
|
||||||
|
|
||||||
|
*Currently external network needs to be created before attaching subnet to the
|
||||||
|
router, because plugin needs to utilize external network information to setup
|
||||||
|
bridge network when handling interface adding operation. This limitation will
|
||||||
|
be removed later.*
|
||||||
|
|
||||||
|
- 2 Create external subnet::
|
||||||
|
|
||||||
|
neutron subnet-create --name ext-subnet --disable-dhcp ext-net 163.3.124.0/24
|
||||||
|
|
||||||
|
- 3 Set router external gateway::
|
||||||
|
|
||||||
|
neutron router-gateway-set router ext-net
|
||||||
|
|
||||||
|
Now virtual machine in the subnet attached to the router should be able to
|
||||||
|
"ping" machines in the external network. In our test, we use hypervisor tool
|
||||||
|
to directly start a virtual machine in the external network to check the
|
||||||
|
network connectivity.
|
||||||
|
|
||||||
|
- 4 Create floating ip::
|
||||||
|
|
||||||
|
neutron floatingip-create ext-net
|
||||||
|
|
||||||
|
- 5 Associate floating ip::
|
||||||
|
|
||||||
|
neutron floatingip-list
|
||||||
|
neutron port-list
|
||||||
|
neutron floatingip-associate $floatingip_id $port_id
|
||||||
|
|
||||||
|
Now you should be able to access virtual machine with floating ip bound from
|
||||||
|
the external network.
|
||||||
|
|
||||||
|
Verification with script
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
|
||||||
|
A sample of admin-openrc.sh and an installation verification script can be
|
||||||
|
found in devstack/ directory. And a demo blog with virtualbox can be found in http://shipengfei92.cn/play_tricircle_with_virtualbox
|
||||||
|
|
||||||
|
Script 'verify_cross_pod_install.sh' is to quickly verify the installation of
|
||||||
|
the Tricircle in Cross Pod OpenStack as the contents above and save the output
|
||||||
|
to logs.
|
||||||
|
Before verifying the installation, some parameters should be modified to your
|
||||||
|
own environment.
|
||||||
|
|
||||||
|
- 1 The default URL is 127.0.0.1, change it if needed,
|
||||||
|
- 2 This script create a external network 10.50.11.0/26 according to the work
|
||||||
|
environment, change it if needed.
|
||||||
|
- 3 This script create 2 subnets 10.0.1.0/24 and 10.0.2.0/24, Change these if
|
||||||
|
needed.
|
||||||
|
- 4 The default created floating-ip is attached to the VM with port 10.0.2.3
|
||||||
|
created by the subnets, modify it according to your environment.
|
||||||
|
|
||||||
|
Then do the following steps in Node1 OpenStack to verify network functions::
|
||||||
|
|
||||||
|
cd tricircle/devstack/
|
||||||
|
./verify_cross_pod_install.sh 2>&1 | tee logs
|
||||||
|
|
1
doc/source/readme.rst
Normal file
1
doc/source/readme.rst
Normal file
@ -0,0 +1 @@
|
|||||||
|
.. include:: ../../README.rst
|
7
doc/source/usage.rst
Normal file
7
doc/source/usage.rst
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
======
|
||||||
|
Usage
|
||||||
|
======
|
||||||
|
|
||||||
|
To use tricircle in a project::
|
||||||
|
|
||||||
|
import tricircle
|
0
releasenotes/notes/.placeholder
Normal file
0
releasenotes/notes/.placeholder
Normal file
0
releasenotes/source/_static/.placeholder
Normal file
0
releasenotes/source/_static/.placeholder
Normal file
0
releasenotes/source/_templates/.placeholder
Normal file
0
releasenotes/source/_templates/.placeholder
Normal file
272
releasenotes/source/conf.py
Normal file
272
releasenotes/source/conf.py
Normal file
@ -0,0 +1,272 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
# implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
# Glance Release Notes documentation build configuration file, created by
|
||||||
|
# sphinx-quickstart on Tue Nov 3 17:40:50 2015.
|
||||||
|
#
|
||||||
|
# This file is execfile()d with the current directory set to its
|
||||||
|
# containing dir.
|
||||||
|
#
|
||||||
|
# Note that not all possible configuration values are present in this
|
||||||
|
# autogenerated file.
|
||||||
|
#
|
||||||
|
# All configuration values have a default; values that are commented out
|
||||||
|
# serve to show the default.
|
||||||
|
|
||||||
|
# If extensions (or modules to document with autodoc) are in another directory,
|
||||||
|
# add these directories to sys.path here. If the directory is relative to the
|
||||||
|
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||||
|
# sys.path.insert(0, os.path.abspath('.'))
|
||||||
|
|
||||||
|
# -- General configuration ------------------------------------------------
|
||||||
|
|
||||||
|
# If your documentation needs a minimal Sphinx version, state it here.
|
||||||
|
# needs_sphinx = '1.0'
|
||||||
|
|
||||||
|
# Add any Sphinx extension module names here, as strings. They can be
|
||||||
|
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
|
||||||
|
# ones.
|
||||||
|
extensions = [
|
||||||
|
'oslosphinx',
|
||||||
|
'reno.sphinxext',
|
||||||
|
]
|
||||||
|
|
||||||
|
# Add any paths that contain templates here, relative to this directory.
|
||||||
|
templates_path = ['_templates']
|
||||||
|
|
||||||
|
# The suffix of source filenames.
|
||||||
|
source_suffix = '.rst'
|
||||||
|
|
||||||
|
# The encoding of source files.
|
||||||
|
# source_encoding = 'utf-8-sig'
|
||||||
|
|
||||||
|
# The master toctree document.
|
||||||
|
master_doc = 'index'
|
||||||
|
|
||||||
|
# General information about the project.
|
||||||
|
project = u'The Tricircle Release Notes'
|
||||||
|
copyright = u'2016, OpenStack Foundation'
|
||||||
|
|
||||||
|
# The version info for the project you're documenting, acts as replacement for
|
||||||
|
# |version| and |release|, also used in various other places throughout the
|
||||||
|
# built documents.
|
||||||
|
#
|
||||||
|
# The short X.Y version.
|
||||||
|
# The full version, including alpha/beta/rc tags.
|
||||||
|
release = ''
|
||||||
|
# The short X.Y version.
|
||||||
|
version = ''
|
||||||
|
|
||||||
|
# The language for content autogenerated by Sphinx. Refer to documentation
|
||||||
|
# for a list of supported languages.
|
||||||
|
# language = None
|
||||||
|
|
||||||
|
# There are two options for replacing |today|: either, you set today to some
|
||||||
|
# non-false value, then it is used:
|
||||||
|
# today = ''
|
||||||
|
# Else, today_fmt is used as the format for a strftime call.
|
||||||
|
# today_fmt = '%B %d, %Y'
|
||||||
|
|
||||||
|
# List of patterns, relative to source directory, that match files and
|
||||||
|
# directories to ignore when looking for source files.
|
||||||
|
exclude_patterns = []
|
||||||
|
|
||||||
|
# The reST default role (used for this markup: `text`) to use for all
|
||||||
|
# documents.
|
||||||
|
# default_role = None
|
||||||
|
|
||||||
|
# If true, '()' will be appended to :func: etc. cross-reference text.
|
||||||
|
# add_function_parentheses = True
|
||||||
|
|
||||||
|
# If true, the current module name will be prepended to all description
|
||||||
|
# unit titles (such as .. function::).
|
||||||
|
# add_module_names = True
|
||||||
|
|
||||||
|
# If true, sectionauthor and moduleauthor directives will be shown in the
|
||||||
|
# output. They are ignored by default.
|
||||||
|
# show_authors = False
|
||||||
|
|
||||||
|
# The name of the Pygments (syntax highlighting) style to use.
|
||||||
|
pygments_style = 'sphinx'
|
||||||
|
|
||||||
|
# A list of ignored prefixes for module index sorting.
|
||||||
|
# modindex_common_prefix = []
|
||||||
|
|
||||||
|
# If true, keep warnings as "system message" paragraphs in the built documents.
|
||||||
|
# keep_warnings = False
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for HTML output ----------------------------------------------
|
||||||
|
|
||||||
|
# The theme to use for HTML and HTML Help pages. See the documentation for
|
||||||
|
# a list of builtin themes.
|
||||||
|
html_theme = 'default'
|
||||||
|
|
||||||
|
# Theme options are theme-specific and customize the look and feel of a theme
|
||||||
|
# further. For a list of options available for each theme, see the
|
||||||
|
# documentation.
|
||||||
|
# html_theme_options = {}
|
||||||
|
|
||||||
|
# Add any paths that contain custom themes here, relative to this directory.
|
||||||
|
# html_theme_path = []
|
||||||
|
|
||||||
|
# The name for this set of Sphinx documents. If None, it defaults to
|
||||||
|
# "<project> v<release> documentation".
|
||||||
|
# html_title = None
|
||||||
|
|
||||||
|
# A shorter title for the navigation bar. Default is the same as html_title.
|
||||||
|
# html_short_title = None
|
||||||
|
|
||||||
|
# The name of an image file (relative to this directory) to place at the top
|
||||||
|
# of the sidebar.
|
||||||
|
# html_logo = None
|
||||||
|
|
||||||
|
# The name of an image file (within the static path) to use as favicon of the
|
||||||
|
# docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32
|
||||||
|
# pixels large.
|
||||||
|
# html_favicon = None
|
||||||
|
|
||||||
|
# Add any paths that contain custom static files (such as style sheets) here,
|
||||||
|
# relative to this directory. They are copied after the builtin static files,
|
||||||
|
# so a file named "default.css" will overwrite the builtin "default.css".
|
||||||
|
html_static_path = ['_static']
|
||||||
|
|
||||||
|
# Add any extra paths that contain custom files (such as robots.txt or
|
||||||
|
# .htaccess) here, relative to this directory. These files are copied
|
||||||
|
# directly to the root of the documentation.
|
||||||
|
# html_extra_path = []
|
||||||
|
|
||||||
|
# If not '', a 'Last updated on:' timestamp is inserted at every page bottom,
|
||||||
|
# using the given strftime format.
|
||||||
|
# html_last_updated_fmt = '%b %d, %Y'
|
||||||
|
|
||||||
|
# If true, SmartyPants will be used to convert quotes and dashes to
|
||||||
|
# typographically correct entities.
|
||||||
|
# html_use_smartypants = True
|
||||||
|
|
||||||
|
# Custom sidebar templates, maps document names to template names.
|
||||||
|
# html_sidebars = {}
|
||||||
|
|
||||||
|
# Additional templates that should be rendered to pages, maps page names to
|
||||||
|
# template names.
|
||||||
|
# html_additional_pages = {}
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
# html_domain_indices = True
|
||||||
|
|
||||||
|
# If false, no index is generated.
|
||||||
|
# html_use_index = True
|
||||||
|
|
||||||
|
# If true, the index is split into individual pages for each letter.
|
||||||
|
# html_split_index = False
|
||||||
|
|
||||||
|
# If true, links to the reST sources are added to the pages.
|
||||||
|
# html_show_sourcelink = True
|
||||||
|
|
||||||
|
# If true, "Created using Sphinx" is shown in the HTML footer. Default is True.
|
||||||
|
# html_show_sphinx = True
|
||||||
|
|
||||||
|
# If true, "(C) Copyright ..." is shown in the HTML footer. Default is True.
|
||||||
|
# html_show_copyright = True
|
||||||
|
|
||||||
|
# If true, an OpenSearch description file will be output, and all pages will
|
||||||
|
# contain a <link> tag referring to it. The value of this option must be the
|
||||||
|
# base URL from which the finished HTML is served.
|
||||||
|
# html_use_opensearch = ''
|
||||||
|
|
||||||
|
# This is the file name suffix for HTML files (e.g. ".xhtml").
|
||||||
|
# html_file_suffix = None
|
||||||
|
|
||||||
|
# Output file base name for HTML help builder.
|
||||||
|
htmlhelp_basename = 'GlanceReleaseNotesdoc'
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for LaTeX output ---------------------------------------------
|
||||||
|
|
||||||
|
latex_elements = {
|
||||||
|
# The paper size ('letterpaper' or 'a4paper').
|
||||||
|
# 'papersize': 'letterpaper',
|
||||||
|
|
||||||
|
# The font size ('10pt', '11pt' or '12pt').
|
||||||
|
# 'pointsize': '10pt',
|
||||||
|
|
||||||
|
# Additional stuff for the LaTeX preamble.
|
||||||
|
# 'preamble': '',
|
||||||
|
}
|
||||||
|
|
||||||
|
# Grouping the document tree into LaTeX files. List of tuples
|
||||||
|
# (source start file, target name, title,
|
||||||
|
# author, documentclass [howto, manual, or own class]).
|
||||||
|
latex_documents = [
|
||||||
|
('index', 'GlanceReleaseNotes.tex', u'Glance Release Notes Documentation',
|
||||||
|
u'Glance Developers', 'manual'),
|
||||||
|
]
|
||||||
|
|
||||||
|
# The name of an image file (relative to this directory) to place at the top of
|
||||||
|
# the title page.
|
||||||
|
# latex_logo = None
|
||||||
|
|
||||||
|
# For "manual" documents, if this is true, then toplevel headings are parts,
|
||||||
|
# not chapters.
|
||||||
|
# latex_use_parts = False
|
||||||
|
|
||||||
|
# If true, show page references after internal links.
|
||||||
|
# latex_show_pagerefs = False
|
||||||
|
|
||||||
|
# If true, show URL addresses after external links.
|
||||||
|
# latex_show_urls = False
|
||||||
|
|
||||||
|
# Documents to append as an appendix to all manuals.
|
||||||
|
# latex_appendices = []
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
# latex_domain_indices = True
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for manual page output ---------------------------------------
|
||||||
|
|
||||||
|
# One entry per manual page. List of tuples
|
||||||
|
# (source start file, name, description, authors, manual section).
|
||||||
|
man_pages = [
|
||||||
|
('index', 'glancereleasenotes', u'Glance Release Notes Documentation',
|
||||||
|
[u'Glance Developers'], 1)
|
||||||
|
]
|
||||||
|
|
||||||
|
# If true, show URL addresses after external links.
|
||||||
|
# man_show_urls = False
|
||||||
|
|
||||||
|
|
||||||
|
# -- Options for Texinfo output -------------------------------------------
|
||||||
|
|
||||||
|
# Grouping the document tree into Texinfo files. List of tuples
|
||||||
|
# (source start file, target name, title, author,
|
||||||
|
# dir menu entry, description, category)
|
||||||
|
texinfo_documents = [
|
||||||
|
('index', 'GlanceReleaseNotes', u'Glance Release Notes Documentation',
|
||||||
|
u'Glance Developers', 'GlanceReleaseNotes',
|
||||||
|
'One line description of project.',
|
||||||
|
'Miscellaneous'),
|
||||||
|
]
|
||||||
|
|
||||||
|
# Documents to append as an appendix to all manuals.
|
||||||
|
# texinfo_appendices = []
|
||||||
|
|
||||||
|
# If false, no module index is generated.
|
||||||
|
# texinfo_domain_indices = True
|
||||||
|
|
||||||
|
# How to display URL addresses: 'footnote', 'no', or 'inline'.
|
||||||
|
# texinfo_show_urls = 'footnote'
|
||||||
|
|
||||||
|
# If true, do not generate a @detailmenu in the "Top" node's menu.
|
||||||
|
# texinfo_no_detailmenu = False
|
9
releasenotes/source/index.rst
Normal file
9
releasenotes/source/index.rst
Normal file
@ -0,0 +1,9 @@
|
|||||||
|
======================
|
||||||
|
Cinder Release Notes
|
||||||
|
======================
|
||||||
|
|
||||||
|
.. toctree::
|
||||||
|
:maxdepth: 1
|
||||||
|
|
||||||
|
mitaka
|
||||||
|
unreleased
|
6
releasenotes/source/mitaka.rst
Normal file
6
releasenotes/source/mitaka.rst
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
=============================
|
||||||
|
Mitaka Series Release Notes
|
||||||
|
=============================
|
||||||
|
|
||||||
|
.. release-notes::
|
||||||
|
:branch: origin/stable/mitaka
|
5
releasenotes/source/unreleased.rst
Normal file
5
releasenotes/source/unreleased.rst
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
==============================
|
||||||
|
Current Series Release Notes
|
||||||
|
==============================
|
||||||
|
|
||||||
|
.. release-notes::
|
@ -1,8 +1,7 @@
|
|||||||
[metadata]
|
[metadata]
|
||||||
name = tricircle
|
name = tricircle
|
||||||
summary = Tricircle is an OpenStack project that aims to deal with OpenStack deployment across multiple sites.
|
summary = the Tricircle provides an OpenStack API gateway and networking automation to allow multiple OpenStack instances, spanning in one site or multiple sites or in hybrid cloud, to be managed as a single OpenStack cloud
|
||||||
description-file =
|
description-file = README.rst
|
||||||
README.md
|
|
||||||
author = OpenStack
|
author = OpenStack
|
||||||
author-email = openstack-dev@lists.openstack.org
|
author-email = openstack-dev@lists.openstack.org
|
||||||
home-page = http://www.openstack.org/
|
home-page = http://www.openstack.org/
|
||||||
|
Loading…
Reference in New Issue
Block a user