Merge "Refresh install openstack page to Wallaby"
This commit is contained in:
commit
15c60338d2
|
@ -108,7 +108,7 @@ the environment. It should now look very similar to this:
|
|||
.. code-block:: none
|
||||
|
||||
Model Controller Cloud/Region Version SLA Timestamp
|
||||
openstack maas-controller mymaas/default 2.8.7 unsupported 04:28:49Z
|
||||
openstack maas-controller mymaas/default 2.9.0 unsupported 01:51:00Z
|
||||
|
||||
Model "admin/openstack" is empty
|
||||
|
||||
|
|
|
@ -12,107 +12,107 @@ installed from the instructions given on the :doc:`Install OpenStack
|
|||
|
||||
.. code-block:: console
|
||||
|
||||
Model Controller Cloud/Region Version SLA Timestamp
|
||||
openstack maas-controller mymaas/default 2.8.7 unsupported 01:12:49Z
|
||||
Model Controller Cloud/Region Version SLA Timestamp
|
||||
openstack maas-one maas-one/default 2.9.0 unsupported 01:35:20Z
|
||||
|
||||
App Version Status Scale Charm Store Rev OS Notes
|
||||
ceph-mon 15.2.5 active 3 ceph-mon jujucharms 50 ubuntu
|
||||
ceph-osd 15.2.5 active 4 ceph-osd jujucharms 306 ubuntu
|
||||
ceph-radosgw 15.2.5 active 1 ceph-radosgw jujucharms 291 ubuntu
|
||||
cinder 17.0.0 active 1 cinder jujucharms 306 ubuntu
|
||||
cinder-ceph 17.0.0 active 1 cinder-ceph jujucharms 258 ubuntu
|
||||
cinder-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
dashboard-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
glance 21.0.0 active 1 glance jujucharms 301 ubuntu
|
||||
glance-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
keystone 18.0.0 active 1 keystone jujucharms 319 ubuntu
|
||||
keystone-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
mysql-innodb-cluster 8.0.22 active 3 mysql-innodb-cluster jujucharms 3 ubuntu
|
||||
ncc-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
neutron-api 17.0.0 active 1 neutron-api jujucharms 290 ubuntu
|
||||
neutron-api-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
neutron-api-plugin-ovn 17.0.0 active 1 neutron-api-plugin-ovn jujucharms 2 ubuntu
|
||||
nova-cloud-controller 22.0.0 active 1 nova-cloud-controller jujucharms 349 ubuntu
|
||||
nova-compute 22.0.0 active 3 nova-compute jujucharms 323 ubuntu
|
||||
ntp 3.5 active 4 ntp jujucharms 41 ubuntu
|
||||
openstack-dashboard 18.6.1 active 1 openstack-dashboard jujucharms 309 ubuntu
|
||||
ovn-central 20.03.1 active 3 ovn-central jujucharms 2 ubuntu
|
||||
ovn-chassis 20.03.1 active 3 ovn-chassis jujucharms 7 ubuntu
|
||||
placement 4.0.0 active 1 placement jujucharms 15 ubuntu
|
||||
placement-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
rabbitmq-server 3.8.2 active 1 rabbitmq-server jujucharms 106 ubuntu
|
||||
vault 1.5.4 active 1 vault jujucharms 41 ubuntu
|
||||
vault-mysql-router 8.0.22 active 1 mysql-router jujucharms 4 ubuntu
|
||||
App Version Status Scale Charm Store Channel Rev OS Message
|
||||
ceph-mon 16.2.0 active 3 ceph-mon charmstore stable 464 ubuntu Unit is ready and clustered
|
||||
ceph-osd 16.2.0 active 4 ceph-osd charmstore stable 489 ubuntu Unit is ready (2 OSD)
|
||||
ceph-radosgw 16.2.0 active 1 ceph-radosgw charmstore stable 398 ubuntu Unit is ready
|
||||
cinder 18.0.0 active 1 cinder charmstore stable 436 ubuntu Unit is ready
|
||||
cinder-ceph 18.0.0 active 1 cinder-ceph charmstore stable 352 ubuntu Unit is ready
|
||||
cinder-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
dashboard-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
glance 22.0.0 active 1 glance charmstore stable 450 ubuntu Unit is ready
|
||||
glance-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
keystone 19.0.0 active 1 keystone charmstore stable 542 ubuntu Application Ready
|
||||
keystone-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
mysql-innodb-cluster 8.0.23 active 3 mysql-innodb-cluster charmstore stable 74 ubuntu Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
ncc-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
neutron-api 18.0.0 active 1 neutron-api charmstore stable 471 ubuntu Unit is ready
|
||||
neutron-api-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
neutron-api-plugin-ovn 18.0.0 active 1 neutron-api-plugin-ovn charmstore stable 40 ubuntu Unit is ready
|
||||
nova-cloud-controller 23.0.0 active 1 nova-cloud-controller charmstore stable 521 ubuntu Unit is ready
|
||||
nova-compute 23.0.0 active 3 nova-compute charmstore stable 539 ubuntu Unit is ready
|
||||
ntp 3.5 active 4 ntp charmstore stable 45 ubuntu chrony: Ready
|
||||
openstack-dashboard 19.2.0 active 1 openstack-dashboard charmstore stable 505 ubuntu Unit is ready
|
||||
ovn-central 20.12.0 active 3 ovn-central charmstore stable 51 ubuntu Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
ovn-chassis 20.12.0 active 3 ovn-chassis charmstore stable 63 ubuntu Unit is ready
|
||||
placement 5.0.0 active 1 placement charmstore stable 47 ubuntu Unit is ready
|
||||
placement-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
rabbitmq-server 3.8.2 active 1 rabbitmq-server charmstore stable 406 ubuntu Unit is ready
|
||||
vault 1.5.4 active 1 vault charmstore stable 141 ubuntu Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router 8.0.23 active 1 mysql-router charmstore stable 48 ubuntu Unit is ready
|
||||
|
||||
Unit Workload Agent Machine Public address Ports Message
|
||||
ceph-mon/0* active idle 0/lxd/3 10.0.0.191 Unit is ready and clustered
|
||||
ceph-mon/1 active idle 1/lxd/3 10.0.0.189 Unit is ready and clustered
|
||||
ceph-mon/2 active idle 2/lxd/4 10.0.0.190 Unit is ready and clustered
|
||||
ceph-osd/0* active idle 0 10.0.0.171 Unit is ready (1 OSD)
|
||||
ntp/1 active idle 10.0.0.171 123/udp chrony: Ready
|
||||
ceph-osd/1 active idle 1 10.0.0.172 Unit is ready (1 OSD)
|
||||
ntp/0* active idle 10.0.0.172 123/udp chrony: Ready
|
||||
ceph-osd/2 active idle 2 10.0.0.173 Unit is ready (1 OSD)
|
||||
ntp/3 active idle 10.0.0.173 123/udp chrony: Ready
|
||||
ceph-osd/3 active idle 3 10.0.0.174 Unit is ready (1 OSD)
|
||||
ntp/2 active idle 10.0.0.174 123/udp chrony: Ready
|
||||
ceph-radosgw/0* active idle 0/lxd/4 10.0.0.193 80/tcp Unit is ready
|
||||
cinder/0* active idle 1/lxd/4 10.0.0.192 8776/tcp Unit is ready
|
||||
cinder-ceph/0* active idle 10.0.0.192 Unit is ready
|
||||
cinder-mysql-router/0* active idle 10.0.0.192 Unit is ready
|
||||
glance/0* active idle 3/lxd/3 10.0.0.188 9292/tcp Unit is ready
|
||||
glance-mysql-router/0* active idle 10.0.0.188 Unit is ready
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.29 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.29 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.175 Unit is ready: Mode: R/W
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.176 Unit is ready: Mode: R/O
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.177 Unit is ready: Mode: R/O
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.182 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.182 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.182 Unit is ready
|
||||
nova-cloud-controller/0* active idle 3/lxd/1 10.0.0.185 8774/tcp,8775/tcp Unit is ready
|
||||
ncc-mysql-router/0* active idle 10.0.0.185 Unit is ready
|
||||
nova-compute/0* active idle 1 10.0.0.172 Unit is ready
|
||||
ovn-chassis/0* active idle 10.0.0.172 Unit is ready
|
||||
nova-compute/1 active idle 2 10.0.0.173 Unit is ready
|
||||
ovn-chassis/2 active idle 10.0.0.173 Unit is ready
|
||||
nova-compute/2 active idle 3 10.0.0.174 Unit is ready
|
||||
ovn-chassis/1 active idle 10.0.0.174 Unit is ready
|
||||
openstack-dashboard/0* active idle 2/lxd/3 10.0.0.187 80/tcp,443/tcp Unit is ready
|
||||
dashboard-mysql-router/0* active idle 10.0.0.187 Unit is ready
|
||||
ovn-central/0 active idle 0/lxd/1 10.0.0.181 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db)
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.179 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2* active idle 2/lxd/1 10.0.0.180 6641/tcp,6642/tcp Unit is ready (leader: ovnsb_db northd: active)
|
||||
placement/0* active idle 3/lxd/2 10.0.0.186 8778/tcp Unit is ready
|
||||
placement-mysql-router/0* active idle 10.0.0.186 Unit is ready
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.184 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.178 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.178 Unit is ready
|
||||
ceph-mon/0 active idle 0/lxd/3 10.0.0.170 Unit is ready and clustered
|
||||
ceph-mon/1 active idle 1/lxd/3 10.0.0.169 Unit is ready and clustered
|
||||
ceph-mon/2* active idle 2/lxd/4 10.0.0.168 Unit is ready and clustered
|
||||
ceph-osd/0* active idle 0 10.0.0.150 Unit is ready (2 OSD)
|
||||
ntp/3 active idle 10.0.0.150 123/udp chrony: Ready
|
||||
ceph-osd/1 active idle 1 10.0.0.151 Unit is ready (2 OSD)
|
||||
ntp/2 active idle 10.0.0.151 123/udp chrony: Ready
|
||||
ceph-osd/2 active idle 2 10.0.0.152 Unit is ready (2 OSD)
|
||||
ntp/1 active idle 10.0.0.152 123/udp chrony: Ready
|
||||
ceph-osd/3 active idle 3 10.0.0.153 Unit is ready (2 OSD)
|
||||
ntp/0* active idle 10.0.0.153 123/udp chrony: Ready
|
||||
ceph-radosgw/0* active idle 0/lxd/4 10.0.0.172 80/tcp Unit is ready
|
||||
cinder/0* active idle 1/lxd/4 10.0.0.171 8776/tcp Unit is ready
|
||||
cinder-ceph/0* active idle 10.0.0.171 Unit is ready
|
||||
cinder-mysql-router/0* active idle 10.0.0.171 Unit is ready
|
||||
glance/0* active idle 3/lxd/3 10.0.0.167 9292/tcp Unit is ready
|
||||
glance-mysql-router/0* active idle 10.0.0.167 Unit is ready
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.162 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.162 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.154 Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.155 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.156 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.161 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.161 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.161 Unit is ready
|
||||
nova-cloud-controller/0* active idle 3/lxd/1 10.0.0.164 8774/tcp,8775/tcp Unit is ready
|
||||
ncc-mysql-router/0* active idle 10.0.0.164 Unit is ready
|
||||
nova-compute/0* active idle 1 10.0.0.151 Unit is ready
|
||||
ovn-chassis/2 active idle 10.0.0.151 Unit is ready
|
||||
nova-compute/1 active idle 2 10.0.0.152 Unit is ready
|
||||
ovn-chassis/0* active idle 10.0.0.152 Unit is ready
|
||||
nova-compute/2 active idle 3 10.0.0.153 Unit is ready
|
||||
ovn-chassis/1 active idle 10.0.0.153 Unit is ready
|
||||
openstack-dashboard/0* active idle 2/lxd/3 10.0.0.166 80/tcp,443/tcp Unit is ready
|
||||
dashboard-mysql-router/0* active idle 10.0.0.166 Unit is ready
|
||||
ovn-central/0* active idle 0/lxd/1 10.0.0.158 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.159 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2 active idle 2/lxd/1 10.0.0.160 6641/tcp,6642/tcp Unit is ready
|
||||
placement/0* active idle 3/lxd/2 10.0.0.165 8778/tcp Unit is ready
|
||||
placement-mysql-router/0* active idle 10.0.0.165 Unit is ready
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.163 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.157 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.157 Unit is ready
|
||||
|
||||
Machine State DNS Inst id Series AZ Message
|
||||
0 started 10.0.0.171 node2 focal default Deployed
|
||||
0/lxd/0 started 10.0.0.175 juju-bdbf2c-0-lxd-0 focal default Container started
|
||||
0/lxd/1 started 10.0.0.181 juju-bdbf2c-0-lxd-1 focal default Container started
|
||||
0/lxd/2 started 10.0.0.29 juju-bdbf2c-0-lxd-2 focal default Container started
|
||||
0/lxd/3 started 10.0.0.191 juju-bdbf2c-0-lxd-3 focal default Container started
|
||||
0/lxd/4 started 10.0.0.193 juju-bdbf2c-0-lxd-4 focal default Container started
|
||||
1 started 10.0.0.172 node1 focal default Deployed
|
||||
1/lxd/0 started 10.0.0.176 juju-bdbf2c-1-lxd-0 focal default Container started
|
||||
1/lxd/1 started 10.0.0.179 juju-bdbf2c-1-lxd-1 focal default Container started
|
||||
1/lxd/2 started 10.0.0.182 juju-bdbf2c-1-lxd-2 focal default Container started
|
||||
1/lxd/3 started 10.0.0.189 juju-bdbf2c-1-lxd-3 focal default Container started
|
||||
1/lxd/4 started 10.0.0.192 juju-bdbf2c-1-lxd-4 focal default Container started
|
||||
2 started 10.0.0.173 node3 focal default Deployed
|
||||
2/lxd/0 started 10.0.0.177 juju-bdbf2c-2-lxd-0 focal default Container started
|
||||
2/lxd/1 started 10.0.0.180 juju-bdbf2c-2-lxd-1 focal default Container started
|
||||
2/lxd/2 started 10.0.0.184 juju-bdbf2c-2-lxd-2 focal default Container started
|
||||
2/lxd/3 started 10.0.0.187 juju-bdbf2c-2-lxd-3 focal default Container started
|
||||
2/lxd/4 started 10.0.0.190 juju-bdbf2c-2-lxd-4 focal default Container started
|
||||
3 started 10.0.0.174 node4 focal default Deployed
|
||||
3/lxd/0 started 10.0.0.178 juju-bdbf2c-3-lxd-0 focal default Container started
|
||||
3/lxd/1 started 10.0.0.185 juju-bdbf2c-3-lxd-1 focal default Container started
|
||||
3/lxd/2 started 10.0.0.186 juju-bdbf2c-3-lxd-2 focal default Container started
|
||||
3/lxd/3 started 10.0.0.188 juju-bdbf2c-3-lxd-3 focal default Container started
|
||||
0 started 10.0.0.150 node4 focal default Deployed
|
||||
0/lxd/0 started 10.0.0.154 juju-3d942c-0-lxd-0 focal default Container started
|
||||
0/lxd/1 started 10.0.0.158 juju-3d942c-0-lxd-1 focal default Container started
|
||||
0/lxd/2 started 10.0.0.162 juju-3d942c-0-lxd-2 focal default Container started
|
||||
0/lxd/3 started 10.0.0.170 juju-3d942c-0-lxd-3 focal default Container started
|
||||
0/lxd/4 started 10.0.0.172 juju-3d942c-0-lxd-4 focal default Container started
|
||||
1 started 10.0.0.151 node1 focal default Deployed
|
||||
1/lxd/0 started 10.0.0.155 juju-3d942c-1-lxd-0 focal default Container started
|
||||
1/lxd/1 started 10.0.0.159 juju-3d942c-1-lxd-1 focal default Container started
|
||||
1/lxd/2 started 10.0.0.161 juju-3d942c-1-lxd-2 focal default Container started
|
||||
1/lxd/3 started 10.0.0.169 juju-3d942c-1-lxd-3 focal default Container started
|
||||
1/lxd/4 started 10.0.0.171 juju-3d942c-1-lxd-4 focal default Container started
|
||||
2 started 10.0.0.152 node2 focal default Deployed
|
||||
2/lxd/0 started 10.0.0.156 juju-3d942c-2-lxd-0 focal default Container started
|
||||
2/lxd/1 started 10.0.0.160 juju-3d942c-2-lxd-1 focal default Container started
|
||||
2/lxd/2 started 10.0.0.163 juju-3d942c-2-lxd-2 focal default Container started
|
||||
2/lxd/3 started 10.0.0.166 juju-3d942c-2-lxd-3 focal default Container started
|
||||
2/lxd/4 started 10.0.0.168 juju-3d942c-2-lxd-4 focal default Container started
|
||||
3 started 10.0.0.153 node3 focal default Deployed
|
||||
3/lxd/0 started 10.0.0.157 juju-3d942c-3-lxd-0 focal default Container started
|
||||
3/lxd/1 started 10.0.0.164 juju-3d942c-3-lxd-1 focal default Container started
|
||||
3/lxd/2 started 10.0.0.165 juju-3d942c-3-lxd-2 focal default Container started
|
||||
3/lxd/3 started 10.0.0.167 juju-3d942c-3-lxd-3 focal default Container started
|
||||
|
||||
Relation provider Requirer Interface Type Message
|
||||
ceph-mon:client cinder-ceph:ceph ceph-client regular
|
||||
|
@ -178,6 +178,7 @@ installed from the instructions given on the :doc:`Install OpenStack
|
|||
vault:certificates cinder:certificates tls-certificates regular
|
||||
vault:certificates glance:certificates tls-certificates regular
|
||||
vault:certificates keystone:certificates tls-certificates regular
|
||||
vault:certificates mysql-innodb-cluster:certificates tls-certificates regular
|
||||
vault:certificates neutron-api-plugin-ovn:certificates tls-certificates regular
|
||||
vault:certificates neutron-api:certificates tls-certificates regular
|
||||
vault:certificates nova-cloud-controller:certificates tls-certificates regular
|
||||
|
|
|
@ -60,8 +60,8 @@ OpenStack release
|
|||
do use this method).
|
||||
|
||||
As the :doc:`Overview <install-overview>` of the Installation section states,
|
||||
OpenStack Victoria will be deployed atop Ubuntu 20.04 LTS (Focal) cloud nodes.
|
||||
In order to achieve this a cloud archive release of 'cloud:focal-victoria' will
|
||||
OpenStack Wallaby will be deployed atop Ubuntu 20.04 LTS (Focal) cloud nodes.
|
||||
In order to achieve this a cloud archive release of 'cloud:focal-wallaby' will
|
||||
be used during the install of each OpenStack application. Note that some
|
||||
applications are not part of the OpenStack project per se and therefore do not
|
||||
apply (exceptionally, Ceph applications do use this method). Not using a more
|
||||
|
@ -75,7 +75,7 @@ releases and how they are used when upgrading OpenStack.
|
|||
.. important::
|
||||
|
||||
The chosen OpenStack release may impact the installation and configuration
|
||||
instructions. **This guide assumes that OpenStack Victoria is being
|
||||
instructions. **This guide assumes that OpenStack Wallaby is being
|
||||
deployed.**
|
||||
|
||||
Installation progress
|
||||
|
@ -131,7 +131,7 @@ contains the configuration.
|
|||
|
||||
ceph-osd:
|
||||
osd-devices: /dev/sdb
|
||||
source: cloud:focal-victoria
|
||||
source: cloud:focal-wallaby
|
||||
|
||||
To deploy the application we'll make use of the 'compute' tag that we placed on
|
||||
each of these nodes on the :doc:`Install MAAS <install-maas>` page:
|
||||
|
@ -162,10 +162,11 @@ charm. We'll then scale-out the application to two other machines. File
|
|||
.. code-block:: yaml
|
||||
|
||||
nova-compute:
|
||||
config-flags: default_ephemeral_format=ext4
|
||||
enable-live-migration: true
|
||||
enable-resize: true
|
||||
migration-auth-type: ssh
|
||||
openstack-origin: cloud:focal-victoria
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
The initial node must be targeted by machine since there are no more free Juju
|
||||
machines (MAAS nodes) available. This means we're placing multiple services on
|
||||
|
@ -221,13 +222,13 @@ Here are the corresponding commands for Vault:
|
|||
juju add-relation vault-mysql-router:db-router mysql-innodb-cluster:db-router
|
||||
juju add-relation vault-mysql-router:shared-db vault:shared-db
|
||||
|
||||
Vault now needs to be initialised and unsealed. The vault charm will also need
|
||||
to be authorised to carry out certain tasks. These steps are covered in the
|
||||
`vault charm`_ documentation. Perform them now.
|
||||
Vault must now be initialised and unsealed. The vault charm will also need to
|
||||
be authorised to carry out certain tasks. These steps are covered in the `vault
|
||||
charm`_ documentation. Perform them now.
|
||||
|
||||
Vault must now be provided with a CA certificate in order for it to issue
|
||||
certificates to cloud API services. This is covered on the :ref:`Managing TLS
|
||||
certificates <add_ca_certificate>` page. Do this now.
|
||||
Provide Vault with a CA certificate so it can issue certificates to cloud API
|
||||
services. This is covered on the :ref:`Managing TLS certificates
|
||||
<add_ca_certificate>` page. Do this now.
|
||||
|
||||
Once the above is completed the Unit section output to command :command:`juju
|
||||
status` should look similar to this:
|
||||
|
@ -235,18 +236,18 @@ status` should look similar to this:
|
|||
.. code-block:: console
|
||||
|
||||
Unit Workload Agent Machine Public address Ports Message
|
||||
ceph-osd/0* blocked idle 0 10.0.0.171 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.172 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.173 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.174 Missing relation: monitor
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.175 Unit is ready: Mode: R/W
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.176 Unit is ready: Mode: R/O
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.177 Unit is ready: Mode: R/O
|
||||
nova-compute/0* blocked idle 1 10.0.0.172 Missing relations: messaging, image
|
||||
nova-compute/1 blocked idle 2 10.0.0.173 Missing relations: messaging, image
|
||||
nova-compute/2 blocked idle 3 10.0.0.174 Missing relations: messaging, image
|
||||
vault/0* active idle 3/lxd/0 10.0.0.178 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.178 Unit is ready
|
||||
ceph-osd/0* blocked idle 0 10.0.0.150 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.151 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.152 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.153 Missing relation: monitor
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.154 Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.155 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.156 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
nova-compute/0* blocked idle 1 10.0.0.151 Missing relations: messaging, image
|
||||
nova-compute/1 blocked idle 2 10.0.0.152 Missing relations: messaging, image
|
||||
nova-compute/2 blocked idle 3 10.0.0.153 Missing relations: messaging, image
|
||||
vault/0* active idle 3/lxd/0 10.0.0.157 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.157 Unit is ready
|
||||
|
||||
Cloud applications are TLS-enabled via the ``vault:certificates`` relation.
|
||||
Below we start with the cloud database. Although the latter has a self-signed
|
||||
|
@ -278,9 +279,10 @@ File ``neutron.yaml`` contains the configuration necessary for three of them:
|
|||
neutron-api:
|
||||
neutron-security-groups: true
|
||||
flat-network-providers: physnet1
|
||||
openstack-origin: cloud:focal-victoria
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
ovn-central:
|
||||
source: cloud:focal-victoria
|
||||
source: cloud:focal-wallaby
|
||||
|
||||
The ``bridge-interface-mappings`` setting refers to a network interface that
|
||||
the OVN Chassis will bind to. In the above example it is 'eth1' and it should
|
||||
|
@ -341,11 +343,20 @@ Join neutron-api to the cloud database:
|
|||
Keystone
|
||||
~~~~~~~~
|
||||
|
||||
The keystone application will be containerised on machine 0:
|
||||
The keystone application will be containerised on machine 0. File
|
||||
``keystone.yaml`` contains the configuration:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
keystone:
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
To deploy:
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy --to lxd:0 --config openstack-origin=cloud:focal-victoria keystone
|
||||
juju deploy --to lxd:0 --config keystone.yaml keystone
|
||||
|
||||
Join keystone to the cloud database:
|
||||
|
||||
|
@ -385,30 +396,30 @@ look similar to this:
|
|||
.. code-block:: console
|
||||
|
||||
Unit Workload Agent Machine Public address Ports Message
|
||||
ceph-osd/0* blocked idle 0 10.0.0.171 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.172 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.173 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.174 Missing relation: monitor
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.183 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.183 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.175 Unit is ready: Mode: R/W
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.176 Unit is ready: Mode: R/O
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.177 Unit is ready: Mode: R/O
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.182 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.182 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.182 Unit is ready
|
||||
nova-compute/0* blocked idle 1 10.0.0.172 Missing relations: image
|
||||
ovn-chassis/0* active idle 10.0.0.172 Unit is ready
|
||||
nova-compute/1 blocked idle 2 10.0.0.173 Missing relations: image
|
||||
ovn-chassis/2 active idle 10.0.0.173 Unit is ready
|
||||
nova-compute/2 blocked idle 3 10.0.0.174 Missing relations: image
|
||||
ovn-chassis/1 active idle 10.0.0.174 Unit is ready
|
||||
ovn-central/0 active idle 0/lxd/1 10.0.0.181 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.179 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2* active idle 2/lxd/1 10.0.0.180 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.184 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.178 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.178 Unit is ready
|
||||
ceph-osd/0* blocked idle 0 10.0.0.150 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.151 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.152 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.153 Missing relation: monitor
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.162 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.162 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.154 Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.155 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.156 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.161 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.161 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.161 Unit is ready
|
||||
nova-compute/0* blocked idle 1 10.0.0.151 Missing relations: image
|
||||
ovn-chassis/2 active idle 10.0.0.151 Unit is ready
|
||||
nova-compute/1 blocked idle 2 10.0.0.152 Missing relations: image
|
||||
ovn-chassis/0* active idle 10.0.0.152 Unit is ready
|
||||
nova-compute/2 blocked idle 3 10.0.0.153 Missing relations: image
|
||||
ovn-chassis/1 active idle 10.0.0.153 Unit is ready
|
||||
ovn-central/0* active idle 0/lxd/1 10.0.0.158 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.159 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2 active idle 2/lxd/1 10.0.0.160 6641/tcp,6642/tcp Unit is ready
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.163 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.157 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.157 Unit is ready
|
||||
|
||||
Nova cloud controller
|
||||
~~~~~~~~~~~~~~~~~~~~~
|
||||
|
@ -422,7 +433,8 @@ the configuration:
|
|||
|
||||
nova-cloud-controller:
|
||||
network-manager: Neutron
|
||||
openstack-origin: cloud:focal-victoria
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
To deploy:
|
||||
|
||||
|
@ -458,11 +470,19 @@ Placement
|
|||
~~~~~~~~~
|
||||
|
||||
The placement application will be containerised on machine 3 with the
|
||||
`placement`_ charm:
|
||||
`placement`_ charm. File ``placement.yaml`` contains the configuration:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
placement:
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
To deploy:
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy --to lxd:3 --config openstack-origin=cloud:focal-victoria placement
|
||||
juju deploy --to lxd:3 --config placement.yaml placement
|
||||
|
||||
Join placement to the cloud database:
|
||||
|
||||
|
@ -488,7 +508,7 @@ The openstack-dashboard application (Horizon) will be containerised on machine
|
|||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy --to lxd:2 --config openstack-origin=cloud:focal-victoria openstack-dashboard
|
||||
juju deploy --to lxd:2 --config openstack-origin=cloud:focal-wallaby openstack-dashboard
|
||||
|
||||
Join openstack-dashboard to the cloud database:
|
||||
|
||||
|
@ -515,11 +535,19 @@ Glance
|
|||
~~~~~~
|
||||
|
||||
The glance application will be containerised on machine 3 with the `glance`_
|
||||
charm:
|
||||
charm. File ``glance.yaml`` contains the configuration:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
glance:
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
To deploy:
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy --to lxd:3 --config openstack-origin=cloud:focal-victoria glance
|
||||
juju deploy --to lxd:3 --config glance.yaml glance
|
||||
|
||||
Join glance to the cloud database:
|
||||
|
||||
|
@ -544,48 +572,55 @@ look similar to this:
|
|||
.. code-block:: console
|
||||
|
||||
Unit Workload Agent Machine Public address Ports Message
|
||||
ceph-osd/0* blocked idle 0 10.0.0.171 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.172 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.173 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.174 Missing relation: monitor
|
||||
glance/0* active idle 3/lxd/3 10.0.0.188 9292/tcp Unit is ready
|
||||
glance-mysql-router/0* active idle 10.0.0.188 Unit is ready
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.183 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.183 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.175 Unit is ready: Mode: R/W
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.176 Unit is ready: Mode: R/O
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.177 Unit is ready: Mode: R/O
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.182 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.182 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.182 Unit is ready
|
||||
nova-cloud-controller/0* active idle 3/lxd/1 10.0.0.185 8774/tcp,8775/tcp Unit is ready
|
||||
ncc-mysql-router/0* active idle 10.0.0.185 Unit is ready
|
||||
nova-compute/0* active idle 1 10.0.0.172 Unit is ready
|
||||
ovn-chassis/0* active idle 10.0.0.172 Unit is ready
|
||||
nova-compute/1 active idle 2 10.0.0.173 Unit is ready
|
||||
ovn-chassis/2 active idle 10.0.0.173 Unit is ready
|
||||
nova-compute/2 active idle 3 10.0.0.174 Unit is ready
|
||||
ovn-chassis/1 active idle 10.0.0.174 Unit is ready
|
||||
openstack-dashboard/0* active idle 2/lxd/3 10.0.0.187 80/tcp,443/tcp Unit is ready
|
||||
dashboard-mysql-router/0* active idle 10.0.0.187 Unit is ready
|
||||
ovn-central/0 active idle 0/lxd/1 10.0.0.181 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.179 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2* active idle 2/lxd/1 10.0.0.180 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
placement/0* active idle 3/lxd/2 10.0.0.186 8778/tcp Unit is ready
|
||||
placement-mysql-router/0* active idle 10.0.0.186 Unit is ready
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.184 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.178 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.178 Unit is ready
|
||||
ceph-osd/0* blocked idle 0 10.0.0.150 Missing relation: monitor
|
||||
ceph-osd/1 blocked idle 1 10.0.0.151 Missing relation: monitor
|
||||
ceph-osd/2 blocked idle 2 10.0.0.152 Missing relation: monitor
|
||||
ceph-osd/3 blocked idle 3 10.0.0.153 Missing relation: monitor
|
||||
glance/0* active idle 3/lxd/3 10.0.0.167 9292/tcp Unit is ready
|
||||
glance-mysql-router/0* active idle 10.0.0.167 Unit is ready
|
||||
keystone/0* active idle 0/lxd/2 10.0.0.162 5000/tcp Unit is ready
|
||||
keystone-mysql-router/0* active idle 10.0.0.162 Unit is ready
|
||||
mysql-innodb-cluster/0* active idle 0/lxd/0 10.0.0.154 Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/1 active idle 1/lxd/0 10.0.0.155 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
mysql-innodb-cluster/2 active idle 2/lxd/0 10.0.0.156 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
|
||||
neutron-api/0* active idle 1/lxd/2 10.0.0.161 9696/tcp Unit is ready
|
||||
neutron-api-mysql-router/0* active idle 10.0.0.161 Unit is ready
|
||||
neutron-api-plugin-ovn/0* active idle 10.0.0.161 Unit is ready
|
||||
nova-cloud-controller/0* active idle 3/lxd/1 10.0.0.164 8774/tcp,8775/tcp Unit is ready
|
||||
ncc-mysql-router/0* active idle 10.0.0.164 Unit is ready
|
||||
nova-compute/0* active idle 1 10.0.0.151 Unit is ready
|
||||
ovn-chassis/2 active idle 10.0.0.151 Unit is ready
|
||||
nova-compute/1 active idle 2 10.0.0.152 Unit is ready
|
||||
ovn-chassis/0* active idle 10.0.0.152 Unit is ready
|
||||
nova-compute/2 active idle 3 10.0.0.153 Unit is ready
|
||||
ovn-chassis/1 active idle 10.0.0.153 Unit is ready
|
||||
openstack-dashboard/0* active idle 2/lxd/3 10.0.0.166 80/tcp,443/tcp Unit is ready
|
||||
dashboard-mysql-router/0* active idle 10.0.0.166 Unit is ready
|
||||
ovn-central/0* active idle 0/lxd/1 10.0.0.158 6641/tcp,6642/tcp Unit is ready (leader: ovnnb_db, ovnsb_db northd: active)
|
||||
ovn-central/1 active idle 1/lxd/1 10.0.0.159 6641/tcp,6642/tcp Unit is ready
|
||||
ovn-central/2 active idle 2/lxd/1 10.0.0.160 6641/tcp,6642/tcp Unit is ready
|
||||
placement/0* active idle 3/lxd/2 10.0.0.165 8778/tcp Unit is ready
|
||||
placement-mysql-router/0* active idle 10.0.0.165 Unit is ready
|
||||
rabbitmq-server/0* active idle 2/lxd/2 10.0.0.163 5672/tcp Unit is ready
|
||||
vault/0* active idle 3/lxd/0 10.0.0.157 8200/tcp Unit is ready (active: true, mlock: disabled)
|
||||
vault-mysql-router/0* active idle 10.0.0.157 Unit is ready
|
||||
|
||||
Ceph monitor
|
||||
~~~~~~~~~~~~
|
||||
|
||||
The ceph-mon application will be containerised on machines 0, 1, and 2 with the
|
||||
`ceph-mon`_ charm:
|
||||
`ceph-mon`_ charm. File ``ceph-mon.yaml`` contains the configuration:
|
||||
|
||||
.. code-block:: yaml
|
||||
|
||||
ceph-mon:
|
||||
expected-osd-count: 3
|
||||
monitor-count: 3
|
||||
source: cloud:focal-wallaby
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy -n 3 --to lxd:0,lxd:1,lxd:2 --config source=cloud:focal-victoria ceph-mon
|
||||
juju deploy -n 3 --to lxd:0,lxd:1,lxd:2 --config ceph-mon.yaml ceph-mon
|
||||
|
||||
Three relations can be added at this time:
|
||||
|
||||
|
@ -612,9 +647,10 @@ charm. File ``cinder.yaml`` contains the configuration:
|
|||
.. code-block:: yaml
|
||||
|
||||
cinder:
|
||||
glance-api-version: 2
|
||||
block-device: None
|
||||
openstack-origin: cloud:focal-victoria
|
||||
glance-api-version: 2
|
||||
worker-multiplier: 0.25
|
||||
openstack-origin: cloud:focal-wallaby
|
||||
|
||||
To deploy:
|
||||
|
||||
|
@ -671,7 +707,7 @@ The ceph-radosgw application will be containerised on machine 0 with the
|
|||
|
||||
.. code-block:: none
|
||||
|
||||
juju deploy --to lxd:0 ceph-radosgw
|
||||
juju deploy --to lxd:0 --config source=cloud:focal-wallaby ceph-radosgw
|
||||
|
||||
A single relation is needed:
|
||||
|
||||
|
@ -716,40 +752,36 @@ Obtain the address in this way:
|
|||
|
||||
juju status --format=yaml openstack-dashboard | grep public-address | awk '{print $2}' | head -1
|
||||
|
||||
The password is queried from Keystone:
|
||||
In this example, the address is '10.0.0.166'.
|
||||
|
||||
The password can be queried from Keystone:
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
juju run --unit keystone/0 leader-get admin_passwd
|
||||
|
||||
In this example, the address is '10.0.0.187' and the password is
|
||||
'kohy6shoh3diWav5'.
|
||||
juju run --unit keystone/leader leader-get admin_passwd
|
||||
|
||||
The dashboard URL then becomes:
|
||||
|
||||
**http://10.0.0.187/horizon**
|
||||
**http://10.0.0.166/horizon**
|
||||
|
||||
And the credentials are:
|
||||
The final credentials needed to log in are:
|
||||
|
||||
| Domain: **admin_domain**
|
||||
| User Name: **admin**
|
||||
| Password: **kohy6shoh3diWav5**
|
||||
| Password: ********************
|
||||
| Domain: **admin_domain**
|
||||
|
|
||||
|
||||
.. tip::
|
||||
|
||||
To access the dasboard from your desktop you will need SSH local port
|
||||
forwarding. Example: ``sudo ssh -L 8001:10.0.0.187:80 <user>@<host>``, where
|
||||
<host> can contact 10.0.0.187 on port 80. Then go to
|
||||
http://localhost:8001/horizon.
|
||||
|
||||
Once logged in you should see something like this:
|
||||
|
||||
.. figure:: ./media/install-openstack_horizon.png
|
||||
:scale: 70%
|
||||
:alt: Horizon dashboard
|
||||
|
||||
To enable instance console access from within Horizon:
|
||||
VM consoles
|
||||
~~~~~~~~~~~
|
||||
|
||||
Enable a remote access protocol such as novnc (or spice) if you want to connect
|
||||
to VM consoles from within the dashboard:
|
||||
|
||||
.. code-block:: none
|
||||
|
||||
|
@ -758,8 +790,8 @@ To enable instance console access from within Horizon:
|
|||
Next steps
|
||||
----------
|
||||
|
||||
You have successfully deployed OpenStack using both Juju and MAAS. The next
|
||||
step is to render the cloud functional for users. This will involve setting up
|
||||
You have successfully deployed OpenStack using Juju and MAAS. The next step is
|
||||
to render the cloud functional for users. This will involve setting up
|
||||
networks, images, and a user environment. Go to :doc:`Configure OpenStack
|
||||
<configure-openstack>` now.
|
||||
|
||||
|
|
|
@ -13,9 +13,9 @@ The software versions used in this guide are as follows:
|
|||
|
||||
* **Ubuntu 20.04 LTS (Focal)** for the MAAS server, Juju client, Juju
|
||||
controller, and all cloud nodes (including containers)
|
||||
* **MAAS 2.9.1**
|
||||
* **Juju 2.8.7**
|
||||
* **OpenStack Victoria**
|
||||
* **MAAS 2.9.2**
|
||||
* **Juju 2.9.0**
|
||||
* **OpenStack Wallaby**
|
||||
|
||||
Proceed to the :doc:`Install MAAS <install-maas>` page to begin your
|
||||
installation journey. Hardware requirements are also listed there.
|
||||
|
|
Loading…
Reference in New Issue