6b58a1b58e
Change-Id: I1909cf75595fd9da7dfb1481478756711e9bb06c
211 lines
6.7 KiB
Plaintext
211 lines
6.7 KiB
Plaintext
[[local|localrc]]
|
|
|
|
enable_plugin kuryr-kubernetes \
|
|
https://opendev.org/openstack/kuryr-kubernetes
|
|
|
|
enable_plugin dragonflow https://opendev.org/openstack/dragonflow
|
|
|
|
# If you do not want stacking to clone new versions of the enabled services,
|
|
# like for example when you did local modifications and need to ./unstack.sh
|
|
# and ./stack.sh again, uncomment the following
|
|
# RECLONE="no"
|
|
|
|
# Log settings for better readability
|
|
LOGFILE=devstack.log
|
|
LOG_COLOR=False
|
|
|
|
|
|
# Credentials
|
|
ADMIN_PASSWORD=pass
|
|
DATABASE_PASSWORD=pass
|
|
RABBIT_PASSWORD=pass
|
|
SERVICE_PASSWORD=pass
|
|
SERVICE_TOKEN=pass
|
|
# Enable Keystone v3
|
|
IDENTITY_API_VERSION=3
|
|
|
|
# In pro of speed and being lightweight, we will be explicit in regards to
|
|
# which services we enable
|
|
ENABLED_SERVICES=""
|
|
|
|
# DF services
|
|
enable_service df-redis
|
|
enable_service df-redis-server
|
|
enable_service df-controller
|
|
|
|
# Neutron services
|
|
enable_service neutron
|
|
enable_service q-svc
|
|
|
|
# Keystone
|
|
enable_service key
|
|
|
|
# Dependencies
|
|
enable_service mysql
|
|
enable_service rabbit
|
|
|
|
# enable DF local controller
|
|
Q_ENABLE_DRAGONFLOW_LOCAL_CONTROLLER=True
|
|
|
|
# DF settings
|
|
DF_RUNNING_IN_GATE=True
|
|
TUNNEL_TYPE=vxlan
|
|
DF_SELECTIVE_TOPO_DIST=False
|
|
|
|
# OCTAVIA
|
|
# Uncomment it to use L2 communication between loadbalancer and member pods
|
|
# KURYR_K8S_OCTAVIA_MEMBER_MODE=L2
|
|
|
|
# Octavia LBaaSv2
|
|
LIBS_FROM_GIT+=python-octaviaclient
|
|
enable_plugin octavia https://opendev.org/openstack/octavia
|
|
enable_service octavia
|
|
enable_service o-api
|
|
enable_service o-cw
|
|
enable_service o-hm
|
|
enable_service o-hk
|
|
## Octavia Deps
|
|
### Image
|
|
### Barbican
|
|
enable_plugin barbican https://opendev.org/openstack/barbican
|
|
### Nova
|
|
enable_service n-api
|
|
enable_service n-api-meta
|
|
enable_service n-cpu
|
|
enable_service n-cond
|
|
enable_service n-sch
|
|
enable_service placement-api
|
|
enable_service placement-client
|
|
### Glance
|
|
enable_service g-api
|
|
enable_service g-reg
|
|
|
|
# By default use all the services from the kuryr-kubernetes plugin
|
|
|
|
# Docker
|
|
# ======
|
|
# If you already have docker configured, running and with its socket writable
|
|
# by the stack user, you can omit the following line.
|
|
enable_plugin devstack-plugin-container https://opendev.org/openstack/devstack-plugin-container
|
|
|
|
# Etcd
|
|
# ====
|
|
# The default is for devstack to run etcd for you.
|
|
enable_service etcd3
|
|
# If you already have an etcd cluster configured and running, you can just
|
|
# comment out the lines enabling legacy_etcd and etcd3
|
|
# then uncomment and set the following line:
|
|
# KURYR_ETCD_CLIENT_URL="http://etcd_ip:etcd_client_port"
|
|
|
|
# Kubernetes
|
|
# ==========
|
|
#
|
|
# Kubernetes is run from the hyperkube docker image
|
|
# If you already have a Kubernetes deployment, you can use it instead and omit
|
|
# enabling the Kubernetes service (except Kubelet, which must be run by
|
|
# devstack so that it uses our development CNI driver.
|
|
#
|
|
# The default is, again, for devstack to run the Kubernetes services:
|
|
enable_service kubernetes-api
|
|
enable_service kubernetes-controller-manager
|
|
enable_service kubernetes-scheduler
|
|
|
|
# We use hyperkube to run the services. You can select the hyperkube image and/
|
|
# or version by uncommenting and setting the following ENV vars different
|
|
# to the following defaults:
|
|
# KURYR_HYPERKUBE_IMAGE="gcr.io/google_containers/hyperkube-amd64"
|
|
# KURYR_HYPERKUBE_VERSION="v1.6.2"
|
|
#
|
|
# If you have the 8080 port already bound to another service, you will need to
|
|
# have kubernetes API server bind to another port. In order to do that,
|
|
# uncomment and set a different port number in:
|
|
# KURYR_K8S_API_PORT="8080"
|
|
#
|
|
# If you want to test with a different range for the Cluster IPs uncomment and
|
|
# set the following ENV var to a different CIDR
|
|
# KURYR_K8S_CLUSTER_IP_RANGE="10.0.0.0/24"
|
|
#
|
|
# If, however, you are reusing an existing deployment, you should uncomment and
|
|
# set an ENV var so that the Kubelet devstack runs can find the API server:
|
|
# KURYR_K8S_API_URL="http://k8s_api_ip:k8s_api_port"
|
|
#
|
|
|
|
# Kubelet
|
|
# =======
|
|
#
|
|
# Kubelet should almost invariably be run by devstack
|
|
enable_service kubelet
|
|
|
|
# You can specify a different location for the hyperkube binary that will be
|
|
# extracted from the hyperkube container into the Host filesystem:
|
|
# KURYR_HYPERKUBE_BINARY=/usr/local/bin/hyperkube
|
|
#
|
|
# NOTE: KURYR_HYPERKUBE_IMAGE, KURYR_HYPERKUBE_VERSION also affect which
|
|
# the selected binary for the Kubelet.
|
|
|
|
# Kuryr watcher
|
|
# =============
|
|
#
|
|
# Just like the Kubelet, you'll want to have the watcher enabled. It is the
|
|
# part of the codebase that connects to the Kubernetes API server to read the
|
|
# resource events and convert them to Neutron actions
|
|
enable_service kuryr-kubernetes
|
|
|
|
# Kuryr Daemon
|
|
# ============
|
|
#
|
|
# Kuryr runs CNI plugin in daemonized way - i.e. kubelet will run kuryr CNI
|
|
# driver and the driver will pass requests to Kuryr daemon running on the node,
|
|
# instead of processing them on its own. This limits the number of Kubernetes
|
|
# API requests (as only Kuryr Daemon will watch for new pod events) and should
|
|
# increase scalability in environments that often delete and create pods.
|
|
# Since Rocky release this is a default deployment configuration.
|
|
enable_service kuryr-daemon
|
|
|
|
# Kuryr POD VIF Driver
|
|
# ====================
|
|
#
|
|
# Set up the VIF Driver to be used. The default one is the neutron-vif, but if
|
|
# a nested deployment is desired, the corresponding driver need to be set,
|
|
# e.g.: nested-vlan or nested-macvlan
|
|
# KURYR_POD_VIF_DRIVER=neutron-vif
|
|
|
|
# Kuryr Enabled Handlers
|
|
# ======================
|
|
#
|
|
# By default, some Kuryr Handlers are set for DevStack installation. This can be
|
|
# further tweaked in order to enable additional ones such as Network Policy. If
|
|
# you want to add additional handlers those can be set here:
|
|
# KURYR_ENABLED_HANDLERS = vif,lb,lbaasspec
|
|
|
|
# Kuryr Ports Pools
|
|
# =================
|
|
#
|
|
# To speed up containers boot time the kuryr ports pool driver can be enabled
|
|
# by uncommenting the next line, so that neutron port resources are precreated
|
|
# and ready to be used by the pods when needed
|
|
# KURYR_USE_PORTS_POOLS=True
|
|
#
|
|
# By default the pool driver is noop, i.e., there is no pool. If pool
|
|
# optimizations want to be used you need to set it to 'neutron' for the
|
|
# baremetal case, or to 'nested' for the nested case
|
|
# KURYR_VIF_POOL_DRIVER=noop
|
|
#
|
|
# There are extra configuration options for the pools that can be set to decide
|
|
# on the minimum number of ports that should be ready to use at each pool, the
|
|
# maximum (0 to unset), and the batch size for the repopulation actions, i.e.,
|
|
# the number of neutron ports to create in bulk operations. Finally, the update
|
|
# frequency between actions over the pool can be set too
|
|
# KURYR_VIF_POOL_MIN=2
|
|
# KURYR_VIF_POOL_MAX=0
|
|
# KURYR_VIF_POOL_BATCH=5
|
|
# KURYR_VIF_POOL_UPDATE_FREQ=30
|
|
|
|
# Increase Octavia amphorae timeout so that the first LB amphora has time to
|
|
# build and boot
|
|
IMAGE_URLS+=",http://download.cirros-cloud.net/0.3.4/cirros-0.3.4-x86_64-disk.img"
|
|
|
|
[[post-config|$OCTAVIA_CONF]]
|
|
[controller_worker]
|
|
amp_active_retries=9999
|