Remove get-occ-config.sh script for deployed server
Change-Id: I1bd8e1a3fe0a44240bab941540703c39aa8446bd Partial-Bug: #1832441
This commit is contained in:
parent
89e9fe53b2
commit
12c0f3a160
|
@ -72,63 +72,3 @@ parameter_defaults:
|
|||
In this example, 192.168.122.130 is the external management IP of an
|
||||
undercloud, thus it is the default route for the configured local_ip value of
|
||||
192.168.24.1.
|
||||
|
||||
|
||||
os-collect-config
|
||||
-----------------
|
||||
os-collect-config on each deployed server must be manually configured to poll
|
||||
the Heat API for the available SoftwareDeployments. An example configuration
|
||||
for /etc/os-collect-config.conf looks like:
|
||||
|
||||
[DEFAULT]
|
||||
collectors=heat
|
||||
command=os-refresh-config
|
||||
|
||||
[heat]
|
||||
# you can get these values from stackrc on the undercloud
|
||||
user_id=<a user that can connect to heat> # note this must be the ID, not the username
|
||||
password=<a password>
|
||||
auth_url=<keystone url>
|
||||
project_id=<project_id> # note, this must be the ID, not project name
|
||||
stack_id=<stack_id>
|
||||
resource_name=<resource_name>
|
||||
|
||||
Note that the stack_id value is the id of the nested stack containing the
|
||||
resource (identified by resource_name) implemented by the deployed-server.yaml
|
||||
templates.
|
||||
|
||||
Once the configuration for os-collect-config has been defined, the service
|
||||
needs to be restarted. Once restarted, it will start polling Heat and applying
|
||||
the SoftwareDeployments.
|
||||
|
||||
A sample script at deployed-server/scripts/get-occ-config.sh is included that
|
||||
will automatically generate the os-collect-config configuration needed on each
|
||||
server, ssh to each server, copy the configuration, and restart the
|
||||
os-collect-config service.
|
||||
|
||||
.. warning::
|
||||
The get-occ-config.sh script is not intended for production use, as it
|
||||
copies admin credentials to each of the deployed nodes.
|
||||
|
||||
The script can only be used once the stack id's of the nested deployed-server
|
||||
stacks have been created via Heat. This usually only takes a couple of minutes
|
||||
once the deployment command has been started. Once the following output is seen
|
||||
from the deployment command, the script should be ready to run:
|
||||
|
||||
[Controller]: CREATE_IN_PROGRESS state changed
|
||||
[NovaCompute]: CREATE_IN_PROGRESS state changed
|
||||
|
||||
The user running the script must be able to ssh as root to each server. Define
|
||||
the names of your custom roles (if applicable) and hostnames of the deployed
|
||||
servers you intend to use for each role type. For each role name, a
|
||||
corresponding <role-name>_hosts variable should also be defined, e.g.::
|
||||
|
||||
export ROLES="Controller NewtorkNode StorageNode Compute"
|
||||
export Controller_hosts="10.0.0.1 10.0.0.2 10.0.0.3"
|
||||
export NetworkNode_hosts="10.0.0.4 10.0.0.5 10.0.0.6"
|
||||
export StorageNode_hosts="10.0.0.7 10.0.08"
|
||||
export Compute_hosts="10.0.0.9 10.0.0.10 10.0.0.11"
|
||||
|
||||
Then run the script on the undercloud with a stackrc file sourced, and
|
||||
the script will copy the needed os-collect-config.conf configuration to each
|
||||
server and restart the os-collect-config service.
|
||||
|
|
|
@ -1,164 +0,0 @@
|
|||
#!/bin/bash
|
||||
|
||||
set -eux
|
||||
|
||||
SLEEP_TIME=2
|
||||
|
||||
CONTROLLER_HOSTS=${CONTROLLER_HOSTS:-""}
|
||||
COMPUTE_HOSTS=${COMPUTE_HOSTS:-""}
|
||||
BLOCKSTORAGE_HOSTS=${BLOCKSTORAGE_HOSTS:-""}
|
||||
OBJECTSTORAGE_HOSTS=${OBJECTSTORAGE_HOSTS:-""}
|
||||
CEPHSTORAGE_HOSTS=${CEPHSTORAGE_HOSTS:-""}
|
||||
SUBNODES_SSH_KEY=${SUBNODES_SSH_KEY:-"~/.ssh/id_rsa"}
|
||||
SSH_OPTIONS=${SSH_OPTIONS:-"-tt -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o LogLevel=Verbose -o PasswordAuthentication=no -o ConnectionAttempts=32"}
|
||||
OVERCLOUD_ROLES=${OVERCLOUD_ROLES:-"Controller Compute BlockStorage ObjectStorage CephStorage"}
|
||||
STACK_NAME=${STACK_NAME:-"overcloud"}
|
||||
|
||||
# Set the _hosts vars for the default roles based on the old var names that
|
||||
# were all caps for backwards compatibility.
|
||||
Controller_hosts=${Controller_hosts:-"$CONTROLLER_HOSTS"}
|
||||
Compute_hosts=${Compute_hosts:-"$COMPUTE_HOSTS"}
|
||||
BlockStorage_hosts=${BlockStorage_hosts:-"$BLOCKSTORAGE_HOSTS"}
|
||||
ObjectStorage_hosts=${ObjectStorage_hosts:-"$OBJECTSTORAGE_HOSTS"}
|
||||
CephStorage_hosts=${CephStorage_hosts:-"$CEPHSTORAGE_HOSTS"}
|
||||
|
||||
#######################################
|
||||
# Retry with backoff interval
|
||||
#######################################
|
||||
function with_backoff {
|
||||
local max_attempts=${ATTEMPTS:-10}
|
||||
local sleep_timeout=${SLEEP_TIME:-2}
|
||||
local attempt=0
|
||||
local rc=0
|
||||
|
||||
while [ ${attempt} -lt ${max_attempts} ]; do
|
||||
set +e
|
||||
set -o pipefail
|
||||
"$@"
|
||||
rc=$?
|
||||
set +o pipefail
|
||||
set -e
|
||||
|
||||
if [ ${rc} -eq 0 ]; then
|
||||
break
|
||||
fi
|
||||
echo "Warning! Retrying in ${sleep_timeout} seconds ..." 1>&2
|
||||
sleep ${sleep_timeout}
|
||||
attempt=$(( attempt + 1 ))
|
||||
sleep_timeout=$(( sleep_timeout * 2 ))
|
||||
done
|
||||
|
||||
if [ ${rc} -ne 0 ]; then
|
||||
echo "Warning! Return code is not 0 on the last try for ($@)" 1>&2
|
||||
fi
|
||||
|
||||
return ${rc}
|
||||
}
|
||||
|
||||
#######################################
|
||||
# Return 1 if empty output received
|
||||
#######################################
|
||||
function fail_if_empty {
|
||||
local output="$(eval "${@}")"
|
||||
if [ -z "${output}" ]; then
|
||||
echo "Warning! Empty output for ($@)" 1>&2
|
||||
return 1
|
||||
else
|
||||
echo "${output}"
|
||||
fi
|
||||
}
|
||||
|
||||
function check_stack {
|
||||
local stack_to_check=${1:-""}
|
||||
local rc=0
|
||||
|
||||
if [ -z "${stack_to_check}" ]; then
|
||||
echo No Stacks received.
|
||||
return 1
|
||||
fi
|
||||
|
||||
with_backoff openstack stack resource list $stack_to_check
|
||||
rc=${?}
|
||||
|
||||
if [ ${rc} -ne 0 ]; then
|
||||
echo Stack ${stack_to_check} not yet created
|
||||
fi
|
||||
|
||||
return ${rc}
|
||||
}
|
||||
|
||||
# Set the _hosts_a vars for each role defined
|
||||
for role in $OVERCLOUD_ROLES; do
|
||||
eval "hosts=\${${role}_hosts}"
|
||||
read -a ${role}_hosts_a <<< $hosts
|
||||
done
|
||||
|
||||
for role in $OVERCLOUD_ROLES; do
|
||||
while ! check_stack $STACK_NAME; do
|
||||
sleep $SLEEP_TIME
|
||||
done
|
||||
|
||||
rg_stack=$(with_backoff fail_if_empty openstack stack resource show $STACK_NAME $role -c physical_resource_id -f value)
|
||||
while ! check_stack $rg_stack; do
|
||||
rg_stack=$(with_backoff fail_if_empty openstack stack resource show $STACK_NAME $role -c physical_resource_id -f value)
|
||||
done
|
||||
|
||||
stacks=$(with_backoff fail_if_empty "openstack stack resource list $rg_stack -c resource_name -c physical_resource_id -f json | jq -r 'sort_by(.resource_name | tonumber ) | .[] | .physical_resource_id'")
|
||||
rc=${?}
|
||||
while [ ${rc} -ne 0 ]; do
|
||||
stacks=$(with_backoff fail_if_empty "openstack stack resource list $rg_stack -c resource_name -c physical_resource_id -f json | jq -r 'sort_by(.resource_name | tonumber) | .[] | .physical_resource_id'")
|
||||
done
|
||||
|
||||
i=0
|
||||
|
||||
for stack in $stacks; do
|
||||
server_resource_name=$role
|
||||
if [ "$server_resource_name" = "Compute" ]; then
|
||||
server_resource_name="NovaCompute"
|
||||
fi
|
||||
|
||||
server_stack=$(with_backoff fail_if_empty openstack stack resource show $stack $server_resource_name -c physical_resource_id -f value)
|
||||
while ! check_stack $server_stack; do
|
||||
server_stack=$(with_backoff fail_if_empty openstack stack resource show $stack $server_resource_name -c physical_resource_id -f value)
|
||||
done
|
||||
|
||||
while true; do
|
||||
deployed_server_metadata_url=$(with_backoff openstack stack resource metadata $server_stack deployed-server | jq -r '.["os-collect-config"].request.metadata_url')
|
||||
if [ "$deployed_server_metadata_url" != "null" ]; then
|
||||
break
|
||||
fi
|
||||
sleep $SLEEP_TIME
|
||||
done
|
||||
|
||||
echo "======================"
|
||||
echo "$role$i deployed-server.json configuration:"
|
||||
|
||||
config="{
|
||||
\"os-collect-config\": {
|
||||
\"collectors\": [\"request\", \"local\"],
|
||||
\"request\": {
|
||||
\"metadata_url\": \"$deployed_server_metadata_url\"
|
||||
}
|
||||
}
|
||||
}"
|
||||
|
||||
echo "$config"
|
||||
echo "======================"
|
||||
echo
|
||||
|
||||
|
||||
host=
|
||||
eval "host=\${${role}_hosts_a[$i]}"
|
||||
if [ -n "$host" ]; then
|
||||
ssh $SSH_OPTIONS -i $SUBNODES_SSH_KEY $host "echo '$config' > deployed-server.json"
|
||||
ssh $SSH_OPTIONS -i $SUBNODES_SSH_KEY $host sudo mkdir -p -m 0700 /var/lib/os-collect-config/local-data/ || true
|
||||
ssh $SSH_OPTIONS -i $SUBNODES_SSH_KEY $host sudo cp deployed-server.json /var/lib/os-collect-config/local-data/deployed-server.json
|
||||
ssh $SSH_OPTIONS -i $SUBNODES_SSH_KEY $host sudo systemctl start os-collect-config
|
||||
ssh $SSH_OPTIONS -i $SUBNODES_SSH_KEY $host sudo systemctl enable os-collect-config
|
||||
fi
|
||||
|
||||
let i+=1
|
||||
|
||||
done
|
||||
|
||||
done
|
Loading…
Reference in New Issue