f55af442de
The puppet and ansible tasks for deploying the Multipathd service have been moved to a new tripleo_multipathd ansible role. THT uses that role passing in parameter values via ansible variables. Two new THT parameters are supported: - MultipathdCustomConfigFile provides a way for the user to specify a custom multipath.conf file. This makes it easier for the user to manage custom settings that are unique to their deployment, such as vendor specific hardware tuning. - MultipathdSkipKpartx provides a means for overriding multipathd's default behavior whereby it automatically creates disk partitions on multipath devices. Partitions should be managed by VM servers accessing the device, not by the overcloud host itself. Depends-On: Icf9faff31d83f0ea77d00a59a53d6ad36b06da4f Depends-On: I3478312b5117da5c2e819e47c99f574246e84838 Change-Id: I934a1ae5bc0d77fd39b25f5039635e5df6e9004f (cherry picked from commit80fcd33248
) (cherry picked from commita19a3c6aef
)
202 lines
8.3 KiB
YAML
202 lines
8.3 KiB
YAML
resource_registry:
|
|
OS::TripleO::Services::AodhApi: ../../deployment/aodh/aodh-api-container-puppet.yaml
|
|
OS::TripleO::Services::AodhEvaluator: ../../deployment/aodh/aodh-evaluator-container-puppet.yaml
|
|
OS::TripleO::Services::AodhListener: ../../deployment/aodh/aodh-listener-container-puppet.yaml
|
|
OS::TripleO::Services::AodhNotifier: ../../deployment/aodh/aodh-notifier-container-puppet.yaml
|
|
OS::TripleO::Services::CeilometerAgentCentral: ../../deployment/ceilometer/ceilometer-agent-central-container-puppet.yaml
|
|
OS::TripleO::Services::CeilometerAgentNotification: ../../deployment/ceilometer/ceilometer-agent-notification-container-puppet.yaml
|
|
OS::TripleO::Services::ComputeCeilometerAgent: ../../deployment/ceilometer/ceilometer-agent-compute-container-puppet.yaml
|
|
OS::TripleO::Services::CephMgr: ../../deployment/ceph-ansible/ceph-mgr.yaml
|
|
OS::TripleO::Services::CephMon: ../../deployment/ceph-ansible/ceph-mon.yaml
|
|
OS::TripleO::Services::CephOSD: ../../deployment/ceph-ansible/ceph-osd.yaml
|
|
OS::TripleO::Services::CephGrafana: ../../deployment/ceph-ansible/ceph-grafana.yaml
|
|
OS::TripleO::Services::CephClient: ../../deployment/ceph-ansible/ceph-client.yaml
|
|
OS::TripleO::Services::PankoApi: ../../deployment/deprecated/panko/panko-api-container-puppet.yaml
|
|
OS::TripleO::Services::Collectd: ../../deployment/metrics/collectd-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiApi: ../../deployment/gnocchi/gnocchi-api-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiMetricd: ../../deployment/gnocchi/gnocchi-metricd-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiStatsd: ../../deployment/gnocchi/gnocchi-statsd-container-puppet.yaml
|
|
OS::TripleO::Services::HeatApi: ../../deployment/heat/heat-api-container-puppet.yaml
|
|
OS::TripleO::Services::HeatApiCfn: ../../deployment/heat/heat-api-cfn-container-puppet.yaml
|
|
OS::TripleO::Services::HeatApiCloudwatch: ../../deployment/heat/heat-api-cloudwatch-disabled-puppet.yaml
|
|
OS::TripleO::Services::HeatEngine: ../../deployment/heat/heat-engine-container-puppet.yaml
|
|
OS::TripleO::Services::MetricsQdr: ../../deployment/metrics/qdr-container-puppet.yaml
|
|
OS::TripleO::Services::Multipathd: ../../deployment/multipathd/multipathd-container-ansible.yaml
|
|
OS::TripleO::Services::OsloMessagingRpc: ../../deployment/rabbitmq/rabbitmq-messaging-rpc-container-puppet.yaml
|
|
OS::TripleO::Services::OsloMessagingNotify: ../../deployment/rabbitmq/rabbitmq-messaging-notify-shared-puppet.yaml
|
|
OS::TripleO::Services::Redis: ../../deployment/database/redis-container-puppet.yaml
|
|
OS::TripleO::Services::CinderBackup: ../../deployment/cinder/cinder-backup-container-puppet.yaml
|
|
OS::TripleO::Services::CinderVolume: ../../deployment/cinder/cinder-volume-container-puppet.yaml
|
|
OS::TripleO::Services::SwiftProxy: OS::Heat::None
|
|
OS::TripleO::Services::SwiftStorage: OS::Heat::None
|
|
OS::TripleO::Services::SwiftRingBuilder: OS::Heat::None
|
|
OS::TripleO::Services::Horizon: OS::Heat::None
|
|
|
|
parameter_defaults:
|
|
NodeDataLookup:
|
|
AB4114B1-9C9D-409A-BEFB-D88C151BF2C3: {"foo": "bar"}
|
|
8CF1A7EA-7B4B-4433-AC83-17675514B1B8: {"foo2": "bar2"}
|
|
StandaloneExtraConfig:
|
|
# NOTE(sileht): To decrease the time test_telemetry_integration takes We
|
|
# configure Ceilometer to poll more, We configure the 'high' Gnocchi
|
|
# archive policy to keep 1 point every 60s. The test will take 2 minutes
|
|
# instead of 10 minutes. Note that tempest telemetry.alarm_granularity must
|
|
# in sync with the archive policy, 60s too.
|
|
ceilometer::agent::polling::polling_interval: 15
|
|
tripleo::profile::base::metrics::qdr::interior_mesh_nodes: ''
|
|
NotificationDriver: 'messagingv2'
|
|
ManagePolling: true
|
|
HeatConfigureDelegatedRoles: true
|
|
CeilometerEnableGnocchi: true
|
|
PipelinePublishers:
|
|
- gnocchi://?archive_policy=ceilometer-high-rate
|
|
EventPipelinePublishers:
|
|
- gnocchi://?archive_policy=ceilometer-high-rate
|
|
- panko://
|
|
CeilometerQdrPublishEvents: true
|
|
CeilometerQdrPublishMetrics: true
|
|
ManageEventPipeline: true
|
|
ManagePipeline: true
|
|
|
|
Debug: true
|
|
CephAnsibleDisksConfig:
|
|
osd_objectstore: bluestore
|
|
osd_scenario: lvm
|
|
lvm_volumes:
|
|
- data: ceph_lv_data
|
|
data_vg: ceph_vg
|
|
db: ceph_lv_db
|
|
db_vg: ceph_vg
|
|
wal: ceph_lv_wal
|
|
wal_vg: ceph_vg
|
|
CephPoolDefaultPgNum: 32
|
|
CephPoolDefaultSize: 1
|
|
CephEnableDashboard: true
|
|
CephAnsibleRepo: "tripleo-centos-ceph-nautilus"
|
|
CinderRbdExtraPools: altrbd,pool2,pool3
|
|
CephPools:
|
|
- name: altrbd
|
|
pg_num: 8
|
|
rule_name: replicated_rule
|
|
CephAnsibleExtraConfig:
|
|
centos_package_dependencies: []
|
|
ceph_osd_docker_memory_limit: '1g'
|
|
ceph_mds_docker_memory_limit: '1g'
|
|
mon_host_v1: { 'enabled': False }
|
|
handler_health_mon_check_retries: 10
|
|
handler_health_mon_check_delay: 20
|
|
#NOTE: These ID's and keys should be regenerated for
|
|
# a production deployment. What is here is suitable for
|
|
# developer and CI testing only.
|
|
CephClusterFSID: '4b5c8c0a-ff60-454b-a1b4-9747aa737d19'
|
|
CephClientKey: 'AQC+vYNXgDAgAhAAc8UoYt+OTz5uhV7ItLdwUw=='
|
|
CephExtraKeys:
|
|
- name: "client.glance"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "profile rbd pool=images"
|
|
key: "AQBRgQ9eAAAAABAAv84zEilJYZPNuJ0Iwn9Ndg=="
|
|
mode: "0600"
|
|
CephExternalMultiConfig:
|
|
# create client conf and key file for two non-existent external ceph clusters
|
|
- cluster: 'ceph2'
|
|
fsid: 'af25554b-42f6-4d2b-9b9b-d08a1132d3e8'
|
|
external_cluster_mon_ips: '172.18.0.5,172.18.0.6,172.18.0.7'
|
|
keys:
|
|
- name: "client.openstack"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "osd: profile rbd pool=volumes, profile rbd pool=backups, profile rbd pool=vms, profile rbd pool=images"
|
|
key: "AQCwmeRcAAAAABAA6SQU/bGqFjlfLro5KxrB1Q=="
|
|
mode: "0600"
|
|
dashboard_enabled: false
|
|
- cluster: 'ceph3'
|
|
fsid: 'e2cba068-5f14-4b0f-b047-acf375c0004a'
|
|
external_cluster_mon_ips: '172.18.0.8,172.18.0.9,172.18.0.10'
|
|
keys:
|
|
- name: "client.openstack"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "osd: profile rbd pool=volumes, profile rbd pool=backups, profile rbd pool=vms, profile rbd pool=images"
|
|
key: "AQCwmeRcAAAAABAA6SQU/bGqFjlfLro5KxrB2Q=="
|
|
mode: "0600"
|
|
dashboard_enabled: false
|
|
CephAnsiblePlaybookVerbosity: 1
|
|
CephAnsibleEnvironmentVariables:
|
|
ANSIBLE_SSH_RETRIES: 4
|
|
DEFAULT_FORKS: 3
|
|
NovaEnableRbdBackend: true
|
|
CinderEnableRbdBackend: true
|
|
CinderBackupBackend: ceph
|
|
GlanceBackend: cinder
|
|
GlanceStoreDescription: 'Cinder glance store'
|
|
GlanceMultistoreConfig:
|
|
rbd_store:
|
|
GlanceBackend: rbd
|
|
GlanceStoreDescription: 'RBD glance store'
|
|
CephClientUserName: 'glance'
|
|
GnocchiBackend: rbd
|
|
CinderEnableIscsiBackend: false
|
|
GnocchiArchivePolicy: 'ceilometer-high-rate'
|
|
BannerText: |
|
|
******************************************************************
|
|
* This system is for the use of authorized users only. Usage of *
|
|
* this system may be monitored and recorded by system personnel. *
|
|
* Anyone using this system expressly consents to such monitoring *
|
|
* and is advised that if such monitoring reveals possible *
|
|
* evidence of criminal activity, system personnel may provide *
|
|
* the evidence from such monitoring to law enforcement officials.*
|
|
******************************************************************
|
|
EnableSTF: true
|
|
MetricsQdrAddresses:
|
|
- prefix: collectd
|
|
distribution: multicast
|
|
MetricsQdrSSLProfiles:
|
|
- name: sslProfile
|
|
CollectdConnectionType: amqp1
|
|
CollectdAmqpInterval: 5
|
|
CollectdDefaultPollingInterval: 5
|
|
CollectdExtraPlugins:
|
|
- rrdtool
|
|
CollectdEnableSensubility: true
|
|
CollectdEnableLibpodstats: true
|
|
CollectdAmqpInstances:
|
|
notify:
|
|
notify: true
|
|
format: JSON
|
|
presettle: false
|
|
telemetry:
|
|
format: JSON
|
|
presettle: false
|
|
LoggingServers:
|
|
- host: 127.0.0.1
|
|
port: 24224
|
|
TtyValues:
|
|
- console
|
|
- tty1
|
|
- tty2
|
|
- tty3
|
|
- tty4
|
|
- tty5
|
|
- tty6
|
|
ContainerCli: podman
|
|
CephConfigOverrides:
|
|
anotherkey: anothervalue
|
|
global:
|
|
globalkey: globalvalue
|
|
osd:
|
|
osdkey: osdvalue
|
|
foo: bar
|
|
NfsUrl: 127.0.0.1
|
|
CephMsgrSecureMode: true
|
|
LVMFilterEnabled: true
|
|
LVMFilterAllowlist:
|
|
- /dev/loop2
|
|
- /dev/loop3
|
|
- /dev/loop4
|
|
MultipathdEnable: true
|
|
NovaLibvirtVolumeUseMultipath: true
|