4f791902ae
I7717f2ce9f668e9fd77c9b49e2433d729ed3c836 will ensure CephClientConfigVars is set for deployed ceph standalone scenarios. Change-Id: I6fe456fe153d10a9958230bd7dcc6f25e350da53
215 lines
9.1 KiB
YAML
215 lines
9.1 KiB
YAML
resource_registry:
|
|
OS::TripleO::Services::AodhApi: ../../deployment/aodh/aodh-api-container-puppet.yaml
|
|
OS::TripleO::Services::AodhEvaluator: ../../deployment/aodh/aodh-evaluator-container-puppet.yaml
|
|
OS::TripleO::Services::AodhListener: ../../deployment/aodh/aodh-listener-container-puppet.yaml
|
|
OS::TripleO::Services::AodhNotifier: ../../deployment/aodh/aodh-notifier-container-puppet.yaml
|
|
OS::TripleO::Services::CeilometerAgentCentral: ../../deployment/ceilometer/ceilometer-agent-central-container-puppet.yaml
|
|
OS::TripleO::Services::CeilometerAgentNotification: ../../deployment/ceilometer/ceilometer-agent-notification-container-puppet.yaml
|
|
OS::TripleO::Services::ComputeCeilometerAgent: ../../deployment/ceilometer/ceilometer-agent-compute-container-puppet.yaml
|
|
OS::TripleO::Services::CephMgr: ../../deployment/cephadm/ceph-mgr.yaml
|
|
OS::TripleO::Services::CephMon: ../../deployment/cephadm/ceph-mon.yaml
|
|
OS::TripleO::Services::CephOSD: ../../deployment/cephadm/ceph-osd.yaml
|
|
OS::TripleO::Services::CephGrafana: ../../deployment/cephadm/ceph-grafana.yaml
|
|
OS::TripleO::Services::CephClient: ../../deployment/cephadm/ceph-client.yaml
|
|
OS::TripleO::Services::CephRgw: ../../deployment/cephadm/ceph-rgw.yaml
|
|
OS::TripleO::Services::Collectd: ../../deployment/metrics/collectd-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiApi: ../../deployment/gnocchi/gnocchi-api-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiMetricd: ../../deployment/gnocchi/gnocchi-metricd-container-puppet.yaml
|
|
OS::TripleO::Services::GnocchiStatsd: ../../deployment/gnocchi/gnocchi-statsd-container-puppet.yaml
|
|
OS::TripleO::Services::HeatApi: ../../deployment/heat/heat-api-container-puppet.yaml
|
|
OS::TripleO::Services::HeatApiCfn: ../../deployment/heat/heat-api-cfn-container-puppet.yaml
|
|
OS::TripleO::Services::HeatEngine: ../../deployment/heat/heat-engine-container-puppet.yaml
|
|
OS::TripleO::Services::Horizon: ../../deployment/horizon/horizon-container-puppet.yaml
|
|
OS::TripleO::Services::MetricsQdr: ../../deployment/metrics/qdr-container-puppet.yaml
|
|
OS::TripleO::Services::Multipathd: ../../deployment/multipathd/multipathd-container-ansible.yaml
|
|
OS::TripleO::Services::OsloMessagingRpc: ../../deployment/rabbitmq/rabbitmq-messaging-rpc-pacemaker-puppet.yaml
|
|
OS::TripleO::Services::OsloMessagingNotify: ../../deployment/rabbitmq/rabbitmq-messaging-notify-shared-puppet.yaml
|
|
OS::TripleO::Services::Redis: ../../deployment/database/redis-pacemaker-puppet.yaml
|
|
OS::TripleO::Services::Rsyslog: ../../deployment/logging/rsyslog-container-puppet.yaml
|
|
OS::TripleO::Services::CinderBackup: ../../deployment/cinder/cinder-backup-pacemaker-puppet.yaml
|
|
OS::TripleO::Services::CinderVolume: ../../deployment/cinder/cinder-volume-pacemaker-puppet.yaml
|
|
OS::TripleO::Services::SwiftProxy: OS::Heat::None
|
|
OS::TripleO::Services::SwiftStorage: OS::Heat::None
|
|
OS::TripleO::Services::SwiftRingBuilder: OS::Heat::None
|
|
|
|
parameter_defaults:
|
|
NodeDataLookup:
|
|
AB4114B1-9C9D-409A-BEFB-D88C151BF2C3: {"foo": "bar"}
|
|
8CF1A7EA-7B4B-4433-AC83-17675514B1B8: {"foo2": "bar2"}
|
|
StandaloneExtraConfig:
|
|
# NOTE(sileht): To decrease the time test_telemetry_integration takes We
|
|
# configure Ceilometer to poll more, We configure the 'high' Gnocchi
|
|
# archive policy to keep 1 point every 60s. The test will take 2 minutes
|
|
# instead of 10 minutes. Note that tempest telemetry.alarm_granularity must
|
|
# in sync with the archive policy, 60s too.
|
|
ceilometer::agent::polling::polling_interval: 15
|
|
tripleo::profile::base::metrics::qdr::interior_mesh_nodes: ''
|
|
NotificationDriver: 'messagingv2'
|
|
ManagePolling: true
|
|
HeatConfigureDelegatedRoles: true
|
|
CeilometerEnableGnocchi: true
|
|
PipelinePublishers:
|
|
- gnocchi://?archive_policy=ceilometer-high-rate
|
|
EventPipelinePublishers:
|
|
- gnocchi://?archive_policy=ceilometer-high-rate
|
|
CeilometerQdrPublishEvents: true
|
|
CeilometerQdrPublishMetrics: true
|
|
ManageEventPipeline: true
|
|
ManagePipeline: true
|
|
Debug: true
|
|
DeployedCeph: true
|
|
CephEnableDashboard: true
|
|
CephDashboardPort: 8445
|
|
GrafanaDashboardPort: 3200
|
|
CinderRbdExtraPools: altrbd,pool2,pool3
|
|
CephPools:
|
|
- name: altrbd
|
|
rule_name: replicated_rule
|
|
application: rbd
|
|
#NOTE: These ID's and keys should be regenerated for
|
|
# a production deployment. What is here is suitable for
|
|
# developer and CI testing only.
|
|
CephClientKey: 'AQC+vYNXgDAgAhAAc8UoYt+OTz5uhV7ItLdwUw=='
|
|
CephExtraKeys:
|
|
- name: "client.glance"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "profile rbd pool=images"
|
|
key: "AQBRgQ9eAAAAABAAv84zEilJYZPNuJ0Iwn9Ndg=="
|
|
mode: "0600"
|
|
CephExternalMultiConfig:
|
|
# create client conf and key file for two non-existent external ceph clusters
|
|
- cluster: 'ceph2'
|
|
fsid: 'af25554b-42f6-4d2b-9b9b-d08a1132d3e8'
|
|
external_cluster_mon_ips: '172.18.0.5,172.18.0.6,172.18.0.7'
|
|
keys:
|
|
- name: "client.openstack"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "osd: profile rbd pool=volumes, profile rbd pool=backups, profile rbd pool=vms, profile rbd pool=images"
|
|
key: "AQCwmeRcAAAAABAA6SQU/bGqFjlfLro5KxrB1Q=="
|
|
mode: "0600"
|
|
dashboard_enabled: false
|
|
- cluster: 'ceph3'
|
|
fsid: 'e2cba068-5f14-4b0f-b047-acf375c0004a'
|
|
external_cluster_mon_ips: '172.18.0.8,172.18.0.9,172.18.0.10'
|
|
keys:
|
|
- name: "client.openstack"
|
|
caps:
|
|
mgr: "allow *"
|
|
mon: "profile rbd"
|
|
osd: "osd: profile rbd pool=volumes, profile rbd pool=backups, profile rbd pool=vms, profile rbd pool=images"
|
|
key: "AQCwmeRcAAAAABAA6SQU/bGqFjlfLro5KxrB2Q=="
|
|
mode: "0600"
|
|
dashboard_enabled: false
|
|
NovaEnableRbdBackend: true
|
|
CinderEnableRbdBackend: true
|
|
CinderRbdBackendName: tripleo_ceph,tripleo_ceph2,tripleo_ceph3
|
|
CinderRbdMultiConfig:
|
|
tripleo_ceph2:
|
|
CephClusterName: ceph2
|
|
CephClusterFSID: af25554b-42f6-4d2b-9b9b-d08a1132d3e8
|
|
CinderRbdAvailabilityZone: ceph2-AZ
|
|
tripleo_ceph3:
|
|
CephClusterName: ceph3
|
|
CephClusterFSID: e2cba068-5f14-4b0f-b047-acf375c0004a
|
|
CinderRbdAvailabilityZone: ceph3-AZ
|
|
CinderBackupBackend: ceph
|
|
GlanceEnabledImportMethods: 'glance-direct,web-download,copy-image'
|
|
# For copy-image method, we are overriding the policy here to allow
|
|
# everyone and every type of image (private or public) to copy.
|
|
# This way we will be able to test copy image via non-admin as well
|
|
# as on private images. The policy is default to admin only in glance
|
|
GlanceApiPolicies:
|
|
glance-copy_image:
|
|
key: "copy_image"
|
|
value: ""
|
|
GlanceBackend: cinder
|
|
GlanceStoreDescription: 'Cinder glance store'
|
|
GlanceMultistoreConfig:
|
|
rbd_store:
|
|
GlanceBackend: rbd
|
|
GlanceStoreDescription: 'RBD glance store'
|
|
CephClientUserName: 'glance'
|
|
CephClusterName: ceph
|
|
GnocchiBackend: rbd
|
|
CinderEnableIscsiBackend: false
|
|
BannerText: |
|
|
******************************************************************
|
|
* This system is for the use of authorized users only. Usage of *
|
|
* this system may be monitored and recorded by system personnel. *
|
|
* Anyone using this system expressly consents to such monitoring *
|
|
* and is advised that if such monitoring reveals possible *
|
|
* evidence of criminal activity, system personnel may provide *
|
|
* the evidence from such monitoring to law enforcement officials.*
|
|
******************************************************************
|
|
EnableSTF: true
|
|
MetricsQdrAddresses:
|
|
- prefix: collectd
|
|
distribution: multicast
|
|
MetricsQdrSSLProfiles:
|
|
- name: sslProfile
|
|
CollectdConnectionType: amqp1
|
|
CollectdAmqpInterval: 5
|
|
CollectdDefaultPollingInterval: 5
|
|
CollectdEnableSensubility: true
|
|
CollectdEnableLibpodstats: true
|
|
CollectdAmqpInstances:
|
|
notify:
|
|
notify: true
|
|
format: JSON
|
|
presettle: false
|
|
telemetry:
|
|
format: JSON
|
|
presettle: false
|
|
CollectdAmqpSendQueueLimit: 40
|
|
LoggingServers:
|
|
- host: 127.0.0.1
|
|
port: 24224
|
|
TtyValues:
|
|
- console
|
|
- tty1
|
|
- tty2
|
|
- tty3
|
|
- tty4
|
|
- tty5
|
|
- tty6
|
|
ContainerCli: podman
|
|
CephAdmVerbose: true
|
|
CephConfigOverrides:
|
|
# put logs in /var/log/ceph/<cluster-fsid>
|
|
log_to_file: true
|
|
mon_cluster_log_to_file: true
|
|
# disable logging to journald so we don't log twice
|
|
log_to_stderr: false
|
|
mon_cluster_log_to_stderr: false
|
|
log_to_journald: false
|
|
mon_cluster_log_to_journald: false
|
|
# test values
|
|
anotherkey: anothervalue
|
|
global:
|
|
globalkey: globalvalue
|
|
osd:
|
|
osdkey: osdvalue
|
|
mon:
|
|
mon_warn_on_pool_no_redundancy: false
|
|
mon_warn_on_insecure_global_id_reclaim_allowed: false
|
|
NfsUrl: 127.0.0.1
|
|
CephMsgrSecureMode: true
|
|
CephConfigPath: "/etc/ceph"
|
|
DisableCephadm: true
|
|
LVMFilterEnabled: true
|
|
LVMFilterAllowlist:
|
|
- /dev/loop2
|
|
- /dev/loop3
|
|
- /dev/loop4
|
|
MultipathdEnable: true
|
|
NovaLibvirtVolumeUseMultipath: true
|
|
NovaShowHostStatus: all
|
|
NovaApiHostStatusPolicy: >
|
|
(role:reader and system_scope:all) or (role:reader and project_id:%(project_id)s)
|
|
RsyslogElasticsearchSetting:
|
|
Server: '127.0.0.1:9200'
|