resource_registry: OS::TripleO::Services::AodhApi: ../../deployment/aodh/aodh-api-container-puppet.yaml OS::TripleO::Services::AodhEvaluator: ../../deployment/aodh/aodh-evaluator-container-puppet.yaml OS::TripleO::Services::AodhListener: ../../deployment/aodh/aodh-listener-container-puppet.yaml OS::TripleO::Services::AodhNotifier: ../../deployment/aodh/aodh-notifier-container-puppet.yaml OS::TripleO::Services::CeilometerAgentCentral: ../../docker/services/ceilometer-agent-central.yaml OS::TripleO::Services::CeilometerAgentNotification: ../../docker/services/ceilometer-agent-notification.yaml OS::TripleO::Services::ComputeCeilometerAgent: ../../docker/services/ceilometer-agent-compute.yaml OS::TripleO::Services::CephMgr: ../../docker/services/ceph-ansible/ceph-mgr.yaml OS::TripleO::Services::CephMon: ../../docker/services/ceph-ansible/ceph-mon.yaml OS::TripleO::Services::CephOSD: ../../docker/services/ceph-ansible/ceph-osd.yaml OS::TripleO::Services::CephClient: ../../docker/services/ceph-ansible/ceph-client.yaml OS::TripleO::Services::PankoApi: ../../deployment/panko/panko-api-container-puppet.yaml OS::TripleO::Services::Collectd: ../../docker/services/metrics/collectd.yaml OS::TripleO::Services::GnocchiApi: ../../docker/services/gnocchi-api.yaml OS::TripleO::Services::GnocchiMetricd: ../../docker/services/gnocchi-metricd.yaml OS::TripleO::Services::GnocchiStatsd: ../../docker/services/gnocchi-statsd.yaml OS::TripleO::Services::HeatApi: ../../deployment/heat/heat-api-container-puppet.yaml OS::TripleO::Services::HeatApiCfn: ../../deployment/heat/heat-api-cfn-container-puppet.yaml OS::TripleO::Services::HeatApiCloudwatch: ../../deployment/heat/heat-api-cloudwatch-disabled-puppet.yaml OS::TripleO::Services::HeatEngine: ../../deployment/heat/heat-engine-container-puppet.yaml OS::TripleO::Services::MetricsQdr: ../../deployment/metrics/qdr-container-puppet.yaml OS::TripleO::Services::OsloMessagingRpc: ../../deployment/rabbitmq/rabbitmq-messaging-rpc-pacemaker-puppet.yaml OS::TripleO::Services::OsloMessagingNotify: ../../deployment/rabbitmq/rabbitmq-messaging-notify-shared-puppet.yaml OS::TripleO::Services::HAproxy: ../../deployment/haproxy/haproxy-pacemaker-puppet.yaml OS::TripleO::Services::Pacemaker: ../../puppet/services/pacemaker.yaml OS::TripleO::Services::PacemakerRemote: ../../puppet/services/pacemaker_remote.yaml OS::TripleO::Services::Clustercheck: ../../docker/services/pacemaker/clustercheck.yaml OS::TripleO::Services::Redis: ../../docker/services/pacemaker/database/redis.yaml OS::TripleO::Services::MySQL: ../../docker/services/pacemaker/database/mysql.yaml OS::TripleO::Services::CinderBackup: ../../deployment/cinder/cinder-backup-pacemaker-puppet.yaml OS::TripleO::Services::CinderVolume: ../../deployment/cinder/cinder-volume-pacemaker-puppet.yaml OS::TripleO::Services::Keepalived: OS::Heat::None #Needs to run scenario001 OS::TripleO::Services::Fluentd: ../../deployment/logging/fluentd-container-puppet.yaml # NOTE(mmagr): We need to disable Sensu client deployment for now as the container health check is based # on successful RabbitMQ connection, which does not happen in this case. We can enable it again when we # will implement default connection to overcloud RabbitMQ instance, #OS::TripleO::Services::SensuClient: ../../deployment/monitoring/sensu-client-container-puppet.yaml # Some infra instances don't pass the ping test but are otherwise working. # Since the OVB jobs also test this functionality we can shut it off here. OS::TripleO::AllNodes::Validation: ../common/all-nodes-validation-disabled.yaml OS::TripleO::Services::SwiftProxy: OS::Heat::None OS::TripleO::Services::SwiftStorage: OS::Heat::None OS::TripleO::Services::SwiftRingBuilder: OS::Heat::None OS::TripleO::Services::Horizon: OS::Heat::None parameter_defaults: NodeDataLookup: AB4114B1-9C9D-409A-BEFB-D88C151BF2C3: {"foo": "bar"} 8CF1A7EA-7B4B-4433-AC83-17675514B1B8: {"foo2": "bar2"} StandaloneExtraConfig: # NOTE(sileht): To decrease the time test_telemetry_integration takes We # configure Ceilometer to poll more, We configure the 'high' Gnocchi # archive policy to keep 1 point every 60s. The test will take 2 minutes # instead of 10 minutes. Note that tempest telemetry.alarm_granularity must # in sync with the archive policy, 60s too. ceilometer::agent::polling::polling_interval: 15 ManagePolling: true ManagePipeline: true PipelinePublishers: - gnocchi://?archive_policy=ceilometer-high-rate EventPipelinePublishers: - gnocchi://?archive_policy=ceilometer-high-rate - panko:// Debug: true # fetch dir needed for standalone LocalCephAnsibleFetchDirectoryBackup: /var/lib/ceph_ansible_fetch CephAnsibleDisksConfig: devices: - /dev/loop3 journal_size: 512 osd_scenario: collocated CephPoolDefaultPgNum: 32 CephPoolDefaultSize: 1 CephAnsibleExtraConfig: centos_package_dependencies: [] ceph_osd_docker_memory_limit: '1g' ceph_mds_docker_memory_limit: '1g' #NOTE: These ID's and keys should be regenerated for # a production deployment. What is here is suitable for # developer and CI testing only. CephClusterFSID: '4b5c8c0a-ff60-454b-a1b4-9747aa737d19' CephMonKey: 'AQC+Ox1VmEr3BxAALZejqeHj50Nj6wJDvs96OQ==' CephAdminKey: 'AQDLOh1VgEp6FRAAFzT7Zw+Y9V6JJExQAsRnRQ==' CephClientKey: 'AQC+vYNXgDAgAhAAc8UoYt+OTz5uhV7ItLdwUw==' CephAnsiblePlaybookVerbosity: 1 CephAnsibleEnvironmentVariables: ANSIBLE_SSH_RETRIES: '4' DEFAULT_FORKS: '3' NovaEnableRbdBackend: true CinderEnableRbdBackend: true CinderBackupBackend: ceph GlanceBackend: rbd GnocchiBackend: rbd CinderEnableIscsiBackend: false GnocchiArchivePolicy: 'ceilometer-high-rate' BannerText: | ****************************************************************** * This system is for the use of authorized users only. Usage of * * this system may be monitored and recorded by system personnel. * * Anyone using this system expressly consents to such monitoring * * and is advised that if such monitoring reveals possible * * evidence of criminal activity, system personnel may provide * * the evidence from such monitoring to law enforcement officials.* ****************************************************************** CollectdExtraPlugins: - rrdtool LoggingServers: - host: 127.0.0.1 port: 24224 MonitoringRabbitHost: 127.0.0.1 MonitoringRabbitPort: 5676 MonitoringRabbitPassword: sensu TtyValues: - console - tty1 - tty2 - tty3 - tty4 - tty5 - tty6