From dc1ef4c138516cf517911227e73b15f3d490ed3d Mon Sep 17 00:00:00 2001 From: Steve Wilkerson Date: Sat, 2 Feb 2019 16:43:14 -0600 Subject: [PATCH] Update openstack-helm Armada job This updates the openstack-helm Armada job to instead deploy only Ceph, the OpenStack service charts, and their dependencies. This is dependent on the addition of the Armada job for Ceph and the LMA components to openstack-helm-infra. This also updates the jobs definition to use the osh-gate-runner playbook instead, as well as sets the job both to a periodic and experimental job Depends-On: https://review.openstack.org/#/c/634676/ Depends-On: https://review.openstack.org/#/c/633067/ Change-Id: I7e191a153f123e04e123acc33fb691d8117062a9 --- .../armada/020-armada-render-manifests.sh | 11 +- .../armada/025-armada-validate-manifests.sh | 2 +- .../armada/030-armada-apply-manifests.sh | 2 +- ...passwords.sh => generate-osh-passwords.sh} | 14 +- .../armada/manifests/armada-ceph.yaml | 327 ---- .../manifests/armada-cluster-ingress.yaml | 81 - .../armada/manifests/armada-lma.yaml | 1340 ----------------- .../armada/manifests/armada-osh.yaml | 20 +- .../playbooks/armada-fullstack-deploy.yaml | 74 - .../playbooks/gather-armada-manifests.yaml | 1 - zuul.d/jobs-openstack-helm.yaml | 16 +- zuul.d/project.yaml | 3 +- 12 files changed, 35 insertions(+), 1856 deletions(-) rename tools/deployment/armada/{generate-passwords.sh => generate-osh-passwords.sh} (79%) delete mode 100644 tools/deployment/armada/manifests/armada-ceph.yaml delete mode 100644 tools/deployment/armada/manifests/armada-cluster-ingress.yaml delete mode 100644 tools/deployment/armada/manifests/armada-lma.yaml delete mode 100644 tools/gate/playbooks/armada-fullstack-deploy.yaml diff --git a/tools/deployment/armada/020-armada-render-manifests.sh b/tools/deployment/armada/020-armada-render-manifests.sh index b1f4f359d0..d8fdb9f7d4 100755 --- a/tools/deployment/armada/020-armada-render-manifests.sh +++ b/tools/deployment/armada/020-armada-render-manifests.sh @@ -16,10 +16,11 @@ set -xe -source ./tools/deployment/armada/generate-passwords.sh : ${OSH_INFRA_PATH:="../openstack-helm-infra"} : ${OSH_PATH:="./"} +source ./tools/deployment/armada/generate-osh-passwords.sh + [ -s /tmp/ceph-fs-uuid.txt ] || uuidgen > /tmp/ceph-fs-uuid.txt #NOTE(portdirect): to use RBD devices with Ubuntu kernels < 4.5 this # should be set to 'hammer' @@ -42,8 +43,10 @@ export OSH_PATH # fluentd's configuration. This ensures the placeholder value gets rendered appropriately export tag='${tag}' -manifests="armada-cluster-ingress armada-ceph armada-lma armada-osh" -for manifest in $manifests; do +for manifest in armada-cluster-ingress armada-ceph; do echo "Rendering $manifest manifest" - envsubst < ./tools/deployment/armada/manifests/$manifest.yaml > /tmp/$manifest.yaml + envsubst < ${OSH_INFRA_PATH}/tools/deployment/armada/manifests/$manifest.yaml > /tmp/$manifest.yaml done + +echo "Rendering armada-osh manifest" +envsubst < ./tools/deployment/armada/manifests/armada-osh.yaml > /tmp/armada-osh.yaml diff --git a/tools/deployment/armada/025-armada-validate-manifests.sh b/tools/deployment/armada/025-armada-validate-manifests.sh index e37ea52b17..e05230dd6f 100755 --- a/tools/deployment/armada/025-armada-validate-manifests.sh +++ b/tools/deployment/armada/025-armada-validate-manifests.sh @@ -16,7 +16,7 @@ set -xe -manifests="armada-cluster-ingress armada-ceph armada-lma armada-osh" +manifests="armada-cluster-ingress armada-ceph armada-osh" for manifest in $manifests; do echo "Validating $manifest manifest" armada validate /tmp/$manifest.yaml diff --git a/tools/deployment/armada/030-armada-apply-manifests.sh b/tools/deployment/armada/030-armada-apply-manifests.sh index 02725509d1..b82489663e 100755 --- a/tools/deployment/armada/030-armada-apply-manifests.sh +++ b/tools/deployment/armada/030-armada-apply-manifests.sh @@ -16,7 +16,7 @@ set -xe -manifests="armada-cluster-ingress armada-ceph armada-lma armada-osh" +manifests="armada-cluster-ingress armada-ceph armada-osh" for manifest in $manifests; do echo "Applying $manifest manifest" armada apply /tmp/$manifest.yaml diff --git a/tools/deployment/armada/generate-passwords.sh b/tools/deployment/armada/generate-osh-passwords.sh similarity index 79% rename from tools/deployment/armada/generate-passwords.sh rename to tools/deployment/armada/generate-osh-passwords.sh index b6bc32a2ef..73ee217073 100755 --- a/tools/deployment/armada/generate-passwords.sh +++ b/tools/deployment/armada/generate-osh-passwords.sh @@ -22,14 +22,9 @@ passwords="BARBICAN_DB_PASSWORD \ CINDER_DB_PASSWORD \ CINDER_RABBITMQ_USER_PASSWORD \ CINDER_USER_PASSWORD \ - DB_ADMIN_PASSWORD \ - ELASTICSEARCH_ADMIN_PASSWORD \ GLANCE_DB_PASSWORD \ GLANCE_RABBITMQ_USER_PASSWORD \ GLANCE_USER_PASSWORD \ - GRAFANA_ADMIN_PASSWORD \ - GRAFANA_DB_PASSWORD \ - GRAFANA_SESSION_DB_PASSWORD \ HEAT_DB_PASSWORD \ HEAT_RABBITMQ_USER_PASSWORD \ HEAT_STACK_PASSWORD \ @@ -39,7 +34,6 @@ passwords="BARBICAN_DB_PASSWORD \ KEYSTONE_AUTHTOKEN_MEMCACHED_SECRET_KEY \ KEYSTONE_DB_PASSWORD \ KEYSTONE_RABBITMQ_USER_PASSWORD \ - KIBANA_ADMIN_PASSWORD \ NEUTRON_DB_PASSWORD \ NEUTRON_RABBITMQ_USER_PASSWORD \ NEUTRON_USER_PASSWORD \ @@ -48,17 +42,13 @@ passwords="BARBICAN_DB_PASSWORD \ NOVA_RABBITMQ_USER_PASSWORD \ NOVA_USER_PASSWORD \ OPENSTACK_EXPORTER_USER_PASSWORD \ - PROMETHEUS_ADMIN_PASSWORD \ + OSH_MARIADB_ADMIN_PASSWORD \ RABBITMQ_ADMIN_PASSWORD \ - RADOSGW_S3_ADMIN_ACCESS_KEY \ - RADOSGW_S3_ADMIN_SECRET_KEY \ - RADOSGW_S3_ELASTICSEARCH_ACCESS_KEY \ - RADOSGW_S3_ELASTICSEARCH_SECRET_KEY \ SWIFT_USER_PASSWORD" for password in $passwords do value=$(tr -dc A-Za-z0-9 < /dev/urandom 2>/dev/null | head -c 20) export $password=$value - echo "export $password=$value" >> /tmp/passwords.env + echo "export $password=$value" >> /tmp/osh-passwords.env done diff --git a/tools/deployment/armada/manifests/armada-ceph.yaml b/tools/deployment/armada/manifests/armada-ceph.yaml deleted file mode 100644 index 34747aed49..0000000000 --- a/tools/deployment/armada/manifests/armada-ceph.yaml +++ /dev/null @@ -1,327 +0,0 @@ ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: helm-toolkit -data: - chart_name: helm-toolkit - release: helm-toolkit - namespace: helm-toolkit - values: {} - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: helm-toolkit - reference: master - dependencies: [] ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-ingress-controller -data: - chart_name: ceph-ingress-controller - release: ceph-ingress-controller - namespace: ceph - wait: - timeout: 1800 - labels: - release_group: osh-ceph-ingress-controller - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ceph-ingress-controller - values: - release_uuid: ${RELEASE_UUID} - labels: - server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - error_server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - pod: - replicas: - error_page: 2 - ingress: 2 - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ingress - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-mon -data: - chart_name: ceph-mon - release: ceph-mon - namespace: ceph - wait: - timeout: 1800 - labels: - release_group: osh-ceph-mon - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ceph-mon - values: - release_uuid: ${RELEASE_UUID} - endpoints: - ceph_mon: - namespace: ceph - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - storage_secrets: true - ceph: true - bootstrap: - enabled: true - conf: - ceph: - global: - fsid: ${CEPH_FS_ID} - pool: - crush: - tunables: ${CRUSH_TUNABLES} - target: - # NOTE(portdirect): 5 nodes, with one osd per node - osd: 5 - pg_per_osd: 100 - storage: - osd: - - data: - type: directory - location: /var/lib/openstack-helm/ceph/osd/osd-one - journal: - type: directory - location: /var/lib/openstack-helm/ceph/osd/journal-one - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-mon - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-osd -data: - chart_name: ceph-osd - release: ceph-osd - namespace: ceph - wait: - timeout: 1800 - labels: - release_group: osh-ceph-osd - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ceph-osd - values: - release_uuid: ${RELEASE_UUID} - endpoints: - ceph_mon: - namespace: ceph - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - ceph: true - bootstrap: - enabled: true - conf: - ceph: - global: - fsid: ${CEPH_FS_ID} - rgw_ks: - enabled: true - pool: - crush: - tunables: ${CRUSH_TUNABLES} - target: - # NOTE(portdirect): 5 nodes, with one osd per node - osd: 5 - pg_per_osd: 100 - storage: - osd: - - data: - type: directory - location: /var/lib/openstack-helm/ceph/osd/osd-one - journal: - type: directory - location: /var/lib/openstack-helm/ceph/osd/journal-one - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-osd - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-client -data: - chart_name: ceph-client - release: ceph-client - namespace: ceph - wait: - timeout: 1800 - labels: - release_group: osh-ceph-client - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ceph-client - values: - release_uuid: ${RELEASE_UUID} - endpoints: - ceph_mon: - namespace: ceph - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - ceph: true - bootstrap: - enabled: true - conf: - ceph: - global: - fsid: ${CEPH_FS_ID} - pool: - crush: - tunables: ${CRUSH_TUNABLES} - target: - # NOTE(portdirect): 5 nodes, with one osd per node - osd: 5 - pg_per_osd: 100 - storage: - osd: - - data: - type: directory - location: /var/lib/openstack-helm/ceph/osd/osd-one - journal: - type: directory - location: /var/lib/openstack-helm/ceph/osd/journal-one - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-client - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-provisioners -data: - chart_name: ceph-provisioners - release: ceph-provisioners - namespace: ceph - wait: - timeout: 1800 - labels: - release_group: osh-ceph-provisioners - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ceph-provisioners - values: - release_uuid: ${RELEASE_UUID} - endpoints: - ceph_mon: - namespace: ceph - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - ceph: true - rbd_provisioner: true - cephfs_provisioner: true - client_secrets: false - bootstrap: - enabled: true - conf: - ceph: - global: - fsid: ${CEPH_FS_ID} - pool: - crush: - tunables: ${CRUSH_TUNABLES} - target: - # NOTE(portdirect): 5 nodes, with one osd per node - osd: 5 - pg_per_osd: 100 - storage: - osd: - - data: - type: directory - location: /var/lib/openstack-helm/ceph/osd/osd-one - journal: - type: directory - location: /var/lib/openstack-helm/ceph/osd/journal-one - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-provisioners - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/ChartGroup/v1 -metadata: - schema: metadata/Document/v1 - name: ceph-storage -data: - description: "Ceph Storage" - sequenced: True - chart_group: - - ceph-ingress-controller - - ceph-mon - - ceph-osd - - ceph-client - - ceph-provisioners ---- -schema: armada/Manifest/v1 -metadata: - schema: metadata/Document/v1 - name: armada-manifest -data: - release_prefix: osh - chart_groups: - - ceph-storage diff --git a/tools/deployment/armada/manifests/armada-cluster-ingress.yaml b/tools/deployment/armada/manifests/armada-cluster-ingress.yaml deleted file mode 100644 index 5a3ceb8017..0000000000 --- a/tools/deployment/armada/manifests/armada-cluster-ingress.yaml +++ /dev/null @@ -1,81 +0,0 @@ ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: helm-toolkit -data: - chart_name: helm-toolkit - release: helm-toolkit - namespace: helm-toolkit - values: {} - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: helm-toolkit - reference: master - dependencies: [] ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: ingress-kube-system -data: - chart_name: ingress-kube-system - release: ingress-kube-system - namespace: kube-system - wait: - timeout: 1800 - labels: - release_group: osh-ingress-kube-system - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-ingress-kube-system - values: - release_uuid: ${RELEASE_UUID} - labels: - server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - error_server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - pod: - replicas: - error_page: 2 - deployment: - mode: cluster - type: DaemonSet - network: - host_namespace: true - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ingress - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/ChartGroup/v1 -metadata: - schema: metadata/Document/v1 - name: cluster-ingress-controller -data: - description: "Cluster Ingress Controller" - sequenced: False - chart_group: - - ingress-kube-system ---- -schema: armada/Manifest/v1 -metadata: - schema: metadata/Document/v1 - name: armada-manifest -data: - release_prefix: osh - chart_groups: - - cluster-ingress-controller diff --git a/tools/deployment/armada/manifests/armada-lma.yaml b/tools/deployment/armada/manifests/armada-lma.yaml deleted file mode 100644 index 1ffb6fef65..0000000000 --- a/tools/deployment/armada/manifests/armada-lma.yaml +++ /dev/null @@ -1,1340 +0,0 @@ ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: helm-toolkit -data: - chart_name: helm-toolkit - release: helm-toolkit - namespace: helm-toolkit - values: {} - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: helm-toolkit - reference: master - dependencies: [] ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: osh-infra-ingress-controller -data: - chart_name: osh-infra-ingress-controller - release: osh-infra-ingress-controller - namespace: osh-infra - wait: - timeout: 1800 - labels: - release_group: osh-infra-osh-infra-ingress-controller - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-osh-infra-ingress-controller - values: - release_uuid: ${RELEASE_UUID} - labels: - node_selector_key: openstack-control-plane - node_selector_value: enabled - pod: - replicas: - error_page: 2 - ingress: 2 - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ingress - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: osh-infra-ceph-config -data: - chart_name: osh-infra-ceph-config - release: osh-infra-ceph-config - namespace: osh-infra - wait: - timeout: 1800 - labels: - release_group: osh-infra-osh-infra-ceph-config - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-osh-infra-ceph-config - values: - release_uuid: ${RELEASE_UUID} - endpoints: - ceph_mon: - namespace: ceph - labels: - jobs: - node_selector_key: openstack-control-plane - node_selector_value: enabled - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - ceph: False - rbd_provisioner: False - cephfs_provisioner: False - client_secrets: True - bootstrap: - enabled: False - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-provisioners - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: osh-infra-radosgw -data: - chart_name: osh-infra-radosgw - release: osh-infra-radosgw - namespace: osh-infra - wait: - timeout: 1800 - labels: - release_group: osh-infra-osh-infra-radosgw - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-radosgw-osh-infra - values: - release_uuid: ${RELEASE_UUID} - endpoints: - object_store: - namespace: osh-infra - ceph_object_store: - namespace: osh-infra - auth: - admin: - access_key: ${RADOSGW_S3_ADMIN_ACCESS_KEY} - secret_key: ${RADOSGW_S3_ADMIN_SECRET_KEY} - ceph_mon: - namespace: ceph - labels: - job: - node_selector_key: openstack-control-plane - node_selector_value: enabled - bootstrap: - enabled: False - conf: - rgw_ks: - enabled: False - rgw_s3: - enabled: True - network: - public: ${CEPH_NETWORK} - cluster: ${CEPH_NETWORK} - deployment: - ceph: True - rbd_provisioner: False - cephfs_provisioner: False - client_secrets: False - rgw_keystone_user_and_endpoints: False - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ceph-rgw - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: osh-infra-ldap -data: - chart_name: osh-infra-ldap - release: osh-infra-ldap - namespace: osh-infra - wait: - timeout: 1800 - labels: - release_group: osh-infra-osh-infra-ldap - install: - no_hooks: false - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-osh-infra-ldap - values: - release_uuid: ${RELEASE_UUID} - labels: - server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - job: - node_selector_key: openstack-control-plane - node_selector_value: enabled - bootstrap: - enabled: true - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: ldap - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: osh-infra-mariadb -data: - chart_name: osh-infra-mariadb - release: osh-infra-mariadb - namespace: osh-infra - wait: - timeout: 1800 - labels: - release_group: osh-infra-osh-infra-mariadb - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-osh-infra-mariadb - values: - release_uuid: ${RELEASE_UUID} - pod: - replicas: - server: 1 - endpoints: - oslo_db: - auth: - admin: - password: ${DB_ADMIN_PASSWORD} - volume: - enabled: false - labels: - server: - node_selector_key: openstack-control-plane - node_selector_value: enabled - prometheus_mysql_exporter: - node_selector_key: openstack-control-plane - node_selector_value: enabled - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: mariadb - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: elasticsearch -data: - chart_name: elasticsearch - release: elasticsearch - namespace: osh-infra - wait: - timeout: 3600 - labels: - release_group: osh-infra-elasticsearch - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-elasticsearch - - type: pod - labels: - release_group: osh-infra-elasticsearch - component: test - values: - release_uuid: ${RELEASE_UUID} - endpoints: - elasticsearch: - auth: - admin: - password: ${ELASTICSEARCH_ADMIN_PASSWORD} - object_store: - namespace: osh-infra - ceph_object_store: - namespace: osh-infra - auth: - admin: - access_key: ${RADOSGW_S3_ADMIN_ACCESS_KEY} - secret_key: ${RADOSGW_S3_ADMIN_SECRET_KEY} - elasticsearch: - access_key: ${RADOSGW_S3_ELASTICSEARCH_ACCESS_KEY} - secret_key: ${RADOSGW_S3_ELASTICSEARCH_SECRET_KEY} - pod: - replicas: - data: 1 - master: 2 - labels: - elasticsearch: - node_selector_key: openstack-control-plane - node_selector_value: enabled - job: - node_selector_key: openstack-control-plane - node_selector_value: enabled - conf: - elasticsearch: - env: - java_opts: "-Xms512m -Xmx512m" - snapshots: - enabled: true - source: - type: local - location: ${OSH_INFRA_PATH} - subpath: elasticsearch - reference: master - dependencies: - - helm-toolkit ---- -schema: armada/Chart/v1 -metadata: - schema: metadata/Document/v1 - name: fluent-logging -data: - chart_name: fluent-logging - release: fluent-logging - namespace: osh-infra - wait: - timeout: 3600 - labels: - release_group: osh-infra-fluent-logging - test: - enabled: false - install: - no_hooks: False - upgrade: - no_hooks: False - pre: - delete: - - type: job - labels: - release_group: osh-infra-fluent-logging - - type: pod - labels: - release_group: osh-infra-fluent-logging - component: test - values: - release_uuid: ${RELEASE_UUID} - conf: - fluentbit: - - service: - header: service - Flush: 30 - Daemon: Off - Log_Level: info - Parsers_File: parsers.conf - - ceph_cluster_logs: - header: input - Name: tail - Tag: ceph.cluster.* - Path: /var/log/ceph/ceph.log - DB: /var/log/ceph.db - Parsers: syslog - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - ceph_audit_logs: - header: input - Name: tail - Tag: ceph.audit.* - Path: /var/log/ceph/ceph.audit.log - DB: /var/log/ceph.db - Parsers: syslog - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - ceph_mon_logs: - header: input - Name: tail - Tag: ceph.mon.* - Path: /var/log/ceph/ceph-mon**.log - DB: /var/log/ceph.db - Parsers: syslog - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - ceph_osd_logs: - header: input - Name: tail - Tag: ceph.osd.* - Path: /var/log/ceph/ceph-osd**.log - DB: /var/log/ceph.db - Parsers: syslog - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - kernel_messages: - header: input - Name: tail - Tag: kernel - Path: /var/log/kern.log - DB: /var/log/kern.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - kubelet: - header: input - Name: systemd - Tag: journal.* - Path: ${JOURNAL_PATH} - Systemd_Filter: _SYSTEMD_UNIT=kubelet.service - DB: /var/log/kubelet.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - libvirt: - header: input - Name: tail - Tag: libvirt - Path: /var/log/libvirt/libvirtd.log - DB: /var/log/libvirt.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - qemu: - header: input - Name: tail - Tag: qemu - Path: /var/log/libvirt/qemu/*.log - DB: /var/log/qemu.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - nagios: - header: input - Name: tail - Tag: nagios - Path: /opt/nagios/var/nagios.log - DB: /var/log/nagios.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - docker_daemon: - header: input - Name: systemd - Tag: journal.* - Path: ${JOURNAL_PATH} - Systemd_Filter: _SYSTEMD_UNIT=docker.service - DB: /var/log/docker.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - libvirt_record_modifier: - header: filter - Name: record_modifier - Match: libvirt - Record: hostname ${HOSTNAME} - - qemu_record_modifier: - header: filter - Name: record_modifier - Match: qemu - Record: hostname ${HOSTNAME} - - kernel_record_modifier: - header: filter - Name: record_modifier - Match: kernel - Record: hostname ${HOSTNAME} - - systemd_modify_fields: - header: filter - Name: modify - Match: journal.** - Rename: - _BOOT_ID: BOOT_ID - _CAP_EFFECTIVE: CAP_EFFECTIVE - _CMDLINE: CMDLINE - _COMM: COMM - _EXE: EXE - _GID: GID - _HOSTNAME: HOSTNAME - _MACHINE_ID: MACHINE_ID - _PID: PID - _SYSTEMD_CGROUP: SYSTEMD_CGROUP - _SYSTEMD_SLICE: SYSTEMD_SLICE - _SYSTEMD_UNIT: SYSTEMD_UNIT - _UID: UID - _TRANSPORT: TRANSPORT - - containers_tail: - header: input - Name: tail - Tag: kube.* - Path: /var/log/containers/*.log - Parser: docker - DB: /var/log/flb_kube.db - Mem_Buf_Limit: 5MB - DB.Sync: Normal - Buffer_Chunk_Size: 1M - Buffer_Max_Size: 1M - - drop_fluentd_logs: - header: output - Name: "null" - Match: "**.fluentd**" - - kube_filter: - header: filter - Name: kubernetes - Match: kube.* - Merge_JSON_Log: On - - fluentd_output: - header: output - Name: forward - Match: "*" - Host: ${FLUENTD_HOST} - Port: ${FLUENTD_PORT} - parsers: - - docker: - header: parser - Name: docker - Format: json - Time_Key: time - Time_Format: "%Y-%m-%dT%H:%M:%S.%L" - Time_Keep: On - - syslog: - header: parser - Name: syslog - Format: regex - Regex: '^(?