treasuremap/global/software/charts/osh/openstack-tenant-ceph/ceph-client.yaml
Chinasubbareddy M f96c3016a0 Airship-treasuremap: Ceph charts cleanup
This is  to clean ceph charts for unused variables and left over from
ceph chart split

Change-Id: I3155c9079ab4b0b4df30782335833389c452adfd
2019-01-14 19:29:13 +00:00

214 lines
5.7 KiB
YAML

---
schema: armada/Chart/v1
metadata:
schema: metadata/Document/v1
name: tenant-ceph-client-global
layeringDefinition:
abstract: true
layer: global
storagePolicy: cleartext
labels:
name: tenant-ceph-client-global
substitutions:
# Chart source
- src:
schema: pegleg/SoftwareVersions/v1
name: software-versions
path: .charts.ucp.tenant-ceph-client
dest:
path: .source
# Images
- src:
schema: pegleg/SoftwareVersions/v1
name: software-versions
path: .images.ceph.tenant-ceph-client
dest:
path: .values.images.tags
# IP addresses
- src:
schema: pegleg/CommonAddresses/v1
name: common-addresses
path: .storage.ceph.public_cidr
dest:
path: .values.network.public
- src:
schema: pegleg/CommonAddresses/v1
name: common-addresses
path: .storage.ceph.cluster_cidr
dest:
path: .values.network.cluster
- src:
schema: pegleg/EndpointCatalogue/v1
name: ucp_endpoints
path: .ceph.tenant_ceph_mon
dest:
path: .values.endpoints.ceph_mon
- src:
schema: pegleg/EndpointCatalogue/v1
name: ucp_endpoints
path: .ceph.tenant_ceph_mgr
dest:
path: .values.endpoints.ceph_mgr
# Secrets
- dest:
path: .values.conf.ceph.global.fsid
src:
schema: deckhand/Passphrase/v1
name: tenant_ceph_fsid
path: .
data:
chart_name: tenant-ceph-client
release: tenant-ceph-client
namespace: tenant-ceph
protected:
continue_processing: true
wait:
timeout: 900
labels:
release_group: airship-tenant-ceph-client
install:
no_hooks: false
upgrade:
no_hooks: false
pre:
delete:
- type: job
labels:
release_group: airship-tenant-ceph-client
values:
labels:
job:
node_selector_key: tenant-ceph-control-plane
node_selector_value: enabled
mds:
node_selector_key: tenant-ceph-mds
node_selector_value: enabled
mgr:
node_selector_key: tenant-ceph-mgr
node_selector_value: enabled
endpoints:
ceph_mon:
namespace: tenant-ceph
ceph_mgr:
namespace: tenant-ceph
monitoring:
prometheus:
ceph_mgr:
port: 9284
ceph_mgr_modules_config:
prometheus:
server_port: 9284
deployment:
ceph: true
bootstrap:
enabled: true
manifests:
deployment_mds: false
conf:
features:
mds: false
pool:
spec:
# RBD pool
- name: rbd
application: rbd
replication: 3
percent_total_data: 10
# Cinder volumes pool
- name: cinder.volumes
application: cinder-volume
replication: 3
percent_total_data: 40
# RadosGW pools
- name: .rgw.root
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.control
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.data.root
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.gc
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.log
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.intent-log
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.meta
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.usage
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.users.keys
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.users.email
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.users.swift
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.users.uid
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.buckets.extra
application: rgw
replication: 3
percent_total_data: 0.1
- name: default.rgw.buckets.index
application: rgw
replication: 3
percent_total_data: 3
- name: default.rgw.buckets.data
application: rgw
replication: 3
percent_total_data: 30
# NOTE(alanmeadowS) spport 4.x 16.04 kernels (non-HWE)
crush:
tunables: 'hammer'
# NOTE(alanmeadows): This is required ATM for bootstrapping a Ceph
# cluster with only one OSD. Depending on OSD targeting & site
# configuration this can be changed.
target:
osd: 1
pg_per_osd: 100
protected: true
default:
# NOTE(alanmeadows): This is required ATM for bootstrapping a Ceph
# cluster with only one OSD. Depending on OSD targeting & site
# configuration this can be changed.
crush_rule: replicated_rule
ceph:
global:
# NOTE(mb874d): This is required ATM for bootstrapping a Ceph
# cluster with only one OSD. Depending on OSD targeting & site
# configuration this can be changed.
osd_pool_default_size: 1
dependencies:
- tenant-ceph-htk
...