6b2be98f0d
- remove the need to create the openstack namespace: the namespace will be created by default by helm when the first chart is installed using the namespace - the rbd-provisioner chart is now installed in the openstack namespace and doesn't create resources over multiple namespaces Change-Id: I875acd5959ceca033d9233ddf70ac951db7ca6a8 Story: 2004520 Task: 28395 Signed-off-by: Irina Mihai <irina.mihai@windriver.com>
162 lines
4.0 KiB
YAML
162 lines
4.0 KiB
YAML
#
|
|
# Copyright (c) 2018 Wind River Systems, Inc.
|
|
#
|
|
# SPDX-License-Identifier: Apache-2.0
|
|
#
|
|
|
|
#
|
|
# Global options.
|
|
# Defaults should be fine in most cases.
|
|
global:
|
|
#
|
|
# Defines the name of a Provisioner.
|
|
#
|
|
name: "rbd-provisioner"
|
|
#
|
|
# Execute initialization job to verify external Ceph cluster access
|
|
# and setup additional dependencies assumed by dependent helm charts
|
|
# (i.e. configmap and secrets).
|
|
# Skipping is not recommended.
|
|
#
|
|
job_storage_init: true
|
|
#
|
|
# Defines whether to reuse an already defined RBAC policy.
|
|
# Make sure that the serviceAccount defined in the RBAC section matches the one
|
|
# in the policy you reuse.
|
|
#
|
|
reuseRbac: false
|
|
#
|
|
# Defines whether to generate service account and role bindings.
|
|
#
|
|
rbac: true
|
|
#
|
|
# Provision storage class. If false you have to provision storage classes by hand.
|
|
#
|
|
provisionStorageClass: true
|
|
#
|
|
# Choose if rbd-provisioner pod should be deployed as deplyment or DaemonSet
|
|
# Values: none, Deployment, DaemonSet
|
|
#
|
|
deployAs: Deployment
|
|
#
|
|
# If configured, tolerations will add a toleration field to the Pod.
|
|
#
|
|
# Node tolerations for rbd-volume-provisioner scheduling to nodes with taints.
|
|
# Ref: https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/
|
|
# Example:
|
|
# [
|
|
# {
|
|
# "key": "node-role.kubernetes.io/master",
|
|
# "operator": "Exists"
|
|
# }
|
|
# ]
|
|
#
|
|
tolerations: []
|
|
# If configured, resources will set the requests/limits field to the Pod.
|
|
# Ref: https://kubernetes.io/docs/concepts/configuration/manage-compute-resources-container/
|
|
# Example:
|
|
# {
|
|
# "limits": {
|
|
# "memory": "200Mi"
|
|
# },
|
|
# "requests": {
|
|
# "cpu": "100m",
|
|
# "memory": "200Mi"
|
|
# }
|
|
# }
|
|
resources: {}
|
|
#
|
|
# Number of replicas to start when configured as deployment
|
|
#
|
|
replicas: 1
|
|
#
|
|
# Node Selector
|
|
#
|
|
nodeSelector: { node-role.kubernetes.io/master: "" }
|
|
#
|
|
# RBAC options.
|
|
# Defaults should be fine in most cases.
|
|
rbac:
|
|
#
|
|
# Cluster Role name
|
|
#
|
|
clusterRole: rbd-provisioner
|
|
#
|
|
# Cluster Role Binding name
|
|
#
|
|
clusterRoleBinding: rbd-provisioner
|
|
#
|
|
# Role name
|
|
#
|
|
role: rbd-provisioner
|
|
#
|
|
# Role Binding name
|
|
#
|
|
roleBinding: rbd-provisioner
|
|
#
|
|
# Defines a name of the service account which Provisioner will use to communicate with API server.
|
|
#
|
|
serviceAccount: rbd-provisioner
|
|
|
|
#
|
|
# Configure storage classes.
|
|
# Defaults for storage classes. Update this if you have a single Ceph storage cluster.
|
|
# No need to add them to each class.
|
|
#
|
|
classdefaults:
|
|
# Define ip addresses of Ceph Monitors
|
|
monitors:
|
|
- 192.168.204.3:6789
|
|
- 192.168.204.150:6789
|
|
- 192.168.204.4:6789
|
|
# Ceph admin account
|
|
adminId: admin
|
|
# K8 secret name for the admin context
|
|
adminSecretName: ceph-secret
|
|
# Ceph RBD image format version
|
|
imageFormat: 2
|
|
# Ceph RBD image features.
|
|
imageFeatures: layering
|
|
|
|
#
|
|
# Configure storage classes.
|
|
# This section should be tailored to your setup. It allows you to define multiple storage
|
|
# classes for the same cluster (e.g. if you have tiers of drives with different speeds).
|
|
# If you have multiple Ceph clusters take attributes from classdefaults and add them here.
|
|
classes:
|
|
- name: fast-rbd # Name of storage class.
|
|
# Ceph pool name
|
|
pool_name: kube
|
|
# Ceph user name to access this pool
|
|
userId: kube
|
|
# K8 secret name with key for accessing the Ceph pool
|
|
userSecretName: ceph-secret-kube
|
|
# Pool replication
|
|
replication: 1
|
|
# Pool crush rule name
|
|
crush_rule_name: storage_tier_ruleset
|
|
# Pool chunk size / PG_NUM
|
|
chunk_size: 8
|
|
# Configuration data for the ephemeral pool(s)
|
|
ephemeral_pools:
|
|
- chunk_size: 8
|
|
crush_rule_name: storage_tier_ruleset
|
|
pool_name: ephemeral
|
|
replication: 1
|
|
|
|
#
|
|
# Defines:
|
|
# - Provisioner's image name including container registry.
|
|
# - CEPH helper image
|
|
#
|
|
images:
|
|
tags:
|
|
rbd_provisioner: quay.io/external_storage/rbd-provisioner:latest
|
|
rbd_provisioner_storage_init: docker.io/port/ceph-config-helper:v1.10.3
|
|
pull_policy: "IfNotPresent"
|
|
local_registry:
|
|
active: false
|
|
exclude:
|
|
- dep_check
|
|
- image_repo_sync
|