Convert guides-welcome to RST
This commit compiles guides-welcome pages into single philosophy.rst page. I've decided to name the section "Philosophy", as that's the only distinctive content that was in original guides-welcome/*.md files. Partial-Implements: blueprint docs-to-rst Change-Id: I7ab5d6465008fc4be2a8c28dd8ca8abbfde0df25
This commit is contained in:
parent
e795466247
commit
f1e49c75b8
@ -1,6 +0,0 @@
|
||||
## Table of Contents: Welcome Guide
|
||||
|
||||
- [Mission Statement](../readme.md#mission-statement)
|
||||
- [Overview](welcome-overview.md)
|
||||
- [Resiliency](welcome-resiliency.md)
|
||||
- [Scaling](welcome-scaling.md)
|
@ -1,3 +0,0 @@
|
||||
## Welcome: Project Overview
|
||||
|
||||
The goal for OpenStack-Helm is to provide an incredibly customizable *framework* for operators and developers alike. This framework will enable end-users to deploy, maintain, and upgrade a fully functioning OpenStack environment for both simple and complex environments. Administrators or developers can either deploy all or individual OpenStack components along with their required dependencies. It heavily borrows concepts from [Stackanetes](https://github.com/stackanetes/stackanetes) and [other complex Helm application deployments](https://github.com/sapcc/openstack-helm). This project is meant to be a collaborative project that brings Openstack applications into a [Cloud-Native](https://www.cncf.io/about/charter) model.
|
@ -1,9 +0,0 @@
|
||||
## Welcome: Resiliency Philosophy
|
||||
|
||||
One of the goals of this project is to produce a set of charts that can be used in a production setting to deploy and upgrade OpenStack. To achieve this goal, all components must be resilient, including both OpenStack and Infrastructure components leveraged by this project. In addition, this also includes Kubernetes itself. It is part of our mission to ensure that all infrastructure components are highly available and that a deployment can withstand a physical host failure out of the box. This means that:
|
||||
|
||||
- OpenStack components need to support and deploy with multiple replicas out of the box to ensure that each chart is deployed as a single-unit production ready first class citizen (unless development mode is enabled).
|
||||
- Infrastructure elements such as Ceph, RabbitMQ, Galera (MariaDB), Memcached, and all others need to support resiliency and leverage multiple replicas for resiliency where applicable. These components also need to validate that their application level configurations (for instance the underlying Galera cluster) can tolerate host crashes and withstand physical host failures.
|
||||
- Scheduling annotations need to be employed to ensure maximum resiliency for multi-host environments. They also need to be flexible to allow all-in-one deployments. To this end, we promote the usage of `podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution` for most infrastructure elements.
|
||||
- We make the assumption that we can depend on a reliable implementation of centralized storage to create PVCs within Kubernetes to support resiliency and complex application design. Today, this is provided by the included Ceph chart. There is much work to do when making even a single backend production ready. We have chosen to focus on bringing Ceph into a production ready state, which includes handling real world deployment scenarios, resiliency, and pool configurations. In the future we would like to support more options for hardened backend PVC's. In the future, we would like to offer flexibility in choosing a hardened backend.
|
||||
- We will document the best practices for running a resilient Kubernetes cluster in production. This includes documenting the steps necessary to make all components resilient, such as Etcd and SkyDNS where possible, and point out gaps due to missing features.
|
@ -1,10 +0,0 @@
|
||||
## Welcome: Scaling Philosophy
|
||||
|
||||
Scaling is another first class citizen in openstack-helm. We will be working to ensure that we support various deployment models that can support hyperscale, such as:
|
||||
|
||||
- Ensuring that by default, clusters include multiple replicas to verify that scaling issues are identified early and often (unless development mode is enabled).
|
||||
- Ensuring that every chart can support more then one replica and allowing operators to override those replica counts. For some applications, this means that they support clustering.
|
||||
- Ensuring clustering style applications are not limited to fixed replica counts. For instance, we want to ensure that we can support n=Galera members and have those scale linearly, within reason, as opposed to only supporting a fixed count.
|
||||
- Duplicate charts of the same type within the same namespace. For example, deploying rabbitmq twice, to the openstack namespace resulting in two fully functioning clusters.
|
||||
- Allowing charts to be deployed to a diverse set of namespaces. For example, allowing infrastructure to be deployed in one namespace and OpenStack in another, or deploying each chart in its own namespace.
|
||||
- Supporting hyperscale configurations that call for per-component infrastructure, such as a dedicated database and RabbitMQ solely for Ceilometer, or even dedicated infrastructure(s) for every component you deploy. It is unique, large scale deployment designs such as this that only become practical under a Kubernetes/Container framework and we want to ensure that we can support them.
|
@ -12,6 +12,7 @@ Contents:
|
||||
:maxdepth: 3
|
||||
|
||||
readme
|
||||
philosophy
|
||||
contributing
|
||||
|
||||
Indices and tables
|
||||
|
71
doc/source/philosophy.rst
Normal file
71
doc/source/philosophy.rst
Normal file
@ -0,0 +1,71 @@
|
||||
==========
|
||||
Philosophy
|
||||
==========
|
||||
|
||||
Resiliency Philosophy
|
||||
---------------------
|
||||
|
||||
One of the goals of this project is to produce a set of charts that can be used
|
||||
in a production setting to deploy and upgrade OpenStack. To achieve this goal,
|
||||
all components must be resilient, including both OpenStack and Infrastructure
|
||||
components leveraged by this project. In addition, this also includes
|
||||
Kubernetes itself. It is part of our mission to ensure that all infrastructure
|
||||
components are highly available and that a deployment can withstand a physical
|
||||
host failure out of the box. This means that:
|
||||
|
||||
* OpenStack components need to support and deploy with multiple replicas out of
|
||||
the box to ensure that each chart is deployed as a single-unit production
|
||||
ready first class citizen (unless development mode is enabled).
|
||||
* Infrastructure elements such as Ceph, RabbitMQ, Galera (MariaDB), Memcached,
|
||||
and all others need to support resiliency and leverage multiple replicas for
|
||||
resiliency where applicable. These components also need to validate that
|
||||
their application level configurations (for instance the underlying Galera
|
||||
cluster) can tolerate host crashes and withstand physical host failures.
|
||||
* Scheduling annotations need to be employed to ensure maximum resiliency for
|
||||
multi-host environments. They also need to be flexible to allow all-in-one
|
||||
deployments. To this end, we promote the usage of
|
||||
``podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution`` for most
|
||||
infrastructure elements.
|
||||
* We make the assumption that we can depend on a reliable implementation of
|
||||
centralized storage to create PVCs within Kubernetes to support resiliency
|
||||
and complex application design. Today, this is provided by the included Ceph
|
||||
chart. There is much work to do when making even a single backend production
|
||||
ready. We have chosen to focus on bringing Ceph into a production ready
|
||||
state, which includes handling real world deployment scenarios, resiliency,
|
||||
and pool configurations. In the future we would like to support more options
|
||||
for hardened backend PVC's. In the future, we would like to offer flexibility
|
||||
in choosing a hardened backend.
|
||||
* We will document the best practices for running a resilient Kubernetes
|
||||
cluster in production. This includes documenting the steps necessary to make
|
||||
all components resilient, such as Etcd and SkyDNS where possible, and point
|
||||
out gaps due to missing features.
|
||||
|
||||
Scaling Philosophy
|
||||
------------------
|
||||
|
||||
Scaling is another first class citizen in openstack-helm. We will be working
|
||||
to ensure that we support various deployment models that can support
|
||||
hyperscale, such as:
|
||||
|
||||
* Ensuring that by default, clusters include multiple replicas to verify that
|
||||
scaling issues are identified early and often (unless development mode is
|
||||
enabled).
|
||||
* Ensuring that every chart can support more then one replica and allowing
|
||||
operators to override those replica counts. For some applications, this means
|
||||
that they support clustering.
|
||||
* Ensuring clustering style applications are not limited to fixed replica
|
||||
counts. For instance, we want to ensure that we can support n=Galera members
|
||||
and have those scale linearly, within reason, as opposed to only supporting a
|
||||
fixed count.
|
||||
* Duplicate charts of the same type within the same namespace. For example,
|
||||
deploying rabbitmq twice, to the openstack namespace resulting in two fully
|
||||
functioning clusters.
|
||||
* Allowing charts to be deployed to a diverse set of namespaces. For example,
|
||||
allowing infrastructure to be deployed in one namespace and OpenStack in
|
||||
another, or deploying each chart in its own namespace.
|
||||
* Supporting hyperscale configurations that call for per-component
|
||||
infrastructure, such as a dedicated database and RabbitMQ solely for
|
||||
Ceilometer, or even dedicated infrastructure(s) for every component you
|
||||
deploy. It is unique, large scale deployment designs such as this that only
|
||||
become practical under a Kubernetes/Container framework and we want to ensure
|
||||
that we can support them.
|
Loading…
Reference in New Issue
Block a user