Merge "Update Airshipctl and CAPD Documentation"
This commit is contained in:
commit
36d5f9822a
@ -1,719 +1,229 @@
|
||||
# Airshipctl and Cluster API Docker Integration
|
||||
|
||||
## Overview
|
||||
Airshipctl and cluster api docker integration facilitates usage of `airshipctl`
|
||||
to create cluster api management and workload clusters using `docker as
|
||||
infrastructure provider`.
|
||||
|
||||
This document demonstrates usage of airshipctl to create kubernetes clusters
|
||||
locally using docker and kind.
|
||||
|
||||
Airshipctl requires an existing Kubernetes cluster accessible via kubectl.
|
||||
We will use kind as a local bootstrap cluster, to provision a target management
|
||||
cluster on the docker infrastructure provider. The target management cluster
|
||||
will then be used to create a workload cluster with one or more worker nodes.
|
||||
|
||||
## Workflow
|
||||
A simple workflow that can be tested involves the following operations:
|
||||
|
||||
**Initialize the management cluster with cluster api and docker provider
|
||||
components**
|
||||
Below are the steps that we will go through to create the kubernetes clusters
|
||||
locally using docker and kind.
|
||||
|
||||
> airshipctl cluster init --debug
|
||||
- create a single node kubernetes cluster using kind
|
||||
- initialize the kind cluster with cluster api management components
|
||||
- use the kind management cluster to create a target cluster with one control
|
||||
plane
|
||||
- apply cni solution on the target cluster
|
||||
- initialize the target cluster with cluster api management components
|
||||
- move the cluster api management crds from kind cluster to target management
|
||||
cluster
|
||||
- tear down the kind management cluster
|
||||
- use the target management cluster to create worker nodes
|
||||
- increase/decrease the worker count as required
|
||||
|
||||
**Create a workload cluster, with control plane and worker nodes**
|
||||
## Airshipctl Commands Used And Purpose
|
||||
|
||||
> airshipctl phase run controlplane
|
||||
`Pull documents from the remote git repository`
|
||||
> airshipctl document pull
|
||||
|
||||
> airshipctl phase run workers
|
||||
`Initialize the kind cluster with cluster api and docker provider components`
|
||||
> airshipctl phase run clusterctl-init-ephemeral
|
||||
|
||||
Note: `airshipctl phase run initinfra-ephemeral` is not used because all the provider
|
||||
components are initialized using `airshipctl cluster init`
|
||||
`Use the management cluster to create a target cluster with one control plane`
|
||||
> airshipctl phase run controlplane-ephemeral
|
||||
|
||||
The phase `initinfra` is included just to get `validate docs` to
|
||||
pass.
|
||||
`Get multiple contexts for every cluster in the airship site`
|
||||
> airshipctl cluster get-kubeconfig > ~/.airship/kubeconfig-tmp
|
||||
|
||||
## Common Pre-requisites
|
||||
`Initialize CNI on target cluster`
|
||||
> airshipctl phase run initinfra-networking-target
|
||||
|
||||
`Initialize Target Cluster with cluster api and docker proivder components`
|
||||
> airshipctl phase run clusterctl-init-target
|
||||
|
||||
`Move managment CRDs from kind management cluster to target management cluster`
|
||||
> airshipctl phase run clusterctl-move
|
||||
|
||||
`Use target management cluster to deploy workers`
|
||||
> airshipctl phase run workers-target
|
||||
|
||||
## Getting Started
|
||||
|
||||
### Build [Airshipctl](https://docs.airshipit.org/airshipctl/developers.html)
|
||||
|
||||
```
|
||||
$ git clone https://review.opendev.org/airship/airshipctl
|
||||
|
||||
$ cd airshipctl
|
||||
|
||||
$ ./tools/deployment/21_systemwide_executable.sh
|
||||
```
|
||||
|
||||
### Create airship configuration
|
||||
|
||||
```
|
||||
$ cat ~/.airship/config
|
||||
|
||||
apiVersion: airshipit.org/v1alpha1
|
||||
managementConfiguration:
|
||||
dummy_management_config:
|
||||
type: redfish
|
||||
insecure: true
|
||||
useproxy: false
|
||||
systemActionRetries: 30
|
||||
systemRebootDelay: 30
|
||||
contexts:
|
||||
ephemeral-cluster:
|
||||
contextKubeconf: ephemeral-cluster_ephemeral
|
||||
manifest: dummy_manifest
|
||||
managementConfiguration: dummy_management_config
|
||||
target-cluster:
|
||||
contextKubeconf: target-cluster_target
|
||||
manifest: dummy_manifest
|
||||
managementConfiguration: dummy_management_config
|
||||
currentContext: ephemeral-cluster
|
||||
kind: Config
|
||||
manifests:
|
||||
dummy_manifest:
|
||||
phaseRepositoryName: primary
|
||||
repositories:
|
||||
primary:
|
||||
checkout:
|
||||
branch: master
|
||||
force: false
|
||||
remoteRef: ""
|
||||
tag: ""
|
||||
url: https://review.opendev.org/airship/airshipctl
|
||||
metadataPath: manifests/site/docker-test-site/metadata.yaml
|
||||
targetPath: /tmp/airship
|
||||
```
|
||||
|
||||
### Deploy Control plane and Workers
|
||||
|
||||
```
|
||||
$ export KIND_EXPERIMENTAL_DOCKER_NETWORK=bridge
|
||||
|
||||
$ export KUBECONFIG="${HOME}/.airship/kubeconfig"
|
||||
|
||||
$ kind create cluster --name ephemeral-cluster --wait 120s --kubeconfig \
|
||||
"${HOME}/.airship/kubeconfig" \
|
||||
--config ./tools/deployment/templates/kind-cluster-with-extramounts
|
||||
|
||||
$ kubectl config set-context ephemeral-cluster \
|
||||
--cluster kind-ephemeral-cluster \
|
||||
--user kind-ephemeral-cluster --kubeconfig $KUBECONFIG
|
||||
|
||||
$ airshipctl document pull -n --debug
|
||||
|
||||
$ airshipctl phase run clusterctl-init-ephemeral --debug --wait-timeout 300s
|
||||
|
||||
$ airshipctl phase run controlplane-ephemeral --debug --wait-timeout 300s
|
||||
|
||||
$ airshipctl cluster get-kubeconfig > ~/.airship/kubeconfig-tmp
|
||||
|
||||
$ mv ~/.airship/kubeconfig-tmp "${KUBECONFIG}"
|
||||
|
||||
$ airshipctl phase run initinfra-networking-target --debug
|
||||
|
||||
$ kubectl --context target-cluster wait \
|
||||
--for=condition=Ready nodes --all --timeout 300s
|
||||
|
||||
$ kubectl get nodes --context target-cluster -A
|
||||
|
||||
```
|
||||
|
||||
`Note:
|
||||
Please take note of the control plane node name because it is untainted
|
||||
in the next step.
|
||||
For eg. control plane node name could be something like
|
||||
target-cluster-control-plane-twwsv
|
||||
`
|
||||
|
||||
```
|
||||
|
||||
$ kubectl taint node target-cluster-control-plane-twwsv \
|
||||
node-role.kubernetes.io/master- --context target-cluster --request-timeout 10s
|
||||
|
||||
$ airshipctl phase run clusterctl-init-target --debug --wait-timeout 300s
|
||||
|
||||
$ kubectl get pods -A --context target-cluster
|
||||
|
||||
$ airshipctl phase run clusterctl-move --debug --progress
|
||||
|
||||
$ kubectl get machines --context target-cluster
|
||||
|
||||
$ kind delete cluster --name "ephemeral-cluster"
|
||||
|
||||
$ airshipctl phase run workers-target --debug
|
||||
|
||||
$ kubectl get machines --context target-cluster
|
||||
|
||||
NAME PROVIDERID PHASE
|
||||
target-cluster-control-plane-m5jf7 docker:////target-cluster-control-plane-m5jf7 Running
|
||||
target-cluster-md-0-84db44cdff-r8dkr docker:////target-cluster-md-0-84db44cdff-r8dkr Running
|
||||
|
||||
```
|
||||
|
||||
## Scale Workers
|
||||
|
||||
Worker count can be adjusted in airshipctl/manifests/site/docker-test-site/
|
||||
target/workers/machine_count.json.
|
||||
|
||||
In this example, we have changed it to 3.
|
||||
|
||||
```
|
||||
|
||||
$ cat /tmp/airship/airshipctl/manifests/site/docker-test-site/target/workers/machine_count.json
|
||||
|
||||
[
|
||||
{ "op": "replace","path": "/spec/replicas","value": 3 }
|
||||
]
|
||||
|
||||
$ airshipctl phase run workers-target --debug
|
||||
|
||||
$ kubectl get machines --kubeconfig /tmp/target-cluster.kubeconfig
|
||||
|
||||
NAME PROVIDERID PHASE
|
||||
target-cluster-control-plane-m5jf7 docker:////target-cluster-control-plane-m5jf7 Running
|
||||
target-cluster-md-0-84db44cdff-b6zp6 docker:////target-cluster-md-0-84db44cdff-b6zp6 Running
|
||||
target-cluster-md-0-84db44cdff-g4nm7 docker:////target-cluster-md-0-84db44cdff-g4nm7 Running
|
||||
target-cluster-md-0-84db44cdff-r8dkr docker:////target-cluster-md-0-84db44cdff-r8dkr Running
|
||||
```
|
||||
|
||||
## Clean Up
|
||||
|
||||
```
|
||||
$ kind get clusters
|
||||
target-cluster
|
||||
|
||||
$ kind delete cluster --name target-cluster
|
||||
```
|
||||
|
||||
## More Information
|
||||
|
||||
- worker count can be adjusted from airshipctl/manifests/site/docker-test-site/
|
||||
target/workers/machine_count.json
|
||||
|
||||
- control plane count can be adjusted from airshipctl/manifests/site/
|
||||
docker-test-site/ephemeral/controlplane/machine_count.json
|
||||
|
||||
## Reference
|
||||
|
||||
### Pre-requisites
|
||||
|
||||
* Install [Docker](https://www.docker.com/)
|
||||
* Install [Kubectl](https://kubernetes.io/docs/tasks/tools/install-kubectl/)
|
||||
* Install [Kind](https://kind.sigs.k8s.io/)
|
||||
* Install
|
||||
[Kustomize](https://kubernetes-sigs.github.io/kustomize/installation/binaries/)
|
||||
* Install [Airshipctl](https://docs.airshipit.org/airshipctl/developers.html)
|
||||
|
||||
Also, check [Software Version Information](#Software-Version-Information),
|
||||
* Check [Software Version Information](#Software-Version-Information),
|
||||
[Special Instructions](#Special-Instructions) and [Virtual Machine
|
||||
Specification](#Virtual-Machine-Specification)
|
||||
|
||||
## Getting Started
|
||||
|
||||
Kind will be used to setup a kubernetes cluster, that will be later transformed
|
||||
into a management cluster using airshipctl. The kind kubernetes cluster will be
|
||||
initialized with cluster API and Cluster API docker provider components.
|
||||
|
||||
Run the following command to create a kind configuration file, that would mount
|
||||
the docker.sock file from the host operating system into the kind cluster. This
|
||||
is required by the management cluster to create machines on host as docker
|
||||
containers.
|
||||
|
||||
$ vim kind-cluster-with-extramounts.yaml
|
||||
|
||||
```
|
||||
kind: Cluster
|
||||
apiVersion: kind.sigs.k8s.io/v1alpha3
|
||||
nodes:
|
||||
- role: control-plane
|
||||
extraMounts:
|
||||
- hostPath: /var/run/docker.sock
|
||||
containerPath: /var/run/docker.sock
|
||||
|
||||
```
|
||||
|
||||
Save and exit.
|
||||
|
||||
$ export KIND_EXPERIMENTAL_DOCKER_NETWORK=bridge
|
||||
|
||||
$ kind create cluster --name capi-docker --config ~/kind-cluster-with-extramounts.yaml
|
||||
|
||||
```
|
||||
Creating cluster "capi-docker" ...
|
||||
WARNING: Overriding docker network due to KIND_EXPERIMENTAL_DOCKER_NETWORK
|
||||
WARNING: Here be dragons! This is not supported currently.
|
||||
✓ Ensuring node image (kindest/node:v1.18.2) 🖼
|
||||
✓ Preparing nodes 📦
|
||||
✓ Writing configuration 📜
|
||||
✓ Starting control-plane 🕹️
|
||||
✓ Installing CNI 🔌
|
||||
✓ Installing StorageClass 💾
|
||||
Set kubectl context to "kind-capi-docker"
|
||||
You can now use your cluster with:
|
||||
|
||||
kubectl cluster-info --context kind-capi-docker
|
||||
Check if all the pods are up.
|
||||
```
|
||||
|
||||
$ kubectl get pods -A
|
||||
|
||||
```
|
||||
NAMESPACE NAME READY STATUS RESTARTS AGE
|
||||
kube-system coredns-6955765f44-fvg8p 1/1 Running 0 72s
|
||||
kube-system coredns-6955765f44-gm96d 1/1 Running 0 72s
|
||||
kube-system etcd-capi-docker-control-plane 1/1 Running 0 83s
|
||||
kube-system kindnet-7fqv6 1/1 Running 0 72s
|
||||
kube-system kube-apiserver-capi-docker-control-plane 1/1 Running 0 83s
|
||||
kube-system kube-controller-manager-capi-docker-control-plane 1/1 Running 0 83s
|
||||
kube-system kube-proxy-gqlnm 1/1 Running 0 72s
|
||||
kube-system kube-scheduler-capi-docker-control-plane 1/1 Running 0 83s
|
||||
local-path-storage local-path-provisioner-7745554f7f-2qcv7 1/1 Running 0 72s
|
||||
```
|
||||
|
||||
## Create airshipctl configuration files
|
||||
|
||||
$ mkdir ~/.airship
|
||||
|
||||
$ airshipctl config init
|
||||
|
||||
Run the below command to configure docker manifest, and add it to airship config
|
||||
|
||||
```
|
||||
$ airshipctl config set-manifest docker_manifest --repo primary \
|
||||
--url https://opendev.org/airship/airshipctl --branch master \
|
||||
--phase --sub-path manifests/site/docker-test-site --target-path /tmp/airship
|
||||
```
|
||||
|
||||
$ airshipctl config set-context kind-capi-docker --manifest docker_manifest
|
||||
|
||||
```
|
||||
Context "kind-capi-docker" modified.
|
||||
```
|
||||
$ cp ~/.kube/config ~/.airship/kubeconfig
|
||||
|
||||
$ airshipctl config get-context
|
||||
|
||||
```
|
||||
Context: kind-capi-docker
|
||||
contextKubeconf: kind-capi-docker_target
|
||||
manifest: docker_manifest
|
||||
|
||||
LocationOfOrigin: /home/rishabh/.airship/kubeconfig
|
||||
cluster: kind-capi-docker_target
|
||||
user: kind-capi-docker
|
||||
```
|
||||
$ airshipctl config use-context kind-capi-docker
|
||||
|
||||
```
|
||||
Manifest "docker_manifest" created.
|
||||
```
|
||||
|
||||
$ airshipctl document pull --debug
|
||||
|
||||
```
|
||||
[airshipctl] 2020/08/12 14:07:13 Reading current context manifest information from /home/rishabh/.airship/config
|
||||
[airshipctl] 2020/08/12 14:07:13 Downloading primary repository airshipctl from https://review.opendev.org/airship/airshipctl into /tmp/airship
|
||||
[airshipctl] 2020/08/12 14:07:13 Attempting to download the repository airshipctl
|
||||
[airshipctl] 2020/08/12 14:07:13 Attempting to clone the repository airshipctl from https://review.opendev.org/airship/airshipctl
|
||||
[airshipctl] 2020/08/12 14:07:23 Attempting to checkout the repository airshipctl from branch refs/heads/master
|
||||
```
|
||||
$ airshipctl config set-manifest docker_manifest --target-path /tmp/airship/airshipctl
|
||||
|
||||
## Initialize the management cluster
|
||||
|
||||
$ airshipctl cluster init --debug
|
||||
|
||||
```
|
||||
[airshipctl] 2020/08/12 14:08:23 Starting cluster-api initiation
|
||||
Installing the clusterctl inventory CRD
|
||||
Creating CustomResourceDefinition="providers.clusterctl.cluster.x-k8s.io"
|
||||
Fetching providers
|
||||
[airshipctl] 2020/08/12 14:08:23 Creating arishipctl repository implementation interface for provider cluster-api of type CoreProvider
|
||||
[airshipctl] 2020/08/12 14:08:23 Setting up airshipctl provider Components client
|
||||
Provider type: CoreProvider, name: cluster-api
|
||||
[airshipctl] 2020/08/12 14:08:23 Getting airshipctl provider components, skipping variable substitution: true.
|
||||
Provider type: CoreProvider, name: cluster-api
|
||||
Fetching File="components.yaml" Provider="cluster-api" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:23 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/capi/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:24 Creating arishipctl repository implementation interface for provider kubeadm of type BootstrapProvider
|
||||
[airshipctl] 2020/08/12 14:08:24 Setting up airshipctl provider Components client
|
||||
Provider type: BootstrapProvider, name: kubeadm
|
||||
[airshipctl] 2020/08/12 14:08:24 Getting airshipctl provider components, skipping variable substitution: true.
|
||||
Provider type: BootstrapProvider, name: kubeadm
|
||||
Fetching File="components.yaml" Provider="bootstrap-kubeadm" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:24 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/cabpk/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:24 Creating arishipctl repository implementation interface for provider kubeadm of type ControlPlaneProvider
|
||||
[airshipctl] 2020/08/12 14:08:24 Setting up airshipctl provider Components client
|
||||
Provider type: ControlPlaneProvider, name: kubeadm
|
||||
[airshipctl] 2020/08/12 14:08:24 Getting airshipctl provider components, skipping variable substitution: true.
|
||||
Provider type: ControlPlaneProvider, name: kubeadm
|
||||
Fetching File="components.yaml" Provider="control-plane-kubeadm" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:24 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/cacpk/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:24 Creating arishipctl repository implementation interface for provider docker of type InfrastructureProvider
|
||||
[airshipctl] 2020/08/12 14:08:24 Setting up airshipctl provider Components client
|
||||
Provider type: InfrastructureProvider, name: docker
|
||||
[airshipctl] 2020/08/12 14:08:24 Getting airshipctl provider components, skipping variable substitution: true.
|
||||
Provider type: InfrastructureProvider, name: docker
|
||||
Fetching File="components.yaml" Provider="infrastructure-docker" Version="v0.3.7"
|
||||
[airshipctl] 2020/08/12 14:08:24 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/capd/v0.3.7
|
||||
[airshipctl] 2020/08/12 14:08:24 Creating arishipctl repository implementation interface for provider cluster-api of type CoreProvider
|
||||
Fetching File="metadata.yaml" Provider="cluster-api" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:24 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/capi/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:25 Creating arishipctl repository implementation interface for provider kubeadm of type BootstrapProvider
|
||||
Fetching File="metadata.yaml" Provider="bootstrap-kubeadm" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:25 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/cabpk/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:25 Creating arishipctl repository implementation interface for provider kubeadm of type ControlPlaneProvider
|
||||
Fetching File="metadata.yaml" Provider="control-plane-kubeadm" Version="v0.3.3"
|
||||
[airshipctl] 2020/08/12 14:08:25 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/cacpk/v0.3.3
|
||||
[airshipctl] 2020/08/12 14:08:25 Creating arishipctl repository implementation interface for provider docker of type InfrastructureProvider
|
||||
Fetching File="metadata.yaml" Provider="infrastructure-docker" Version="v0.3.7"
|
||||
[airshipctl] 2020/08/12 14:08:25 Building cluster-api provider component documents from kustomize path at /tmp/airship/airshipctl/manifests/function/capd/v0.3.7
|
||||
Installing cert-manager
|
||||
Creating Namespace="cert-manager"
|
||||
Creating CustomResourceDefinition="challenges.acme.cert-manager.io"
|
||||
Creating CustomResourceDefinition="orders.acme.cert-manager.io"
|
||||
Creating CustomResourceDefinition="certificaterequests.cert-manager.io"
|
||||
Creating CustomResourceDefinition="certificates.cert-manager.io"
|
||||
Creating CustomResourceDefinition="clusterissuers.cert-manager.io"
|
||||
Creating CustomResourceDefinition="issuers.cert-manager.io"
|
||||
Creating ServiceAccount="cert-manager-cainjector" Namespace="cert-manager"
|
||||
Creating ServiceAccount="cert-manager" Namespace="cert-manager"
|
||||
Creating ServiceAccount="cert-manager-webhook" Namespace="cert-manager"
|
||||
Creating ClusterRole="cert-manager-cainjector"
|
||||
Creating ClusterRoleBinding="cert-manager-cainjector"
|
||||
Creating Role="cert-manager-cainjector:leaderelection" Namespace="kube-system"
|
||||
Creating RoleBinding="cert-manager-cainjector:leaderelection" Namespace="kube-system"
|
||||
Creating ClusterRoleBinding="cert-manager-webhook:auth-delegator"
|
||||
Creating RoleBinding="cert-manager-webhook:webhook-authentication-reader" Namespace="kube-system"
|
||||
Creating ClusterRole="cert-manager-webhook:webhook-requester"
|
||||
Creating Role="cert-manager:leaderelection" Namespace="kube-system"
|
||||
Creating RoleBinding="cert-manager:leaderelection" Namespace="kube-system"
|
||||
Creating ClusterRole="cert-manager-controller-issuers"
|
||||
Creating ClusterRole="cert-manager-controller-clusterissuers"
|
||||
Creating ClusterRole="cert-manager-controller-certificates"
|
||||
Creating ClusterRole="cert-manager-controller-orders"
|
||||
Creating ClusterRole="cert-manager-controller-challenges"
|
||||
Creating ClusterRole="cert-manager-controller-ingress-shim"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-issuers"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-clusterissuers"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-certificates"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-orders"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-challenges"
|
||||
Creating ClusterRoleBinding="cert-manager-controller-ingress-shim"
|
||||
Creating ClusterRole="cert-manager-view"
|
||||
Creating ClusterRole="cert-manager-edit"
|
||||
Creating Service="cert-manager" Namespace="cert-manager"
|
||||
Creating Service="cert-manager-webhook" Namespace="cert-manager"
|
||||
Creating Deployment="cert-manager-cainjector" Namespace="cert-manager"
|
||||
Creating Deployment="cert-manager" Namespace="cert-manager"
|
||||
Creating Deployment="cert-manager-webhook" Namespace="cert-manager"
|
||||
Creating APIService="v1beta1.webhook.cert-manager.io"
|
||||
Creating MutatingWebhookConfiguration="cert-manager-webhook"
|
||||
Creating ValidatingWebhookConfiguration="cert-manager-webhook"
|
||||
Waiting for cert-manager to be available...
|
||||
Installing Provider="cluster-api" Version="v0.3.3" TargetNamespace="capi-system"
|
||||
Creating shared objects Provider="cluster-api" Version="v0.3.3"
|
||||
Creating Namespace="capi-webhook-system"
|
||||
Creating CustomResourceDefinition="clusters.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="machinedeployments.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="machinehealthchecks.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="machinepools.exp.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="machines.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="machinesets.cluster.x-k8s.io"
|
||||
Creating MutatingWebhookConfiguration="capi-mutating-webhook-configuration"
|
||||
Creating Service="capi-webhook-service" Namespace="capi-webhook-system"
|
||||
Creating Deployment="capi-controller-manager" Namespace="capi-webhook-system"
|
||||
Creating Certificate="capi-serving-cert" Namespace="capi-webhook-system"
|
||||
Creating Issuer="capi-selfsigned-issuer" Namespace="capi-webhook-system"
|
||||
Creating ValidatingWebhookConfiguration="capi-validating-webhook-configuration"
|
||||
Creating instance objects Provider="cluster-api" Version="v0.3.3" TargetNamespace="capi-system"
|
||||
Creating Namespace="capi-system"
|
||||
Creating Role="capi-leader-election-role" Namespace="capi-system"
|
||||
Creating ClusterRole="capi-system-capi-aggregated-manager-role"
|
||||
Creating ClusterRole="capi-system-capi-manager-role"
|
||||
Creating ClusterRole="capi-system-capi-proxy-role"
|
||||
Creating RoleBinding="capi-leader-election-rolebinding" Namespace="capi-system"
|
||||
Creating ClusterRoleBinding="capi-system-capi-manager-rolebinding"
|
||||
Creating ClusterRoleBinding="capi-system-capi-proxy-rolebinding"
|
||||
Creating Service="capi-controller-manager-metrics-service" Namespace="capi-system"
|
||||
Creating Deployment="capi-controller-manager" Namespace="capi-system"
|
||||
Creating inventory entry Provider="cluster-api" Version="v0.3.3" TargetNamespace="capi-system"
|
||||
Installing Provider="bootstrap-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-bootstrap-system"
|
||||
Creating shared objects Provider="bootstrap-kubeadm" Version="v0.3.3"
|
||||
Creating CustomResourceDefinition="kubeadmconfigs.bootstrap.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="kubeadmconfigtemplates.bootstrap.cluster.x-k8s.io"
|
||||
Creating Service="capi-kubeadm-bootstrap-webhook-service" Namespace="capi-webhook-system"
|
||||
Creating Deployment="capi-kubeadm-bootstrap-controller-manager" Namespace="capi-webhook-system"
|
||||
Creating Certificate="capi-kubeadm-bootstrap-serving-cert" Namespace="capi-webhook-system"
|
||||
Creating Issuer="capi-kubeadm-bootstrap-selfsigned-issuer" Namespace="capi-webhook-system"
|
||||
Creating instance objects Provider="bootstrap-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-bootstrap-system"
|
||||
Creating Namespace="capi-kubeadm-bootstrap-system"
|
||||
Creating Role="capi-kubeadm-bootstrap-leader-election-role" Namespace="capi-kubeadm-bootstrap-system"
|
||||
Creating ClusterRole="capi-kubeadm-bootstrap-system-capi-kubeadm-bootstrap-manager-role"
|
||||
Creating ClusterRole="capi-kubeadm-bootstrap-system-capi-kubeadm-bootstrap-proxy-role"
|
||||
Creating RoleBinding="capi-kubeadm-bootstrap-leader-election-rolebinding" Namespace="capi-kubeadm-bootstrap-system"
|
||||
Creating ClusterRoleBinding="capi-kubeadm-bootstrap-system-capi-kubeadm-bootstrap-manager-rolebinding"
|
||||
Creating ClusterRoleBinding="capi-kubeadm-bootstrap-system-capi-kubeadm-bootstrap-proxy-rolebinding"
|
||||
Creating Service="capi-kubeadm-bootstrap-controller-manager-metrics-service" Namespace="capi-kubeadm-bootstrap-system"
|
||||
Creating Deployment="capi-kubeadm-bootstrap-controller-manager" Namespace="capi-kubeadm-bootstrap-system"
|
||||
Creating inventory entry Provider="bootstrap-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-bootstrap-system"
|
||||
Installing Provider="control-plane-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-control-plane-system"
|
||||
Creating shared objects Provider="control-plane-kubeadm" Version="v0.3.3"
|
||||
Creating CustomResourceDefinition="kubeadmcontrolplanes.controlplane.cluster.x-k8s.io"
|
||||
Creating MutatingWebhookConfiguration="capi-kubeadm-control-plane-mutating-webhook-configuration"
|
||||
Creating Service="capi-kubeadm-control-plane-webhook-service" Namespace="capi-webhook-system"
|
||||
Creating Deployment="capi-kubeadm-control-plane-controller-manager" Namespace="capi-webhook-system"
|
||||
Creating Certificate="capi-kubeadm-control-plane-serving-cert" Namespace="capi-webhook-system"
|
||||
Creating Issuer="capi-kubeadm-control-plane-selfsigned-issuer" Namespace="capi-webhook-system"
|
||||
Creating ValidatingWebhookConfiguration="capi-kubeadm-control-plane-validating-webhook-configuration"
|
||||
Creating instance objects Provider="control-plane-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-control-plane-system"
|
||||
Creating Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating Role="capi-kubeadm-control-plane-leader-election-role" Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating Role="capi-kubeadm-control-plane-manager-role" Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating ClusterRole="capi-kubeadm-control-plane-system-capi-kubeadm-control-plane-manager-role"
|
||||
Creating ClusterRole="capi-kubeadm-control-plane-system-capi-kubeadm-control-plane-proxy-role"
|
||||
Creating RoleBinding="capi-kubeadm-control-plane-leader-election-rolebinding" Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating ClusterRoleBinding="capi-kubeadm-control-plane-system-capi-kubeadm-control-plane-manager-rolebinding"
|
||||
Creating ClusterRoleBinding="capi-kubeadm-control-plane-system-capi-kubeadm-control-plane-proxy-rolebinding"
|
||||
Creating Service="capi-kubeadm-control-plane-controller-manager-metrics-service" Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating Deployment="capi-kubeadm-control-plane-controller-manager" Namespace="capi-kubeadm-control-plane-system"
|
||||
Creating inventory entry Provider="control-plane-kubeadm" Version="v0.3.3" TargetNamespace="capi-kubeadm-control-plane-system"
|
||||
Installing Provider="infrastructure-docker" Version="v0.3.7" TargetNamespace="capd-system"
|
||||
Creating shared objects Provider="infrastructure-docker" Version="v0.3.7"
|
||||
Creating CustomResourceDefinition="dockerclusters.infrastructure.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="dockermachines.infrastructure.cluster.x-k8s.io"
|
||||
Creating CustomResourceDefinition="dockermachinetemplates.infrastructure.cluster.x-k8s.io"
|
||||
Creating ValidatingWebhookConfiguration="capd-validating-webhook-configuration"
|
||||
Creating instance objects Provider="infrastructure-docker" Version="v0.3.7" TargetNamespace="capd-system"
|
||||
Creating Namespace="capd-system"
|
||||
Creating Role="capd-leader-election-role" Namespace="capd-system"
|
||||
Creating ClusterRole="capd-system-capd-manager-role"
|
||||
Creating ClusterRole="capd-system-capd-proxy-role"
|
||||
Creating RoleBinding="capd-leader-election-rolebinding" Namespace="capd-system"
|
||||
Creating ClusterRoleBinding="capd-system-capd-manager-rolebinding"
|
||||
Creating ClusterRoleBinding="capd-system-capd-proxy-rolebinding"
|
||||
Creating Service="capd-controller-manager-metrics-service" Namespace="capd-system"
|
||||
Creating Service="capd-webhook-service" Namespace="capd-system"
|
||||
Creating Deployment="capd-controller-manager" Namespace="capd-system"
|
||||
Creating Certificate="capd-serving-cert" Namespace="capd-system"
|
||||
Creating Issuer="capd-selfsigned-issuer" Namespace="capd-system"
|
||||
Creating inventory entry Provider="infrastructure-docker" Version="v0.3.7" TargetNamespace="capd-system"
|
||||
```
|
||||
|
||||
Wait for all the pods to be up.
|
||||
|
||||
$ kubectl get pods -A
|
||||
```
|
||||
NAMESPACE NAME READY STATUS RESTARTS AGE
|
||||
capd-system capd-controller-manager-75f5d546d7-frrm5 2/2 Running 0 77s
|
||||
capi-kubeadm-bootstrap-system capi-kubeadm-bootstrap-controller-manager-5bb9bfdc46-mhbqz 2/2 Running 0 85s
|
||||
capi-kubeadm-control-plane-system capi-kubeadm-control-plane-controller-manager-77466c7666-t69m5 2/2 Running 0 81s
|
||||
capi-system capi-controller-manager-5798474d9f-tp2c2 2/2 Running 0 89s
|
||||
capi-webhook-system capi-controller-manager-5d64dd9dfb-r6mb2 2/2 Running 1 91s
|
||||
capi-webhook-system capi-kubeadm-bootstrap-controller-manager-7c78fff45-dmnlc 2/2 Running 0 88s
|
||||
capi-webhook-system capi-kubeadm-control-plane-controller-manager-58465bb88f-c6j5q 2/2 Running 0 84s
|
||||
cert-manager cert-manager-69b4f77ffc-8vchm 1/1 Running 0 117s
|
||||
cert-manager cert-manager-cainjector-576978ffc8-frsxg 1/1 Running 0 117s
|
||||
cert-manager cert-manager-webhook-c67fbc858-qxrcj 1/1 Running 1 117s
|
||||
kube-system coredns-6955765f44-f28p7 1/1 Running 0 3m12s
|
||||
kube-system coredns-6955765f44-nq5qk 1/1 Running 0 3m12s
|
||||
kube-system etcd-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kindnet-nxm6k 1/1 Running 0 3m12s
|
||||
kube-system kube-apiserver-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kube-controller-manager-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kube-proxy-5jmc5 1/1 Running 0 3m12s
|
||||
kube-system kube-scheduler-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
local-path-storage local-path-provisioner-7745554f7f-ms989 1/1 Running 0 3m12s
|
||||
```
|
||||
Now, the management cluster is initialized with cluster api and cluster api
|
||||
docker provider components.
|
||||
|
||||
$ kubectl get providers -A
|
||||
|
||||
```
|
||||
NAMESPACE NAME TYPE PROVIDER VERSION WATCH NAMESPACE
|
||||
capd-system infrastructure-docker InfrastructureProvider v0.3.7
|
||||
capi-kubeadm-bootstrap-system bootstrap-kubeadm BootstrapProvider v0.3.3
|
||||
capi-kubeadm-control-plane-system control-plane-kubeadm ControlPlaneProvider v0.3.3
|
||||
capi-system cluster-api CoreProvider v0.3.3
|
||||
```
|
||||
|
||||
$ docker ps
|
||||
|
||||
```
|
||||
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
|
||||
b9690cecdcf2 kindest/node:v1.18.2 "/usr/local/bin/entr…" 14 minutes ago Up 14 minutes 127.0.0.1:32773->6443/tcp capi-docker-control-plane
|
||||
```
|
||||
|
||||
|
||||
## Create your first workload cluster
|
||||
|
||||
$ airshipctl phase run controlplane --debug
|
||||
|
||||
```
|
||||
[airshipctl] 2020/08/12 14:10:12 building bundle from kustomize path /tmp/airship/airshipctl/manifests/site/docker-test-site/target/controlplane
|
||||
[airshipctl] 2020/08/12 14:10:12 Applying bundle, inventory id: kind-capi-docker-target-controlplane
|
||||
[airshipctl] 2020/08/12 14:10:12 Inventory Object config Map not found, auto generating Invetory object
|
||||
[airshipctl] 2020/08/12 14:10:12 Injecting Invetory Object: {"apiVersion":"v1","kind":"ConfigMap","metadata":{"creationTimestamp":null,"labels":{"cli-utils.sigs.k8s.io/inventory-id":"kind-capi-docker-target-controlplane"},"name":"airshipit-kind-capi-docker-target-controlplane","namespace":"airshipit"}}{nsfx:false,beh:unspecified} into bundle
|
||||
[airshipctl] 2020/08/12 14:10:12 Making sure that inventory object namespace airshipit exists
|
||||
configmap/airshipit-kind-capi-docker-target-controlplane-87efb53a created
|
||||
cluster.cluster.x-k8s.io/dtc created
|
||||
machinehealthcheck.cluster.x-k8s.io/dtc-mhc-0 created
|
||||
kubeadmcontrolplane.controlplane.cluster.x-k8s.io/dtc-control-plane created
|
||||
dockercluster.infrastructure.cluster.x-k8s.io/dtc created
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-control-plane created
|
||||
6 resource(s) applied. 6 created, 0 unchanged, 0 configured
|
||||
machinehealthcheck.cluster.x-k8s.io/dtc-mhc-0 is NotFound: Resource not found
|
||||
kubeadmcontrolplane.controlplane.cluster.x-k8s.io/dtc-control-plane is NotFound: Resource not found
|
||||
dockercluster.infrastructure.cluster.x-k8s.io/dtc is NotFound: Resource not found
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-control-plane is NotFound: Resource not found
|
||||
configmap/airshipit-kind-capi-docker-target-controlplane-87efb53a is NotFound: Resource not found
|
||||
cluster.cluster.x-k8s.io/dtc is NotFound: Resource not found
|
||||
configmap/airshipit-kind-capi-docker-target-controlplane-87efb53a is Current: Resource is always ready
|
||||
cluster.cluster.x-k8s.io/dtc is Current: Resource is current
|
||||
machinehealthcheck.cluster.x-k8s.io/dtc-mhc-0 is Current: Resource is current
|
||||
kubeadmcontrolplane.controlplane.cluster.x-k8s.io/dtc-control-plane is Current: Resource is current
|
||||
dockercluster.infrastructure.cluster.x-k8s.io/dtc is Current: Resource is current
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-control-plane is Current: Resource is current
|
||||
all resources has reached the Current status
|
||||
```
|
||||
|
||||
$ kubectl get pods -A
|
||||
```
|
||||
NAMESPACE NAME READY STATUS RESTARTS AGE
|
||||
capd-system capd-controller-manager-75f5d546d7-frrm5 2/2 Running 0 77s
|
||||
capi-kubeadm-bootstrap-system capi-kubeadm-bootstrap-controller-manager-5bb9bfdc46-mhbqz 2/2 Running 0 85s
|
||||
capi-kubeadm-control-plane-system capi-kubeadm-control-plane-controller-manager-77466c7666-t69m5 2/2 Running 0 81s
|
||||
capi-system capi-controller-manager-5798474d9f-tp2c2 2/2 Running 0 89s
|
||||
capi-webhook-system capi-controller-manager-5d64dd9dfb-r6mb2 2/2 Running 1 91s
|
||||
capi-webhook-system capi-kubeadm-bootstrap-controller-manager-7c78fff45-dmnlc 2/2 Running 0 88s
|
||||
capi-webhook-system capi-kubeadm-control-plane-controller-manager-58465bb88f-c6j5q 2/2 Running 0 84s
|
||||
cert-manager cert-manager-69b4f77ffc-8vchm 1/1 Running 0 117s
|
||||
cert-manager cert-manager-cainjector-576978ffc8-frsxg 1/1 Running 0 117s
|
||||
cert-manager cert-manager-webhook-c67fbc858-qxrcj 1/1 Running 1 117s
|
||||
kube-system coredns-6955765f44-f28p7 1/1 Running 0 3m12s
|
||||
kube-system coredns-6955765f44-nq5qk 1/1 Running 0 3m12s
|
||||
kube-system etcd-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kindnet-nxm6k 1/1 Running 0 3m12s
|
||||
kube-system kube-apiserver-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kube-controller-manager-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
kube-system kube-proxy-5jmc5 1/1 Running 0 3m12s
|
||||
kube-system kube-scheduler-capi-docker-control-plane 1/1 Running 0 3m25s
|
||||
local-path-storage local-path-provisioner-7745554f7f-ms989 1/1 Running 0 3m12s
|
||||
```
|
||||
|
||||
$ kubectl logs capd-controller-manager-75f5d546d7-frrm5 -n capd-system --all-containers=true -f
|
||||
|
||||
```
|
||||
I0812 21:11:24.761608 1 controller.go:272] controller-runtime/controller "msg"="Successfully Reconciled" "controller"="dockermachine" "name"="dtc-control-plane-zc5bw" "namespace"="default"
|
||||
I0812 21:11:25.189401 1 controller.go:272] controller-runtime/controller "msg"="Successfully Reconciled" "controller"="dockermachine" "name"="dtc-control-plane-zc5bw" "namespace"="default"
|
||||
I0812 21:11:26.219320 1 generic_predicates.go:38] controllers/DockerMachine "msg"="One of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="All"
|
||||
I0812 21:11:26.219774 1 cluster_predicates.go:143] controllers/DockerMachine "msg"="Cluster was not unpaused, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateUnpaused"
|
||||
I0812 21:11:26.222004 1 cluster_predicates.go:111] controllers/DockerMachine "msg"="Cluster infrastructure did not become ready, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateInfraReady"
|
||||
I0812 21:11:26.223003 1 generic_predicates.go:89] controllers/DockerMachine "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="Any"
|
||||
I0812 21:11:26.223239 1 generic_predicates.go:89] controllers/DockerMachine "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="Any"
|
||||
I0812 21:11:26.219658 1 cluster_predicates.go:143] controllers/DockerCluster "msg"="Cluster was not unpaused, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateUnpaused"
|
||||
I0812 21:11:26.229665 1 generic_predicates.go:89] controllers/DockerCluster "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpaused" "predicateAggregation"="Any"
|
||||
```
|
||||
|
||||
$ kubectl get machines
|
||||
```
|
||||
NAME PROVIDERID PHASE
|
||||
dtc-control-plane-p4fsx docker:////dtc-dtc-control-plane-p4fsx Running
|
||||
```
|
||||
|
||||
$ airshipctl phase run workers --debug
|
||||
|
||||
```
|
||||
[airshipctl] 2020/08/12 14:11:55 building bundle from kustomize path /tmp/airship/airshipctl/manifests/site/docker-test-site/target/worker
|
||||
[airshipctl] 2020/08/12 14:11:55 Applying bundle, inventory id: kind-capi-docker-target-worker
|
||||
[airshipctl] 2020/08/12 14:11:55 Inventory Object config Map not found, auto generating Invetory object
|
||||
[airshipctl] 2020/08/12 14:11:55 Injecting Invetory Object: {"apiVersion":"v1","kind":"ConfigMap","metadata":{"creationTimestamp":null,"labels":{"cli-utils.sigs.k8s.io/inventory-id":"kind-capi-docker-target-worker"},"name":"airshipit-kind-capi-docker-target-worker","namespace":"airshipit"}}{nsfx:false,beh:unspecified} into bundle
|
||||
[airshipctl] 2020/08/12 14:11:55 Making sure that inventory object namespace airshipit exists
|
||||
configmap/airshipit-kind-capi-docker-target-worker-b56f83 created
|
||||
kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/dtc-md-0 created
|
||||
machinedeployment.cluster.x-k8s.io/dtc-md-0 created
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-md-0 created
|
||||
4 resource(s) applied. 4 created, 0 unchanged, 0 configured
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-md-0 is NotFound: Resource not found
|
||||
configmap/airshipit-kind-capi-docker-target-worker-b56f83 is NotFound: Resource not found
|
||||
kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/dtc-md-0 is NotFound: Resource not found
|
||||
machinedeployment.cluster.x-k8s.io/dtc-md-0 is NotFound: Resource not found
|
||||
configmap/airshipit-kind-capi-docker-target-worker-b56f83 is Current: Resource is always ready
|
||||
kubeadmconfigtemplate.bootstrap.cluster.x-k8s.io/dtc-md-0 is Current: Resource is current
|
||||
machinedeployment.cluster.x-k8s.io/dtc-md-0 is Current: Resource is current
|
||||
dockermachinetemplate.infrastructure.cluster.x-k8s.io/dtc-md-0 is Current: Resource is current
|
||||
```
|
||||
|
||||
$ kubectl get machines
|
||||
```
|
||||
NAME PROVIDERID PHASE
|
||||
dtc-control-plane-p4fsx docker:////dtc-dtc-control-plane-p4fsx Running
|
||||
dtc-md-0-94c79cf9c-8ct2g Provisioning
|
||||
```
|
||||
|
||||
$ kubectl logs capd-controller-manager-75f5d546d7-frrm5 -n capd-system --all-containers=true -f
|
||||
|
||||
```
|
||||
I0812 21:10:14.071166 1 cluster_predicates.go:111] controllers/DockerMachine "msg"="Cluster infrastructure did not become ready, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateInfraReady"
|
||||
I0812 21:10:14.071204 1 generic_predicates.go:89] controllers/DockerMachine "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="Any"
|
||||
I0812 21:10:14.071325 1 generic_predicates.go:89] controllers/DockerMachine "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="Any"
|
||||
I0812 21:10:14.082937 1 generic_predicates.go:38] controllers/DockerMachine "msg"="One of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpausedAndInfrastructureReady" "predicateAggregation"="All"
|
||||
I0812 21:10:14.082981 1 cluster_predicates.go:143] controllers/DockerMachine "msg"="Cluster was not unpaused, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateUnpaused"
|
||||
I0812 21:10:14.082994 1 cluster_predicates.go:143] controllers/DockerCluster "msg"="Cluster was not unpaused, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateUnpaused"
|
||||
I0812 21:10:14.083012 1 cluster_predicates.go:111] controllers/DockerMachine "msg"="Cluster infrastructure did not become ready, blocking further processing" "cluster"="dtc" "eventType"="update" "namespace"="default" "predicate"="ClusterUpdateInfraReady"
|
||||
I0812 21:10:14.083024 1 generic_predicates.go:89] controllers/DockerCluster "msg"="All of the provided predicates returned false, blocking further processing" "predicate"="ClusterUnpaused" "predicateAggregation"="Any"
|
||||
I0812 21:10:14.083036 1 generic_predicates.go:89] controllers/DockerMachine "msg"="All of the provided predicates returned false, blocking further processing"
|
||||
```
|
||||
$ kubectl get machines
|
||||
```
|
||||
NAME PROVIDERID PHASE
|
||||
dtc-control-plane-p4fsx docker:////dtc-dtc-control-plane-p4fsx Running
|
||||
dtc-md-0-94c79cf9c-8ct2g docker:////dtc-dtc-md-0-94c79cf9c-8ct2g Running
|
||||
```
|
||||
|
||||
$ kubectl --namespace=default get secret/dtc-kubeconfig -o jsonpath={.data.value} | base64 --decode > ./dtc.kubeconfig
|
||||
|
||||
$ kubectl get nodes --kubeconfig dtc.kubeconfig
|
||||
|
||||
```
|
||||
NAME STATUS ROLES AGE VERSION
|
||||
dtc-dtc-control-plane-p4fsx Ready master 5m45s v1.18.6
|
||||
dtc-dtc-md-0-94c79cf9c-8ct2g Ready <none> 4m45s v1.18.6
|
||||
```
|
||||
|
||||
$ kubectl get pods -A --kubeconfig dtc.kubeconfig
|
||||
```
|
||||
NAMESPACE NAME READY STATUS RESTARTS AGE
|
||||
kube-system calico-kube-controllers-59b699859f-xp8dv 1/1 Running 0 5m40s
|
||||
kube-system calico-node-5drwf 1/1 Running 0 5m39s
|
||||
kube-system calico-node-bqw5j 1/1 Running 0 4m53s
|
||||
kube-system coredns-6955765f44-8kg27 1/1 Running 0 5m40s
|
||||
kube-system coredns-6955765f44-lqzzq 1/1 Running 0 5m40s
|
||||
kube-system etcd-dtc-dtc-control-plane-p4fsx 1/1 Running 0 5m49s
|
||||
kube-system kube-apiserver-dtc-dtc-control-plane-p4fsx 1/1 Running 0 5m49s
|
||||
kube-system kube-controller-manager-dtc-dtc-control-plane-p4fsx 1/1 Running 0 5m49s
|
||||
kube-system kube-proxy-cjcls 1/1 Running 0 5m39s
|
||||
kube-system kube-proxy-fkvpc 1/1 Running 0 4m53s
|
||||
kube-system kube-scheduler-dtc-dtc-control-plane-p4fsx 1/1 Running 0 5m49s
|
||||
```
|
||||
|
||||
## Tear Down Clusters
|
||||
|
||||
The command shown below can be used to delete the control plane, worker nodes and all
|
||||
associated cluster resources
|
||||
|
||||
$ airshipctl phase render --phase-name controlplane -k Cluster | kubectl delete -f -
|
||||
|
||||
```
|
||||
cluster.cluster.x-k8s.io "dtc" deleted
|
||||
```
|
||||
|
||||
To delete the kind cluster, use the following command:
|
||||
|
||||
$ kind delete cluster --name capi-docker
|
||||
|
||||
```
|
||||
Deleting cluster "capi-docker" ...
|
||||
```
|
||||
|
||||
## Reference
|
||||
|
||||
### Provider Manifests
|
||||
|
||||
Provider Configuration is referenced from
|
||||
[config](https://github.com/kubernetes-sigs/cluster-api/tree/master/test/infrastructure/docker/config)
|
||||
Cluster API does not support docker out of the box. Therefore, the metadata
|
||||
infromation is added using files in `airshipctl/manifests/function/capd/data`
|
||||
|
||||
$ tree airshipctl/manifests/function/capd
|
||||
|
||||
```
|
||||
airshipctl/manifests/function/capd
|
||||
└── v0.3.7
|
||||
├── certmanager
|
||||
│ ├── certificate.yaml
|
||||
│ ├── kustomization.yaml
|
||||
│ └── kustomizeconfig.yaml
|
||||
├── crd
|
||||
│ ├── bases
|
||||
│ │ ├── infrastructure.cluster.x-k8s.io_dockerclusters.yaml
|
||||
│ │ ├── infrastructure.cluster.x-k8s.io_dockermachines.yaml
|
||||
│ │ └── infrastructure.cluster.x-k8s.io_dockermachinetemplates.yaml
|
||||
│ ├── kustomization.yaml
|
||||
│ ├── kustomizeconfig.yaml
|
||||
│ └── patches
|
||||
│ ├── cainjection_in_dockerclusters.yaml
|
||||
│ ├── cainjection_in_dockermachines.yaml
|
||||
│ ├── webhook_in_dockerclusters.yaml
|
||||
│ └── webhook_in_dockermachines.yaml
|
||||
├── data
|
||||
│ ├── kustomization.yaml
|
||||
│ └── metadata.yaml
|
||||
├── default
|
||||
│ ├── kustomization.yaml
|
||||
│ └── namespace.yaml
|
||||
├── kustomization.yaml
|
||||
├── manager
|
||||
│ ├── kustomization.yaml
|
||||
│ ├── manager_auth_proxy_patch.yaml
|
||||
│ ├── manager_image_patch.yaml
|
||||
│ ├── manager_prometheus_metrics_patch.yaml
|
||||
│ ├── manager_pull_policy.yaml
|
||||
│ └── manager.yaml
|
||||
├── rbac
|
||||
│ ├── auth_proxy_role_binding.yaml
|
||||
│ ├── auth_proxy_role.yaml
|
||||
│ ├── auth_proxy_service.yaml
|
||||
│ ├── kustomization.yaml
|
||||
│ ├── leader_election_role_binding.yaml
|
||||
│ ├── leader_election_role.yaml
|
||||
│ ├── role_binding.yaml
|
||||
│ └── role.yaml
|
||||
└── webhook
|
||||
├── kustomization.yaml
|
||||
├── kustomizeconfig.yaml
|
||||
├── manager_webhook_patch.yaml
|
||||
├── manifests.yaml
|
||||
├── service.yaml
|
||||
└── webhookcainjection_patch.yaml
|
||||
|
||||
10 directories, 37 files
|
||||
```
|
||||
|
||||
### Cluster Templates
|
||||
|
||||
`manifests/function/k8scontrol-capd` contains cluster.yaml, controlplane.yaml
|
||||
templates referenced from
|
||||
[cluster-template](https://github.com/kubernetes-sigs/cluster-api/blob/master/test/e2e/data/infrastructure-docker/cluster-template.yaml)
|
||||
|
||||
|
||||
| Template Name | CRDs |
|
||||
| ----------------- | ---- |
|
||||
| cluster.yaml | Cluster, DockerCluster |
|
||||
| controlplane.yaml | KubeadmControlPlane, DockerMachineTemplate, MachineHealthCheck |
|
||||
|
||||
$ tree airshipctl/manifests/function/k8scontrol-capd
|
||||
|
||||
```
|
||||
airshipctl/manifests/function/k8scontrol-capd
|
||||
├── cluster.yaml
|
||||
├── controlplane.yaml
|
||||
└── kustomization.yaml
|
||||
```
|
||||
|
||||
`airshipctl/manifests/function/workers-capd` contains workers.yaml referenced
|
||||
from
|
||||
[cluster-template](https://github.com/kubernetes-sigs/cluster-api/blob/master/test/e2e/data/infrastructure-docker/cluster-template.yaml)
|
||||
|
||||
| Template Name | CRDs |
|
||||
| ----------------- | ---- |
|
||||
| workers.yaml | KubeadmConfigTemplate, MachineDeployment, DockerMachineTemplate |
|
||||
|
||||
$ tree airshipctl/manifests/function/workers-capd
|
||||
|
||||
```
|
||||
airshipctl/manifests/function/workers-capd
|
||||
├── kustomization.yaml
|
||||
└── workers.yaml
|
||||
```
|
||||
|
||||
### Test Site Manifests
|
||||
|
||||
#### docker-test-site/shared
|
||||
|
||||
`airshipctl cluster init` uses `airshipctl/manifests/site/docker-test-site/shared/clusterctl`
|
||||
to initialize management cluster with defined provider components and version.
|
||||
|
||||
`$ tree airshipctl/manifests/site/docker-test-site/shared`
|
||||
|
||||
```
|
||||
/tmp/airship/airshipctl/manifests/site/docker-test-site/shared
|
||||
└── clusterctl
|
||||
├── clusterctl.yaml
|
||||
└── kustomization.yaml
|
||||
|
||||
1 directory, 2 files
|
||||
```
|
||||
|
||||
#### docker-test-site/target
|
||||
|
||||
There are 3 phases currently available in `docker-test-site/target`.
|
||||
|
||||
| Phase Name | Purpose |
|
||||
| ----------- | --------- |
|
||||
| controlplane | Patches templates in manifests/function/k8scontrol-capd |
|
||||
| workers | Patches template in manifests/function/workers-capd |
|
||||
| initinfra | Simply calls `docker-test-site/shared/clusterctl` |
|
||||
|
||||
Note: `airshipctl cluster init` initializes all the provider components
|
||||
including the docker infrastructure provider component. As a result, `airshipctl
|
||||
phase run initinfra-ephemeral` is not used.
|
||||
|
||||
At the moment, `phase initinfra` is only present for two reasons:
|
||||
- `airshipctl` complains if the phase is not found
|
||||
- `validate site docs to pass`
|
||||
|
||||
#### Patch Merge Strategy
|
||||
|
||||
Json patch to patch `control plane machine count` is applied on templates in `manifests/function/k8scontrol-capd`
|
||||
from `airshipctl/manifests/site/docker-test-site/target/controlplane` when `airshipctl phase run
|
||||
controlplane` is executed
|
||||
|
||||
Json patch to patch `workers machine count` is applied on templates in `manifests/function/workers-capd`
|
||||
from `airshipctl/manifests/site/docker-test-site/target/workers` when `airshipctl phase run
|
||||
workers` is executed.
|
||||
|
||||
| Patch Name | Purpose |
|
||||
| ---------- | -------- |
|
||||
| controlplane/machine_count.json | patches control plane machine count in template function/k8scontrol-capd |
|
||||
| workers/machine_count.json | patches workers machine count in template function/workers-capd |
|
||||
|
||||
$ tree airshipctl/manifests/site/docker-test-site/target/
|
||||
|
||||
```
|
||||
airshipctl/manifests/site/docker-test-site/target/
|
||||
├── controlplane
|
||||
│ ├── kustomization.yaml
|
||||
│ └── machine_count.json
|
||||
├── initinfra
|
||||
│ └── kustomization.yaml
|
||||
└── workers
|
||||
├── kustomization.yaml
|
||||
└── machine_count.json
|
||||
|
||||
3 directories, 7 files
|
||||
```
|
||||
|
||||
### Software Version Information
|
||||
|
||||
@ -726,10 +236,9 @@ All the instructions in the document were perfomed on a Oracle Virtual Box(6.1)
|
||||
VM running Ubuntu 18.04.4 LTS (Bionic Beaver) with 16G of memory and 4 VCPUs
|
||||
|
||||
#### Docker
|
||||
|
||||
```
|
||||
$ docker version
|
||||
|
||||
```
|
||||
Client: Docker Engine - Community
|
||||
Version: 19.03.9
|
||||
API version: 1.40
|
||||
@ -761,42 +270,31 @@ Server: Docker Engine - Community
|
||||
|
||||
#### Kind
|
||||
|
||||
```
|
||||
$ kind version
|
||||
|
||||
```
|
||||
kind v0.8.1 go1.14.2 linux/amd64
|
||||
```
|
||||
|
||||
#### Kubectl
|
||||
|
||||
```
|
||||
$ kubectl version
|
||||
|
||||
```
|
||||
Client Version: version.Info{Major:"1", Minor:"17", GitVersion:"v1.17.4", GitCommit:"8d8aa39598534325ad77120c120a22b3a990b5ea", GitTreeState:"clean", BuildDate:"2020-03-12T21:03:42Z", GoVersion:"go1.13.8", Compiler:"gc", Platform:"linux/amd64"}
|
||||
Server Version: version.Info{Major:"1", Minor:"17", GitVersion:"v1.17.0", GitCommit:"70132b0f130acc0bed193d9ba59dd186f0e634cf", GitTreeState:"clean", BuildDate:"2020-01-14T00:09:19Z", GoVersion:"go1.13.4", Compiler:"gc", Platform:"linux/amd64"}
|
||||
```
|
||||
|
||||
#### Go
|
||||
|
||||
$ go version
|
||||
|
||||
```
|
||||
$ go version
|
||||
go version go1.14.1 linux/amd64
|
||||
```
|
||||
|
||||
#### Kustomize
|
||||
|
||||
$ kustomize version
|
||||
|
||||
```
|
||||
{Version:kustomize/v3.8.0 GitCommit:6a50372dd5686df22750b0c729adaf369fbf193c BuildDate:2020-07-05T14:08:42Z GoOs:linux GoArch:amd64}
|
||||
```
|
||||
|
||||
#### OS
|
||||
|
||||
```
|
||||
$ cat /etc/os-release
|
||||
|
||||
```
|
||||
NAME="Ubuntu"
|
||||
VERSION="18.04.4 LTS (Bionic Beaver)"
|
||||
ID=ubuntu
|
||||
@ -811,6 +309,6 @@ VERSION_CODENAME=bionic
|
||||
UBUNTU_CODENAME=bionic
|
||||
```
|
||||
|
||||
## Special Instructions
|
||||
### Special Instructions
|
||||
|
||||
Swap was disabled on the VM using `sudo swapoff -a`
|
||||
|
Loading…
Reference in New Issue
Block a user