Pronounced: ClustARRRR-APIARRRR-plunder.
Warning: This provider is so untested that I would only recommend using it against your worst enemies, that being said.. if you're feeling brave then be my guest :D
The cluster-api-plunder is a Cluster-API provider that extends the capabilities of a Kubernetes cluster so that it can be used to not only manage the provisioning of applications and services, but also the provisioning of actual infrastructure to host additional Kubernetes clusters.
This provider does this by "translating" infrastructure requests from Cluster-API and using plunder to provision clusters and the required machines that make up a complete Kubernetes cluster.
Cloud environments typically have a massive agile advantage when it comes to provisioning infrastructure, typically the ground work (bare-metal provisioning) it already taken care of (I'm presuming through the automation of api-human). Meaning that when you ask for some infrastructure then the cloud provider will be cloning VM templates etc.. to quickly get some machine infrastructure back to you.
In order to handle quick provisioning today plunder watches for machines starting up that typically are blank/new and need to try and boot, it will register their MAC addresses and place them in a reboot loop (until we need them).
$ pldrctl get unleased Mac Address Hardware Vendor Time Seen Time since 00:50:56:a5:11:20 VMware, Inc. Sun Nov 3 10:54:18 2019 0s 00:50:56:a5:b5:f1 VMware, Inc. Sun Nov 3 10:54:18 2019 0s NOTE: There are more efficent ways of doing this, just haven't had the time ¯\_(ツ)_/¯
When it comes to provisioning, we simply flip a server from reboot to provision and "hey presto"
At the moment, there is still a few steps that are needed to get this all up and running and i've yet to get the provider in a kubernetes deployment (someone feel free to raise an issue).
Add the seret that will contain the plunder config, the --from-file should point to the location of the client certificate.
k create secret generic plunder --from-file=./plunderclient.yaml --namespace=capi-system To use the created ones:
kubectl apply -f https://github.com/plunder-app/cluster-api-plunder/raw/master/cluster-api-plunder-components.yaml To generate them within the source code / project:
make install Then verify them with kubectl get crds | grep plunder.
If you are using the manifest that is part of this repository and have created the certificate, then the controller should be up and running as expected.
Alternatively you can build/run the controller locally as detailed below
Copy the plunderclient.yaml file to the same location that the controller will run.
make run will then start the controller.
Cluster.yaml should typically look like below the cidrBlocks will define the range of addresses used by pods started within the cluster.
apiVersion: cluster.x-k8s.io/v1alpha2 kind: Cluster metadata: name: cluster-plunder spec: clusterNetwork: pods: cidrBlocks: ["192.168.0.0/16"] infrastructureRef: apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderCluster name: cluster-plunder --- apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderCluster metadata: name: cluster-plunder IPAM isn't completed (lol.. it's not started), so currently you'll need to specify addresses for machines, this will need fixing for machineSets
Also:
- In
plunderMachine.spec=>deploymentTypeis required in order for Plunder to know what to provision. - In
machine.spec=>versionis required to determine the version of Kubernetes to provision.
Machine.yaml should looks something like below:
apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderMachine metadata: name: controlplane namespace: default spec: ipaddress: 192.168.1.123 deploymentType: preseed --- apiVersion: cluster.x-k8s.io/v1alpha2 kind: Machine metadata: labels: cluster.x-k8s.io/cluster-name: cluster-plunder cluster.x-k8s.io/control-plane: "true" name: controlplane namespace: default spec: version: "v1.14.2" bootstrap: data: "" infrastructureRef: apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderMachine name: controlplane namespace: default --- apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderMachine metadata: name: worker namespace: default spec: ipaddress: 192.168.1.124 deploymentType: preseed --- apiVersion: cluster.x-k8s.io/v1alpha2 kind: Machine metadata: labels: cluster.x-k8s.io/cluster-name: cluster-plunder name: worker namespace: default spec: version: "v1.14.2" bootstrap: data: "" infrastructureRef: apiVersion: infrastructure.cluster.x-k8s.io/v1alpha1 kind: PlunderMachine name: worker namespace: default The same manifests are in examples/simple and can be deployed through kubectl with the command:
kubectl -f examples/simple/cluster.yaml and kubectl create -f examples/simple/machine.yaml
k get machines NAME PROVIDERID PHASE controlplane provisioning worker provisioning < 7-ish mins later> k get machines NAME PROVIDERID PHASE controlplane plunder://00:50:56:a5:b5:f1 provisioned worker plunder://00:50:56:a5:11:20 provisioned k get events LAST SEEN TYPE REASON OBJECT MESSAGE 50m Warning No Hardware found plundermachine/controlplane Plunder has no available hardware to provision 47m Normal PlunderProvision plundermachine/controlplane Plunder has begun provisioning the Operating System 40m Normal PlunderProvision plundermachine/controlplane Host has been succesfully provisioned OS in 7m1s Seconds 40m Normal PlunderProvision plundermachine/worker Plunder has begun provisioning the Operating System 33m Normal PlunderProvision plundermachine/worker Host has been succesfully provisioned OS in 7m6s Seconds There are two methods for removing the deployed machines:
kubectl delete machines --all or kubectl delete -f ./examples/simple/machine.yaml
This process will wipe the boot sector and beginning of the disk which will result in it booting into a "blank enough" state for plunder to add it back to the reboot loop.