Skip to content

satyapatibandla/kubernetes-multi-master

 
 

Repository files navigation

Multi Master Kubernetes Cluster

Create a multi master kubernetes cluster

Steps to create a Multi Master kubernetes cluster

Servers -

  1. adminnode - Loadbalancer (10.142.0.8)
  2. 3 Masters - Kmaster1, kmaster2, kmaster3 (10.142.0.2, 10.142.0.3, 10.142.0.4)
  3. 3 Nodes - knode1, knode2, knode3 (10.142.0.5, 10.142.0.6, 10.142.0.7)
  4. Pre- requisites - a. SSH enabled on all servers b. servers are able to ssh to each other

Admin Node

  1. cd multimaster

  2. Run : ./1_install_cffsl.sh to install cffsl (used to generate SSL keys)

  3. Run : ./2_install_kubectl_admin.sh to install kubectl executable

  4. Run : ./3_install_haproxy.sh to install haproxy for Loadbalancing 3 masters

  5. As root : cat haproxy_cfg >> /etc/haproxy/haproxy.cfg - This will add haproxy configurations to haproxy.cfg for frontend and backend

  6. As root run sudo systemctl restart haproxy

  7. Generate certificates for CA and ETCD cluster as below

    a. cd certs

    b. ca-config.json and ca-csr.json is already provided with dummy examples. Make changes accordingly as per your organization standards

    c. execute - ./4_generateca.sh to generate the root CA file and private key

    d. Files kubernetes-csr.json is already provided with dummy examples. Make changes to them accordingly

    e. execute - ./5_generateetcdcert.sh to generate certificates for ETCD cluster. Make sure to make changes related to Master IP addresses inside 5_generateetcdcert.sh script.

    f. scp ca.pem kubernetes.pem kubernetes-key.pem to all nodes from admin Node

Master Nodes and minion nodes -

On all the nodes master + minion - run the below -

  1. cd multimaster/install_kubeadm

  2. Run - ./6_install_docker_kubeadm.sh to install docker + kubeadm + kubelet + kubectl

Only on Master Nodes - Install etcd on master nodes as below -

  1. cd multimaster/install_etcd

  2. Run ./7_install_etcd.sh - This will install ETCD and will copy the certificates from your home account to /etc/etcd

  3. The files etcd.service.kmaster1/2/3 are provided with etcd default configurations. Please make sure that you edit them according to the IP addresses used to generate the certificates.

  4. cp etcd.service.kmaster1 /etc/systemd/system/etcd.service - on kmaster1

  5. cp etcd.service.kmaster2 /etc/systemd/system/etcd.service - on kmaster2

  6. cp etcd.service.kmaster3 /etc/systemd/system/etcd.service - on kmaster3

  7. Run systemctl daemon-reload

  8. Run systemctl enable etcd

  9. Run systemctl start etcd

  10. Verify the cluster using - ETCDCTL_API=3 etcdctl member list

Output as below - root@kmaster1:~# ETCDCTL_API=3 etcdctl member list 56fbe3d44c581b82, started, kmaster2, https://10.142.0.3:2380, https://10.142.0.3:2379 80bd93ff16829bfc, started, kmaster3, https://10.142.0.4:2380, https://10.142.0.4:2379 a3da5fa43de07c2f, started, kmaster1, https://10.142.0.2:2380, https://10.142.0.2:2379

Only on master nodes -

Initialize cluster as below -

  1. Login to any one of the master (kmaster1)

  2. cd multimaster/init_kubeadm

  3. File config.yaml is provided with default kubeadm configuration required to initialize kubeadm. Please modify the ip addresses or names of certificates accordingly. You can also edit the pod subnet in the default configuration. Additionally - apiServerCertSANs and controlPlaneEndpoint must be the ip address of the loadbalancer or the adminnode.

kubeadm config migrate --old-config config.yaml --new-config new.config.yaml

  1. Run : 9_kubeadminit.sh to initialize the cluster with the above config.yaml. Make sure to copy the last few lines for kubeadm join for nodes

  2. on the root home of kmaster1 -

    a. tar cvf pki.tar /etc/kubernetes/pki

    b. scp this tar to kmaster2 and kmaster3

  3. Login to Kmaster2 and kmaster3 and do the below

  4. tar xvf pki.tar

  5. cd inside pki directory and run - rm apiserver.*

  6. mv pki /etc/kubernetes/

  7. cd multimaster/init_kubeadm

  8. execute - ./9_kubeadminit.sh

On Worker / Minion Nodes -

  1. Run the kubeadm join command on all nodes

  2. Login to any master and execute - kubectl --kubeconfig /etc/kubernetes/admin.conf get nodes to see if all nodes and masters are joined

Initialize the cluster on admin node for secure access -

  1. Login to kmaster and run - chmod +r /etc/kubernetes/admin.conf

  2. scp admin.conf to adminode root home directory

  3. login to adminnode

  4. mkdir ~/.kube

  5. cp admin.conf ~/.kube/config

  6. chmod 600 ~/.kube/config

  7. on kmaster1 chmod 600 /etc/kubernetes/admin.conf

  8. You can now start using adminnode for all kubernetes cluster related activities without giving master access to everyone.

  9. Add CNI from admin Node: kubectl apply -f "https://cloud.weave.works/k8s/net?k8s-version=$(kubectl version | base64 | tr -d '\n')"

About

Create a multi masterkubernetes cluster

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Shell 100.0%