Getting Started with kOps on GCE

Make sure you have installed kOps and installed kubectl, and installed the gcloud tools.

You’ll need a Google Cloud account, and make sure that gcloud is logged in to your account using gcloud init.

You should confirm that basic commands like gcloud compute zones list are working.

You’ll also need to configure default credentials, using gcloud auth application-default login.

Creating a state store

kOps needs a state store, to hold the configuration for your clusters. The simplest configuration for Google Cloud is to store it in a Google Cloud Storage bucket in the same account, so that’s how we’ll start.

So, just create an empty bucket - you can use any (available) name - e.g. gsutil mb gs://kubernetes-clusters/

Further, rather than typing the --state argument every time, it’s much easier to export the KOPS_STATE_STORE environment variable:

  1. export KOPS_STATE_STORE=gs://kubernetes-clusters/

You can also put this in your ~/.bashrc or similar.

Creating our first cluster

kops create cluster creates the Cluster object and InstanceGroup object you’ll be working with in kOps:

  1. PROJECT=`gcloud config get-value project`
  2. export KOPS_FEATURE_FLAGS=AlphaAllowGCE # to unlock the GCE features
  3. kops create cluster simple.k8s.local --zones us-central1-a --state ${KOPS_STATE_STORE}/ --project=${PROJECT}

You can now list the Cluster objects in your kOps state store (the GCS bucket we created).

  1. > kops get cluster --state ${KOPS_STATE_STORE}
  2. NAME CLOUD ZONES
  3. simple.k8s.local gce us-central1-a

This shows that you have one Cluster object configured, named simple.k8s.local. The cluster holds the cluster-wide configuration for a kubernetes cluster - things like the kubernetes version, and the authorization policy in use.

The kops tool should feel a lot like kubectl - kOps uses the same API machinery as kubernetes, so it should behave similarly, although now you are managing kubernetes clusters, instead of managing objects on a kubernetes cluster.

You can see the details of your Cluster object by doing:

kops get cluster --state ${KOPS_STATE_STORE}/ simple.k8s.local -oyaml

  1. apiVersion: kops.k8s.io/v1alpha2
  2. kind: Cluster
  3. metadata:
  4. name: simple.k8s.local
  5. spec:
  6. api:
  7. loadBalancer:
  8. type: Public
  9. authorization:
  10. alwaysAllow: {}
  11. channel: stable
  12. cloudProvider: gce
  13. configBase: gs://kubernetes-clusters/simple.k8s.local
  14. etcdClusters:
  15. - etcdMembers:
  16. - instanceGroup: master-us-central1-a
  17. name: a
  18. name: main
  19. - etcdMembers:
  20. - instanceGroup: master-us-central1-a
  21. name: a
  22. name: events
  23. iam:
  24. legacy: false
  25. kubernetesApiAccess:
  26. - 0.0.0.0/0
  27. kubernetesVersion: 1.7.2
  28. masterPublicName: api.simple.k8s.local
  29. networking:
  30. kubenet: {}
  31. nonMasqueradeCIDR: 100.64.0.0/10
  32. project: my-gce-project
  33. sshAccess:
  34. - 0.0.0.0/0
  35. subnets:
  36. - name: us-central1
  37. region: us-central1
  38. type: Public
  39. topology:
  40. dns:
  41. type: Public
  42. masters: public
  43. nodes: public

Similarly, you can also see your InstanceGroups using:

  1. > kops get instancegroup --state ${KOPS_STATE_STORE}/ --name simple.k8s.local
  2. NAME ROLE MACHINETYPE MIN MAX SUBNETS
  3. master-us-central1-a Master n1-standard-1 1 1 us-central1
  4. nodes Node n1-standard-2 2 2 us-central1

InstanceGroups are the other main kOps object - an InstanceGroup manages a set of cloud instances, which then are registered in kubernetes as Nodes. You have multiple InstanceGroups for different types of instances / Nodes - in our simple example we have one for our master (which only has a single member), and one for our nodes (and we have two nodes configured).

We’ll see a lot more of Cluster objects and InstanceGroups as we use kOps to reconfigure clusters. But let’s get on with our first cluster.

Creating a cluster

kops create cluster created the Cluster object & InstanceGroup object in our state store, but didn’t actually create any instances or other cloud objects in GCE. To do that, we’ll use kops update cluster.

kops update cluster without --yes will show you a preview of all the changes will be made; it is very useful to see what kOps is about to do, before actually making the changes.

Run kops update cluster simple.k8s.local and peruse the changes.

We’re now finally ready to create the object: kops update cluster simple.k8s.local --yes

(If you haven’t created an SSH key, you’ll have to ssh-keygen -t rsa)

Your cluster is created in the background - kOps actually creates GCE Managed Instance Groups that run the instances; this ensures that even if instances are terminated, they will automatically be relaunched by GCE and your cluster will self-heal.

After a few minutes, you should be able to do kubectl get nodes and your first cluster should be ready!

Enjoy

At this point you have a kubernetes cluster - the core commands to do so are as simple as kops create cluster and kops update cluster. There’s a lot more power in kOps, and even more power in kubernetes itself, so we’ve put a few jumping off places here. But when you’re done, don’t forget to delete your cluster.

Deleting the cluster

When you’re done using the cluster, you should delete it to release the cloud resources. kops delete cluster is the command. When run without --yes it shows a preview of the objects it will delete:

  1. > kops delete cluster simple.k8s.local
  2. TYPE NAME ID
  3. Address api-simple-k8s-local api-simple-k8s-local
  4. Disk a-etcd-events-simple-k8s-local a-etcd-events-simple-k8s-local
  5. Disk a-etcd-main-simple-k8s-local a-etcd-main-simple-k8s-local
  6. ForwardingRule api-simple-k8s-local api-simple-k8s-local
  7. Instance master-us-central1-a-9847 us-central1-a/master-us-central1-a-9847
  8. Instance nodes-0s0w us-central1-a/nodes-0s0w
  9. Instance nodes-dvlq us-central1-a/nodes-dvlq
  10. InstanceGroupManager a-master-us-central1-a-simple-k8s-local us-central1-a/a-master-us-central1-a-simple-k8s-local
  11. InstanceGroupManager a-nodes-simple-k8s-local us-central1-a/a-nodes-simple-k8s-local
  12. InstanceTemplate master-us-central1-a-simple-k8s-local-1507008700 master-us-central1-a-simple-k8s-local-1507008700
  13. InstanceTemplate nodes-simple-k8s-local-1507008700 nodes-simple-k8s-local-1507008700
  14. Route simple-k8s-local-715bb0c7-a7fc-11e7-93d7-42010a800002 simple-k8s-local-715bb0c7-a7fc-11e7-93d7-42010a800002
  15. Route simple-k8s-local-9a2a08e8-a7fc-11e7-93d7-42010a800002 simple-k8s-local-9a2a08e8-a7fc-11e7-93d7-42010a800002
  16. Route simple-k8s-local-9c17a4e6-a7fc-11e7-93d7-42010a800002 simple-k8s-local-9c17a4e6-a7fc-11e7-93d7-42010a800002
  17. TargetPool api-simple-k8s-local api-simple-k8s-local
  18. Must specify --yes to delete cluster

After you’ve double-checked you’re deleting exactly what you want to delete, run kops delete cluster simple.k8s.local --yes.

Next steps

Now that you have a working kOps cluster, read through the recommendations for production setups guide to learn more about how to configure kOps for production workloads.