OpenStack

Known compatible public clouds

Kubespray has been tested on a number of OpenStack Public Clouds including (in alphabetical order):

The in-tree cloud provider

To deploy Kubespray on OpenStack uncomment the cloud_provider option in group_vars/all/all.yml and set it to openstack.

After that make sure to source in your OpenStack credentials like you would do when using nova-client or neutron-client by using source path/to/your/openstack-rc or . path/to/your/openstack-rc.

For those who prefer to pass the OpenStack CA certificate as a string, one can base64 encode the cacert file and store it in the variable openstack_cacert.

The next step is to make sure the hostnames in your inventory file are identical to your instance names in OpenStack. Otherwise cinder won’t work as expected.

Unless you are using calico or kube-router you can now run the playbook.

Additional step needed when using calico or kube-router:

Being L3 CNI, calico and kube-router do not encapsulate all packages with the hosts’ ip addresses. Instead the packets will be routed with the PODs ip addresses directly.

OpenStack will filter and drop all packets from ips it does not know to prevent spoofing.

In order to make L3 CNIs work on OpenStack you will need to tell OpenStack to allow pods packets by allowing the network they use.

First you will need the ids of your OpenStack instances that will run kubernetes:

  1. openstack server list --project YOUR_PROJECT
  2. +--------------------------------------+--------+----------------------------------+--------+-------------+
  3. | ID | Name | Tenant ID | Status | Power State |
  4. +--------------------------------------+--------+----------------------------------+--------+-------------+
  5. | e1f48aad-df96-4bce-bf61-62ae12bf3f95 | k8s-1 | fba478440cb2444a9e5cf03717eb5d6f | ACTIVE | Running |
  6. | 725cd548-6ea3-426b-baaa-e7306d3c8052 | k8s-2 | fba478440cb2444a9e5cf03717eb5d6f | ACTIVE | Running |

Then you can use the instance ids to find the connected neutron ports (though they are now configured through using OpenStack):

  1. openstack port list -c id -c device_id --project YOUR_PROJECT
  2. +--------------------------------------+--------------------------------------+
  3. | id | device_id |
  4. +--------------------------------------+--------------------------------------+
  5. | 5662a4e0-e646-47f0-bf88-d80fbd2d99ef | e1f48aad-df96-4bce-bf61-62ae12bf3f95 |
  6. | e5ae2045-a1e1-4e99-9aac-4353889449a7 | 725cd548-6ea3-426b-baaa-e7306d3c8052 |

Given the port ids on the left, you can set the two allowed-address(es) in OpenStack. Note that you have to allow both kube_service_addresses (default 10.233.0.0/18) and kube_pods_subnet (default 10.233.64.0/18.)

  1. # allow kube_service_addresses and kube_pods_subnet network
  2. openstack port set 5662a4e0-e646-47f0-bf88-d80fbd2d99ef --allowed-address ip-address=10.233.0.0/18 --allowed-address ip-address=10.233.64.0/18
  3. openstack port set e5ae2045-a1e1-4e99-9aac-4353889449a7 --allowed-address ip-address=10.233.0.0/18 --allowed-address ip-address=10.233.64.0/18

If all the VMs in the tenant correspond to Kubespray deployment, you can “sweep run” above with:

  1. openstack port list --device-owner=compute:nova -c ID -f value | xargs -tI@ openstack port set @ --allowed-address ip-address=10.233.0.0/18 --allowed-address ip-address=10.233.64.0/18

Now you can finally run the playbook.

The external cloud provider

The in-tree cloud provider is deprecated and will be removed in a future version of Kubernetes. The target release for removing all remaining in-tree cloud providers is set to 1.21.

The new cloud provider is configured to have Octavia by default in Kubespray.

  • Enable the new external cloud provider in group_vars/all/all.yml:

    1. cloud_provider: external
    2. external_cloud_provider: openstack
  • Enable Cinder CSI in group_vars/all/openstack.yml:

    1. cinder_csi_enabled: true
  • Enable topology support (optional), if your openstack provider has custom Zone names you can override the default “nova” zone by setting the variable cinder_topology_zones

    1. cinder_topology: true
  • If you are using OpenStack loadbalancer(s) replace the openstack_lbaas_subnet_id with the new external_openstack_lbaas_subnet_id. Note The new cloud provider is using Octavia instead of Neutron LBaaS by default!

  • Enable 3 feature gates to allow migration of all volumes and storage classes (if you have any feature gates already set just add the 3 listed below):

    1. kube_feature_gates:
    2. - CSIMigration=true
    3. - CSIMigrationOpenStack=true
    4. - ExpandCSIVolumes=true
  • If you are in a case of a multi-nic OpenStack VMs (see kubernetes/cloud-provider-openstack#407 and #6083 for explanation), you should override the default OpenStack networking configuration:

    1. external_openstack_network_ipv6_disabled: false
    2. external_openstack_network_internal_networks: []
    3. external_openstack_network_public_networks: []
  • You can override the default OpenStack metadata configuration (see #6338 for explanation):

    1. external_openstack_metadata_search_order: "configDrive,metadataService"
  • Available variables for configuring lbaas:

    1. external_openstack_lbaas_create_monitor: false
    2. external_openstack_lbaas_monitor_delay: "1m"
    3. external_openstack_lbaas_monitor_timeout: "30s"
    4. external_openstack_lbaas_monitor_max_retries: "3"
    5. external_openstack_lbaas_provider: octavia
    6. external_openstack_lbaas_use_octavia: false
    7. external_openstack_lbaas_network_id: "Neutron network ID to create LBaaS VIP"
    8. external_openstack_lbaas_subnet_id: "Neutron subnet ID to create LBaaS VIP"
    9. external_openstack_lbaas_floating_network_id: "Neutron network ID to get floating IP from"
    10. external_openstack_lbaas_floating_subnet_id: "Neutron subnet ID to get floating IP from"
    11. external_openstack_lbaas_method: "ROUND_ROBIN"
    12. external_openstack_lbaas_manage_security_groups: false
    13. external_openstack_lbaas_internal_lb: false
  • Run source path/to/your/openstack-rc to read your OpenStack credentials like OS_AUTH_URL, OS_USERNAME, OS_PASSWORD, etc. Those variables are used for accessing OpenStack from the external cloud provider.

  • Run the cluster.yml playbook