Introduction
What does Cilium provide in your Kubernetes Cluster?
The following functionality is provided as your run Cilium in your Kubernetes cluster:
CNI
plugin support to provide pod_connectivity with Networking.- Identity based implementation of the NetworkPolicy resource to isolate
pod
topod
connectivity on Layer 3 and 4. - An extension to NetworkPolicy in the form of a
CustomResourceDefinition
which extends policy control to add:- Layer 7 policy enforcement on ingress and egress for the following application protocols:
- HTTP
- Kafka
- Egress support for CIDRs to secure access to external services
- Enforcement to external headless services to automatically restrict to the set of Kubernetes endpoints configured for a service.
- Layer 7 policy enforcement on ingress and egress for the following application protocols:
- ClusterIP implementation to provide distributed load-balancing for pod to pod traffic.
- Fully compatible with existing kube-proxy model
Pod-to-Pod Connectivity
In Kubernetes, containers are deployed within units referred to as Pod
, which include one or more containers reachable via a single IP address. With Cilium, each Pod gets an IP address from the node prefix of the Linux node running the Pod. See IP Address Management (IPAM) for additional details. In the absence of any network security policies, all Pods can reach each other.
Pod IP addresses are typically local to the Kubernetes cluster. If pods need to reach services outside the cluster as a client, the network traffic is automatically masqueraded as it leaves the node.
Service Load-balancing
Kubernetes has developed the Services abstraction which provides the user the ability to load balance network traffic to different pods. This abstraction allows the pods reaching out to other pods by a single IP address, a virtual IP address, without knowing all the pods that are running that particular service.
Without Cilium, kube-proxy is installed on every node, watches for endpoints and services addition and removal on the kube-master which allows it to apply the necessary enforcement on iptables. Thus, the received and sent traffic from and to the pods are properly routed to the node and port serving for that service. For more information you can check out the kubernetes user guide for Services.
When implementing ClusterIP, Cilium acts on the same principles as kube-proxy, it watches for services addition or removal, but instead of doing the enforcement on the iptables, it updates eBPF map entries on each node. For more information, see the Pull Request.
Further Reading
The Kubernetes documentation contains more background on the Kubernetes Networking Model and Kubernetes Network Plugins .