64 KiB
cilium
Cilium is open source software for providing and transparently securing network connectivity and loadbalancing between application workloads such as application containers or processes. Cilium operates at Layer 3/4 to provide traditional networking and security services as well as Layer 7 to protect and secure use of modern application protocols such as HTTP, gRPC and Kafka.
A new Linux kernel technology called eBPF is at the foundation of Cilium. It supports dynamic insertion of eBPF bytecode into the Linux kernel at various integration points such as: network IO, application sockets, and tracepoints to implement security, networking and visibility logic. eBPF is highly efficient and flexible.
Prerequisites
- Kubernetes:
>= 1.16.0-0
- Helm:
>= 3.0
Getting Started
Try Cilium on any Kubernetes distribution in under 15 minutes:
Minikube | Self-Managed K8s | Amazon EKS | Google GKE | Microsoft AKS |
---|---|---|---|---|
Or, for a quick install with the default configuration:
$ helm repo add cilium https://helm.cilium.io/
$ helm install cilium cilium/cilium --namespace=kube-system
After Cilium is installed, you can explore the features that Cilium has to offer from the Getting Started Guides page.
Source Code
Getting Help
The best way to get help if you get stuck is to ask a question on the Cilium Slack channel. With Cilium contributors across the globe, there is almost always someone available to help.
Values
Key | Type | Default | Description |
---|---|---|---|
affinity | object | {"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}} |
Affinity for cilium-agent. |
agent | bool | true |
Install the cilium agent resources. |
agentNotReadyTaintKey | string | "node.cilium.io/agent-not-ready" |
Configure the key of the taint indicating that Cilium is not ready on the node. When set to a value starting with ignore-taint.cluster-autoscaler.kubernetes.io/ , the Cluster Autoscaler will ignore the taint on its decisions, allowing the cluster to scale up. |
aksbyocni.enabled | bool | false |
Enable AKS BYOCNI integration. Note that this is incompatible with AKS clusters not created in BYOCNI mode: use Azure integration (azure.enabled ) instead. |
alibabacloud.enabled | bool | false |
Enable AlibabaCloud ENI integration |
annotateK8sNode | bool | false |
Annotate k8s node upon initialization with Cilium's metadata. |
autoDirectNodeRoutes | bool | false |
Enable installation of PodCIDR routes between worker nodes if worker nodes share a common L2 network segment. |
azure.enabled | bool | false |
Enable Azure integration. Note that this is incompatible with AKS clusters created in BYOCNI mode: use AKS BYOCNI integration (aksbyocni.enabled ) instead. |
bandwidthManager | object | {"bbr":false,"enabled":false} |
Enable bandwidth manager to optimize TCP and UDP workloads and allow for rate-limiting traffic from individual Pods with EDT (Earliest Departure Time) through the "kubernetes.io/egress-bandwidth" Pod annotation. |
bandwidthManager.bbr | bool | false |
Activate BBR TCP congestion control for Pods |
bandwidthManager.enabled | bool | false |
Enable bandwidth manager infrastructure (also prerequirement for BBR) |
bgp | object | {"announce":{"loadbalancerIP":false,"podCIDR":false},"enabled":false} |
Configure BGP |
bgp.announce.loadbalancerIP | bool | false |
Enable allocation and announcement of service LoadBalancer IPs |
bgp.announce.podCIDR | bool | false |
Enable announcement of node pod CIDR |
bgp.enabled | bool | false |
Enable BGP support inside Cilium; embeds a new ConfigMap for BGP inside cilium-agent and cilium-operator |
bgpControlPlane | object | {"enabled":false} |
This feature set enables virtual BGP routers to be created via CiliumBGPPeeringPolicy CRDs. |
bgpControlPlane.enabled | bool | false |
Enables the BGP control plane. |
bpf.clockProbe | bool | false |
Enable BPF clock source probing for more efficient tick retrieval. |
bpf.lbExternalClusterIP | bool | false |
Allow cluster external access to ClusterIP services. |
bpf.lbMapMax | int | 65536 |
Configure the maximum number of service entries in the load balancer maps. |
bpf.monitorAggregation | string | "medium" |
Configure the level of aggregation for monitor notifications. Valid options are none, low, medium, maximum. |
bpf.monitorFlags | string | "all" |
Configure which TCP flags trigger notifications when seen for the first time in a connection. |
bpf.monitorInterval | string | "5s" |
Configure the typical time between monitor notifications for active connections. |
bpf.policyMapMax | int | 16384 |
Configure the maximum number of entries in endpoint policy map (per endpoint). |
bpf.preallocateMaps | bool | false |
Enables pre-allocation of eBPF map values. This increases memory usage but can reduce latency. |
bpf.root | string | "/sys/fs/bpf" |
Configure the mount point for the BPF filesystem |
certgen | object | {"image":{"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/certgen","tag":"v0.1.8@sha256:4a456552a5f192992a6edcec2febb1c54870d665173a33dc7d876129b199ddbd"},"podLabels":{},"tolerations":[],"ttlSecondsAfterFinished":1800} |
Configure certificate generation for Hubble integration. If hubble.tls.auto.method=cronJob, these values are used for the Kubernetes CronJob which will be scheduled regularly to (re)generate any certificates not provided manually. |
certgen.podLabels | object | {} |
Labels to be added to hubble-certgen pods |
certgen.tolerations | list | [] |
Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
certgen.ttlSecondsAfterFinished | int | 1800 |
Seconds after which the completed job pod will be deleted |
cgroup | object | {"autoMount":{"enabled":true},"hostRoot":"/run/cilium/cgroupv2"} |
Configure cgroup related configuration |
cgroup.autoMount.enabled | bool | true |
Enable auto mount of cgroup2 filesystem. When autoMount is enabled, cgroup2 filesystem is mounted at cgroup.hostRoot path on the underlying host and inside the cilium agent pod. If users disable autoMount , it's expected that users have mounted cgroup2 filesystem at the specified cgroup.hostRoot volume, and then the volume will be mounted inside the cilium agent pod at the same path. |
cgroup.hostRoot | string | "/run/cilium/cgroupv2" |
Configure cgroup root where cgroup2 filesystem is mounted on the host (see also: cgroup.autoMount ) |
cleanBpfState | bool | false |
Clean all eBPF datapath state from the initContainer of the cilium-agent DaemonSet. WARNING: Use with care! |
cleanState | bool | false |
Clean all local Cilium state from the initContainer of the cilium-agent DaemonSet. Implies cleanBpfState: true. WARNING: Use with care! |
cluster.id | int | 0 |
(int) Unique ID of the cluster. Must be unique across all connected clusters and in the range of 1 to 255. Only required for Cluster Mesh, may be 0 if Cluster Mesh is not used. |
cluster.name | string | "default" |
Name of the cluster. Only required for Cluster Mesh. |
clustermesh.apiserver.affinity | object | {"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"clustermesh-apiserver"}},"topologyKey":"kubernetes.io/hostname"}]}} |
Affinity for clustermesh.apiserver |
clustermesh.apiserver.etcd.image | object | {"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/coreos/etcd","tag":"v3.5.4@sha256:795d8660c48c439a7c3764c2330ed9222ab5db5bb524d8d0607cac76f7ba82a3"} |
Clustermesh API server etcd image. |
clustermesh.apiserver.extraEnv | list | [] |
Additional clustermesh-apiserver environment variables. |
clustermesh.apiserver.image | object | {"digest":"sha256:3f5a6298bd70a2b555c88e291eec1583a6478c3e2272e3fc721aa03b3300d299","override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/clustermesh-apiserver","tag":"v1.12.0","useDigest":true} |
Clustermesh API server image. |
clustermesh.apiserver.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
clustermesh.apiserver.podAnnotations | object | {} |
Annotations to be added to clustermesh-apiserver pods |
clustermesh.apiserver.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
clustermesh.apiserver.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
clustermesh.apiserver.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
clustermesh.apiserver.podLabels | object | {} |
Labels to be added to clustermesh-apiserver pods |
clustermesh.apiserver.priorityClassName | string | "" |
The priority class to use for clustermesh-apiserver |
clustermesh.apiserver.replicas | int | 1 |
Number of replicas run for the clustermesh-apiserver deployment. |
clustermesh.apiserver.resources | object | {} |
Resource requests and limits for the clustermesh-apiserver |
clustermesh.apiserver.service.annotations | object | {} |
Annotations for the clustermesh-apiserver For GKE LoadBalancer, use annotation cloud.google.com/load-balancer-type: "Internal" For EKS LoadBalancer, use annotation service.beta.kubernetes.io/aws-load-balancer-internal: 0.0.0.0/0 |
clustermesh.apiserver.service.nodePort | int | 32379 |
Optional port to use as the node port for apiserver access. |
clustermesh.apiserver.service.type | string | "NodePort" |
The type of service used for apiserver access. |
clustermesh.apiserver.tls.admin | object | {"cert":"","key":""} |
base64 encoded PEM values for the clustermesh-apiserver admin certificate and private key. Used if 'auto' is not enabled. |
clustermesh.apiserver.tls.auto | object | {"certManagerIssuerRef":{},"certValidityDuration":1095,"enabled":true,"method":"helm"} |
Configure automatic TLS certificates generation. A Kubernetes CronJob is used the generate any certificates not provided by the user at installation time. |
clustermesh.apiserver.tls.auto.certManagerIssuerRef | object | {} |
certmanager issuer used when clustermesh.apiserver.tls.auto.method=certmanager. If not specified, a CA issuer will be created. |
clustermesh.apiserver.tls.auto.certValidityDuration | int | 1095 |
Generated certificates validity duration in days. |
clustermesh.apiserver.tls.auto.enabled | bool | true |
When set to true, automatically generate a CA and certificates to enable mTLS between clustermesh-apiserver and external workload instances. If set to false, the certs to be provided by setting appropriate values below. |
clustermesh.apiserver.tls.ca | object | {"cert":"","key":""} |
base64 encoded PEM values for the ExternalWorkload CA certificate and private key. |
clustermesh.apiserver.tls.ca.cert | string | "" |
Optional CA cert. If it is provided, it will be used by the 'cronJob' method to generate all other certificates. Otherwise, an ephemeral CA is generated. |
clustermesh.apiserver.tls.ca.key | string | "" |
Optional CA private key. If it is provided, it will be used by the 'cronJob' method to generate all other certificates. Otherwise, an ephemeral CA is generated. |
clustermesh.apiserver.tls.client | object | {"cert":"","key":""} |
base64 encoded PEM values for the clustermesh-apiserver client certificate and private key. Used if 'auto' is not enabled. |
clustermesh.apiserver.tls.remote | object | {"cert":"","key":""} |
base64 encoded PEM values for the clustermesh-apiserver remote cluster certificate and private key. Used if 'auto' is not enabled. |
clustermesh.apiserver.tls.server | object | {"cert":"","extraDnsNames":[],"extraIpAddresses":[],"key":""} |
base64 encoded PEM values for the clustermesh-apiserver server certificate and private key. Used if 'auto' is not enabled. |
clustermesh.apiserver.tls.server.extraDnsNames | list | [] |
Extra DNS names added to certificate when it's auto generated |
clustermesh.apiserver.tls.server.extraIpAddresses | list | [] |
Extra IP addresses added to certificate when it's auto generated |
clustermesh.apiserver.tolerations | list | [] |
Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
clustermesh.apiserver.updateStrategy | object | {"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"} |
clustermesh-apiserver update strategy |
clustermesh.config | object | {"clusters":[],"domain":"mesh.cilium.io","enabled":false} |
Clustermesh explicit configuration. |
clustermesh.config.clusters | list | [] |
List of clusters to be peered in the mesh. |
clustermesh.config.domain | string | "mesh.cilium.io" |
Default dns domain for the Clustermesh API servers This is used in the case cluster addresses are not provided and IPs are used. |
clustermesh.config.enabled | bool | false |
Enable the Clustermesh explicit configuration. |
clustermesh.useAPIServer | bool | false |
Deploy clustermesh-apiserver for clustermesh |
cni.binPath | string | "/opt/cni/bin" |
Configure the path to the CNI binary directory on the host. |
cni.chainingMode | string | "none" |
Configure chaining on top of other CNI plugins. Possible values: - none - aws-cni - flannel - generic-veth - portmap |
cni.confFileMountPath | string | "/tmp/cni-configuration" |
Configure the path to where to mount the ConfigMap inside the agent pod. |
cni.confPath | string | "/etc/cni/net.d" |
Configure the path to the CNI configuration directory on the host. |
cni.configMapKey | string | "cni-config" |
Configure the key in the CNI ConfigMap to read the contents of the CNI configuration from. |
cni.customConf | bool | false |
Skip writing of the CNI configuration. This can be used if writing of the CNI configuration is performed by external automation. |
cni.exclusive | bool | true |
Make Cilium take ownership over the /etc/cni/net.d directory on the node, renaming all non-Cilium CNI configurations to *.cilium_bak . This ensures no Pods can be scheduled using other CNI plugins during Cilium agent downtime. |
cni.hostConfDirMountPath | string | "/host/etc/cni/net.d" |
Configure the path to where the CNI configuration directory is mounted inside the agent pod. |
cni.install | bool | true |
Install the CNI configuration and binary files into the filesystem. |
cni.logFile | string | "/var/run/cilium/cilium-cni.log" |
Configure the log file for CNI logging with retention policy of 7 days. Disable CNI file logging by setting this field to empty explicitly. |
containerRuntime | object | {"integration":"none"} |
Configure container runtime specific integration. |
containerRuntime.integration | string | "none" |
Enables specific integrations for container runtimes. Supported values: - containerd - crio - docker - none - auto (automatically detect the container runtime) |
customCalls | object | {"enabled":false} |
Tail call hooks for custom eBPF programs. |
customCalls.enabled | bool | false |
Enable tail call hooks for custom eBPF programs. |
daemon.runPath | string | "/var/run/cilium" |
Configure where Cilium runtime state should be stored. |
datapathMode | string | "veth" |
Configure which datapath mode should be used for configuring container connectivity. Valid options are "veth" or "ipvlan". Deprecated, to be removed in v1.12. |
debug.enabled | bool | false |
Enable debug logging |
disableEndpointCRD | string | "false" |
Disable the usage of CiliumEndpoint CRD. |
dnsPolicy | string | "" |
DNS policy for Cilium agent pods. Ref: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy |
dnsProxy.dnsRejectResponseCode | string | "refused" |
DNS response code for rejecting DNS requests, available options are '[nameError refused]'. |
dnsProxy.enableDnsCompression | bool | true |
Allow the DNS proxy to compress responses to endpoints that are larger than 512 Bytes or the EDNS0 option, if present. |
dnsProxy.endpointMaxIpPerHostname | int | 50 |
Maximum number of IPs to maintain per FQDN name for each endpoint. |
dnsProxy.idleConnectionGracePeriod | string | "0s" |
Time during which idle but previously active connections with expired DNS lookups are still considered alive. |
dnsProxy.maxDeferredConnectionDeletes | int | 10000 |
Maximum number of IPs to retain for expired DNS lookups with still-active connections. |
dnsProxy.minTtl | int | 3600 |
The minimum time, in seconds, to use DNS data for toFQDNs policies. |
dnsProxy.preCache | string | "" |
DNS cache data at this path is preloaded on agent startup. |
dnsProxy.proxyPort | int | 0 |
Global port on which the in-agent DNS proxy should listen. Default 0 is a OS-assigned port. |
dnsProxy.proxyResponseMaxDelay | string | "100ms" |
The maximum time the DNS proxy holds an allowed DNS response before sending it along. Responses are sent as soon as the datapath is updated with the new IP information. |
egressGateway | object | {"enabled":false,"installRoutes":false} |
Enables egress gateway to redirect and SNAT the traffic that leaves the cluster. |
egressGateway.installRoutes | bool | false |
Install egress gateway IP rules and routes in order to properly steer egress gateway traffic to the correct ENI interface |
enableCiliumEndpointSlice | bool | false |
Enable CiliumEndpointSlice feature. |
enableCnpStatusUpdates | bool | false |
Whether to enable CNP status updates. |
enableCriticalPriorityClass | bool | true |
Explicitly enable or disable priority class. .Capabilities.KubeVersion is unsettable in helm template calls, it depends on k8s libraries version that Helm was compiled against. This option allows to explicitly disable setting the priority class, which is useful for rendering charts for gke clusters in advance. |
enableIPv4Masquerade | bool | true |
Enables masquerading of IPv4 traffic leaving the node from endpoints. |
enableIPv6Masquerade | bool | true |
Enables masquerading of IPv6 traffic leaving the node from endpoints. |
enableK8sEventHandover | bool | false |
Configures the use of the KVStore to optimize Kubernetes event handling by mirroring it into the KVstore for reduced overhead in large clusters. |
enableK8sTerminatingEndpoint | bool | true |
Configure whether to enable auto detect of terminating state for endpoints in order to support graceful termination. |
enableRuntimeDeviceDetection | bool | false |
Enables experimental support for the detection of new and removed datapath devices. When devices change the eBPF datapath is reloaded and services updated. If "devices" is set then only those devices, or devices matching a wildcard will be considered. |
enableXTSocketFallback | bool | true |
Enables the fallback compatibility solution for when the xt_socket kernel module is missing and it is needed for the datapath L7 redirection to work properly. See documentation for details on when this can be disabled: https://docs.cilium.io/en/stable/operations/system_requirements/#linux-kernel. |
encryption.enabled | bool | false |
Enable transparent network encryption. |
encryption.interface | string | "" |
Deprecated in favor of encryption.ipsec.interface. The interface to use for encrypted traffic. This option is only effective when encryption.type is set to ipsec. |
encryption.ipsec.interface | string | "" |
The interface to use for encrypted traffic. |
encryption.ipsec.keyFile | string | "" |
Name of the key file inside the Kubernetes secret configured via secretName. |
encryption.ipsec.mountPath | string | "" |
Path to mount the secret inside the Cilium pod. |
encryption.ipsec.secretName | string | "" |
Name of the Kubernetes secret containing the encryption keys. |
encryption.keyFile | string | "keys" |
Deprecated in favor of encryption.ipsec.keyFile. Name of the key file inside the Kubernetes secret configured via secretName. This option is only effective when encryption.type is set to ipsec. |
encryption.mountPath | string | "/etc/ipsec" |
Deprecated in favor of encryption.ipsec.mountPath. Path to mount the secret inside the Cilium pod. This option is only effective when encryption.type is set to ipsec. |
encryption.nodeEncryption | bool | false |
Enable encryption for pure node to node traffic. This option is only effective when encryption.type is set to ipsec. |
encryption.secretName | string | "cilium-ipsec-keys" |
Deprecated in favor of encryption.ipsec.secretName. Name of the Kubernetes secret containing the encryption keys. This option is only effective when encryption.type is set to ipsec. |
encryption.type | string | "ipsec" |
Encryption method. Can be either ipsec or wireguard. |
encryption.wireguard.userspaceFallback | bool | false |
Enables the fallback to the user-space implementation. |
endpointHealthChecking.enabled | bool | true |
Enable connectivity health checking between virtual endpoints. |
endpointRoutes.enabled | bool | false |
Enable use of per endpoint routes instead of routing via the cilium_host interface. |
endpointStatus | object | {"enabled":false,"status":""} |
Enable endpoint status. Status can be: policy, health, controllers, logs and / or state. For 2 or more options use a comma. |
eni.awsEnablePrefixDelegation | bool | false |
Enable ENI prefix delegation |
eni.awsReleaseExcessIPs | bool | false |
Release IPs not used from the ENI |
eni.ec2APIEndpoint | string | "" |
EC2 API endpoint to use |
eni.enabled | bool | false |
Enable Elastic Network Interface (ENI) integration. |
eni.eniTags | object | {} |
Tags to apply to the newly created ENIs |
eni.iamRole | string | "" |
If using IAM role for Service Accounts will not try to inject identity values from cilium-aws kubernetes secret. Adds annotation to service account if managed by Helm. See https://github.com/aws/amazon-eks-pod-identity-webhook |
eni.instanceTagsFilter | string | "" |
Filter via AWS EC2 Instance tags (k=v) which will dictate which AWS EC2 Instances are going to be used to create new ENIs |
eni.subnetIDsFilter | string | "" |
Filter via subnet IDs which will dictate which subnets are going to be used to create new ENIs Important note: This requires that each instance has an ENI with a matching subnet attached when Cilium is deployed. If you only want to control subnets for ENIs attached by Cilium, use the CNI configuration file settings (cni.customConf) instead. |
eni.subnetTagsFilter | string | "" |
Filter via tags (k=v) which will dictate which subnets are going to be used to create new ENIs Important note: This requires that each instance has an ENI with a matching subnet attached when Cilium is deployed. If you only want to control subnets for ENIs attached by Cilium, use the CNI configuration file settings (cni.customConf) instead. |
eni.updateEC2AdapterLimitViaAPI | bool | false |
Update ENI Adapter limits from the EC2 API |
etcd.clusterDomain | string | "cluster.local" |
Cluster domain for cilium-etcd-operator. |
etcd.enabled | bool | false |
Enable etcd mode for the agent. |
etcd.endpoints | list | ["https://CHANGE-ME:2379"] |
List of etcd endpoints (not needed when using managed=true). |
etcd.extraArgs | list | [] |
Additional cilium-etcd-operator container arguments. |
etcd.image | object | {"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/cilium-etcd-operator","tag":"v2.0.7@sha256:04b8327f7f992693c2cb483b999041ed8f92efc8e14f2a5f3ab95574a65ea2dc"} |
cilium-etcd-operator image. |
etcd.k8sService | bool | false |
If etcd is behind a k8s service set this option to true so that Cilium does the service translation automatically without requiring a DNS to be running. |
etcd.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for cilium-etcd-operator pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
etcd.podAnnotations | object | {} |
Annotations to be added to cilium-etcd-operator pods |
etcd.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
etcd.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
etcd.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
etcd.podLabels | object | {} |
Labels to be added to cilium-etcd-operator pods |
etcd.priorityClassName | string | "" |
The priority class to use for cilium-etcd-operator |
etcd.resources | object | {} |
cilium-etcd-operator resource limits & requests ref: https://kubernetes.io/docs/user-guide/compute-resources/ |
etcd.securityContext | object | {} |
Security context to be added to cilium-etcd-operator pods |
etcd.ssl | bool | false |
Enable use of TLS/SSL for connectivity to etcd. (auto-enabled if managed=true) |
etcd.tolerations | list | [{"operator":"Exists"}] |
Node tolerations for cilium-etcd-operator scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
etcd.updateStrategy | object | {"rollingUpdate":{"maxSurge":1,"maxUnavailable":1},"type":"RollingUpdate"} |
cilium-etcd-operator update strategy |
externalIPs.enabled | bool | false |
Enable ExternalIPs service support. |
externalWorkloads | object | {"enabled":false} |
Configure external workloads support |
externalWorkloads.enabled | bool | false |
Enable support for external workloads, such as VMs (false by default). |
extraArgs | list | [] |
Additional agent container arguments. |
extraConfig | object | {} |
extraConfig allows you to specify additional configuration parameters to be included in the cilium-config configmap. |
extraEnv | list | [] |
Additional agent container environment variables. |
extraHostPathMounts | list | [] |
Additional agent hostPath mounts. |
extraVolumeMounts | list | [] |
Additional agent volumeMounts. |
extraVolumes | list | [] |
Additional agent volumes. |
gke.enabled | bool | false |
Enable Google Kubernetes Engine integration |
healthChecking | bool | true |
Enable connectivity health checking. |
healthPort | int | 9879 |
TCP port for the agent health API. This is not the port for cilium-health. |
hostFirewall | object | {"enabled":false} |
Configure the host firewall. |
hostFirewall.enabled | bool | false |
Enables the enforcement of host policies in the eBPF datapath. |
hostPort.enabled | bool | false |
Enable hostPort service support. |
hubble.enabled | bool | true |
Enable Hubble (true by default). |
hubble.listenAddress | string | ":4244" |
An additional address for Hubble to listen to. Set this field ":4244" if you are enabling Hubble Relay, as it assumes that Hubble is listening on port 4244. |
hubble.metrics | object | {"enabled":null,"port":9965,"serviceAnnotations":{},"serviceMonitor":{"annotations":{},"enabled":false,"labels":{}}} |
Hubble metrics configuration. See https://docs.cilium.io/en/stable/operations/metrics/#hubble-metrics for more comprehensive documentation about Hubble metrics. |
hubble.metrics.enabled | string | nil |
Configures the list of metrics to collect. If empty or null, metrics are disabled. Example: enabled: - dns:query;ignoreAAAA - drop - tcp - flow - icmp - http You can specify the list of metrics from the helm CLI: --set metrics.enabled="{dns:query;ignoreAAAA,drop,tcp,flow,icmp,http}" |
hubble.metrics.port | int | 9965 |
Configure the port the hubble metric server listens on. |
hubble.metrics.serviceAnnotations | object | {} |
Annotations to be added to hubble-metrics service. |
hubble.metrics.serviceMonitor.annotations | object | {} |
Annotations to add to ServiceMonitor hubble |
hubble.metrics.serviceMonitor.enabled | bool | false |
Create ServiceMonitor resources for Prometheus Operator. This requires the prometheus CRDs to be available. ref: https://github.com/prometheus-operator/prometheus-operator/blob/master/example/prometheus-operator-crd/monitoring.coreos.com_servicemonitors.yaml) |
hubble.metrics.serviceMonitor.labels | object | {} |
Labels to add to ServiceMonitor hubble |
hubble.peerService.clusterDomain | string | "cluster.local" |
The cluster domain to use to query the Hubble Peer service. It should be the local cluster. |
hubble.peerService.enabled | bool | true |
Enable a K8s Service for the Peer service, so that it can be accessed by a non-local client |
hubble.peerService.targetPort | int | 4244 |
Target Port for the Peer service. |
hubble.relay.affinity | object | {"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}} |
Affinity for hubble-replay |
hubble.relay.dialTimeout | string | nil |
Dial timeout to connect to the local hubble instance to receive peer information (e.g. "30s"). |
hubble.relay.enabled | bool | false |
Enable Hubble Relay (requires hubble.enabled=true) |
hubble.relay.extraEnv | list | [] |
Additional hubble-relay environment variables. |
hubble.relay.image | object | {"digest":"sha256:ca8033ea8a3112d838f958862fa76c8d895e3c8d0f5590de849b91745af5ac4d","override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/hubble-relay","tag":"v1.12.0","useDigest":true} |
Hubble-relay container image. |
hubble.relay.listenHost | string | "" |
Host to listen to. Specify an empty string to bind to all the interfaces. |
hubble.relay.listenPort | string | "4245" |
Port to listen to. |
hubble.relay.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
hubble.relay.podAnnotations | object | {} |
Annotations to be added to hubble-relay pods |
hubble.relay.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
hubble.relay.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
hubble.relay.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
hubble.relay.podLabels | object | {} |
Labels to be added to hubble-relay pods |
hubble.relay.priorityClassName | string | "" |
The priority class to use for hubble-relay |
hubble.relay.prometheus | object | {"enabled":false,"port":9966,"serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","labels":{}}} |
Enable prometheus metrics for hubble-relay on the configured port at /metrics |
hubble.relay.prometheus.serviceMonitor.annotations | object | {} |
Annotations to add to ServiceMonitor hubble-relay |
hubble.relay.prometheus.serviceMonitor.enabled | bool | false |
Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/master/example/prometheus-operator-crd/monitoring.coreos.com_servicemonitors.yaml) |
hubble.relay.prometheus.serviceMonitor.interval | string | "10s" |
Interval for scrape metrics. |
hubble.relay.prometheus.serviceMonitor.labels | object | {} |
Labels to add to ServiceMonitor hubble-relay |
hubble.relay.replicas | int | 1 |
Number of replicas run for the hubble-relay deployment. |
hubble.relay.resources | object | {} |
Specifies the resources for the hubble-relay pods |
hubble.relay.retryTimeout | string | nil |
Backoff duration to retry connecting to the local hubble instance in case of failure (e.g. "30s"). |
hubble.relay.rollOutPods | bool | false |
Roll out Hubble Relay pods automatically when configmap is updated. |
hubble.relay.securityContext | object | {} |
hubble-relay security context |
hubble.relay.service | object | {"nodePort":31234,"type":"ClusterIP"} |
hubble-relay service configuration. |
hubble.relay.service.nodePort | int | 31234 |
- The port to use when the service type is set to NodePort. |
hubble.relay.service.type | string | "ClusterIP" |
- The type of service used for Hubble Relay access, either ClusterIP or NodePort. |
hubble.relay.sortBufferDrainTimeout | string | nil |
When the per-request flows sort buffer is not full, a flow is drained every time this timeout is reached (only affects requests in follow-mode) (e.g. "1s"). |
hubble.relay.sortBufferLenMax | string | nil |
Max number of flows that can be buffered for sorting before being sent to the client (per request) (e.g. 100). |
hubble.relay.terminationGracePeriodSeconds | int | 1 |
Configure termination grace period for hubble relay Deployment. |
hubble.relay.tls | object | {"client":{"cert":"","key":""},"server":{"cert":"","enabled":false,"extraDnsNames":[],"extraIpAddresses":[],"key":""}} |
TLS configuration for Hubble Relay |
hubble.relay.tls.client | object | {"cert":"","key":""} |
base64 encoded PEM values for the hubble-relay client certificate and private key This keypair is presented to Hubble server instances for mTLS authentication and is required when hubble.tls.enabled is true. These values need to be set manually if hubble.tls.auto.enabled is false. |
hubble.relay.tls.server | object | {"cert":"","enabled":false,"extraDnsNames":[],"extraIpAddresses":[],"key":""} |
base64 encoded PEM values for the hubble-relay server certificate and private key |
hubble.relay.tls.server.extraDnsNames | list | [] |
extra DNS names added to certificate when its auto gen |
hubble.relay.tls.server.extraIpAddresses | list | [] |
extra IP addresses added to certificate when its auto gen |
hubble.relay.tolerations | list | [] |
Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
hubble.relay.updateStrategy | object | {"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"} |
hubble-relay update strategy |
hubble.socketPath | string | "/var/run/cilium/hubble.sock" |
Unix domain socket path to listen to when Hubble is enabled. |
hubble.tls | object | {"auto":{"certManagerIssuerRef":{},"certValidityDuration":1095,"enabled":true,"method":"helm","schedule":"0 0 1 */4 *"},"ca":{"cert":"","key":""},"enabled":true,"server":{"cert":"","extraDnsNames":[],"extraIpAddresses":[],"key":""}} |
TLS configuration for Hubble |
hubble.tls.auto | object | {"certManagerIssuerRef":{},"certValidityDuration":1095,"enabled":true,"method":"helm","schedule":"0 0 1 */4 *"} |
Configure automatic TLS certificates generation. |
hubble.tls.auto.certManagerIssuerRef | object | {} |
certmanager issuer used when hubble.tls.auto.method=certmanager. If not specified, a CA issuer will be created. |
hubble.tls.auto.certValidityDuration | int | 1095 |
Generated certificates validity duration in days. |
hubble.tls.auto.enabled | bool | true |
Auto-generate certificates. When set to true, automatically generate a CA and certificates to enable mTLS between Hubble server and Hubble Relay instances. If set to false, the certs for Hubble server need to be provided by setting appropriate values below. |
hubble.tls.auto.method | string | "helm" |
Set the method to auto-generate certificates. Supported values: - helm: This method uses Helm to generate all certificates. - cronJob: This method uses a Kubernetes CronJob the generate any certificates not provided by the user at installation time. - certmanager: This method use cert-manager to generate & rotate certificates. |
hubble.tls.auto.schedule | string | "0 0 1 */4 *" |
Schedule for certificates regeneration (regardless of their expiration date). Only used if method is "cronJob". If nil, then no recurring job will be created. Instead, only the one-shot job is deployed to generate the certificates at installation time. Defaults to midnight of the first day of every fourth month. For syntax, see https://kubernetes.io/docs/tasks/job/automated-tasks-with-cron-jobs/#schedule |
hubble.tls.ca | object | {"cert":"","key":""} |
Deprecated in favor of tls.ca. To be removed in 1.13. base64 encoded PEM values for the Hubble CA certificate and private key. |
hubble.tls.ca.cert | string | "" |
Deprecated in favor of tls.ca.cert. To be removed in 1.13. |
hubble.tls.ca.key | string | "" |
Deprecated in favor of tls.ca.key. To be removed in 1.13. The CA private key (optional). If it is provided, then it will be used by hubble.tls.auto.method=cronJob to generate all other certificates. Otherwise, a ephemeral CA is generated if hubble.tls.auto.enabled=true. |
hubble.tls.enabled | bool | true |
Enable mutual TLS for listenAddress. Setting this value to false is highly discouraged as the Hubble API provides access to potentially sensitive network flow metadata and is exposed on the host network. |
hubble.tls.server | object | {"cert":"","extraDnsNames":[],"extraIpAddresses":[],"key":""} |
base64 encoded PEM values for the Hubble server certificate and private key |
hubble.tls.server.extraDnsNames | list | [] |
Extra DNS names added to certificate when it's auto generated |
hubble.tls.server.extraIpAddresses | list | [] |
Extra IP addresses added to certificate when it's auto generated |
hubble.ui.affinity | object | {} |
Affinity for hubble-ui |
hubble.ui.backend.extraEnv | list | [] |
Additional hubble-ui backend environment variables. |
hubble.ui.backend.image | object | {"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/hubble-ui-backend","tag":"v0.9.0@sha256:000df6b76719f607a9edefb9af94dfd1811a6f1b6a8a9c537cba90bf12df474b"} |
Hubble-ui backend image. |
hubble.ui.backend.resources | object | {} |
Resource requests and limits for the 'backend' container of the 'hubble-ui' deployment. |
hubble.ui.enabled | bool | false |
Whether to enable the Hubble UI. |
hubble.ui.frontend.extraEnv | list | [] |
Additional hubble-ui frontend environment variables. |
hubble.ui.frontend.image | object | {"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/hubble-ui","tag":"v0.9.0@sha256:0ef04e9a29212925da6bdfd0ba5b581765e41a01f1cc30563cef9b30b457fea0"} |
Hubble-ui frontend image. |
hubble.ui.frontend.resources | object | {} |
Resource requests and limits for the 'frontend' container of the 'hubble-ui' deployment. |
hubble.ui.ingress | object | {"annotations":{},"className":"","enabled":false,"hosts":["chart-example.local"],"tls":[]} |
hubble-ui ingress configuration. |
hubble.ui.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
hubble.ui.podAnnotations | object | {} |
Annotations to be added to hubble-ui pods |
hubble.ui.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
hubble.ui.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
hubble.ui.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
hubble.ui.podLabels | object | {} |
Labels to be added to hubble-ui pods |
hubble.ui.priorityClassName | string | "" |
The priority class to use for hubble-ui |
hubble.ui.replicas | int | 1 |
The number of replicas of Hubble UI to deploy. |
hubble.ui.rollOutPods | bool | false |
Roll out Hubble-ui pods automatically when configmap is updated. |
hubble.ui.securityContext | object | {"enabled":true,"fsGroup":1001,"runAsGroup":1001,"runAsUser":1001} |
Security context to be added to Hubble UI pods |
hubble.ui.securityContext.enabled | bool | true |
Deprecated in favor of hubble.ui.securityContext. Whether to set the security context on the Hubble UI pods. |
hubble.ui.service | object | {"nodePort":31235,"type":"ClusterIP"} |
hubble-ui service configuration. |
hubble.ui.service.nodePort | int | 31235 |
- The port to use when the service type is set to NodePort. |
hubble.ui.service.type | string | "ClusterIP" |
- The type of service used for Hubble UI access, either ClusterIP or NodePort. |
hubble.ui.standalone.enabled | bool | false |
When true, it will allow installing the Hubble UI only, without checking dependencies. It is useful if a cluster already has cilium and Hubble relay installed and you just want Hubble UI to be deployed. When installed via helm, installing UI should be done via helm upgrade and when installed via the cilium cli, then cilium hubble enable --ui |
hubble.ui.standalone.tls.certsVolume | object | {} |
When deploying Hubble UI in standalone, with tls enabled for Hubble relay, it is required to provide a volume for mounting the client certificates. |
hubble.ui.tls.client | object | {"cert":"","key":""} |
base64 encoded PEM values used to connect to hubble-relay This keypair is presented to Hubble Relay instances for mTLS authentication and is required when hubble.relay.tls.server.enabled is true. These values need to be set manually if hubble.tls.auto.enabled is false. |
hubble.ui.tolerations | list | [] |
Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
hubble.ui.updateStrategy | object | {"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"} |
hubble-ui update strategy. |
identityAllocationMode | string | "crd" |
Method to use for identity allocation (crd or kvstore ). |
image | object | {"digest":"sha256:079baa4fa1b9fe638f96084f4e0297c84dd4fb215d29d2321dcbe54273f63ade","override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/cilium","tag":"v1.12.0","useDigest":true} |
Agent container image. |
imagePullSecrets | string | nil |
Configure image pull secrets for pulling container images |
ingressController.enabled | bool | false |
Enable cilium ingress controller This will automatically set enable-envoy-config as well. |
ingressController.enforceHttps | bool | true |
Enforce https for host having matching TLS host in Ingress. Incoming traffic to http listener will return 308 http error code with respective location in header. |
ingressController.secretsNamespace | object | {"create":true,"name":"cilium-secrets","sync":true} |
SecretsNamespace is the namespace in which envoy SDS will retrieve TLS secrets from. |
ingressController.secretsNamespace.create | bool | true |
Create secrets namespace for Ingress. |
ingressController.secretsNamespace.name | string | "cilium-secrets" |
Name of Ingress secret namespace. |
ingressController.secretsNamespace.sync | bool | true |
Enable secret sync, which will make sure all TLS secrets used by Ingress are synced to secretsNamespace.name. If disabled, TLS secrets must be maintained externally. |
installIptablesRules | bool | true |
Configure whether to install iptables rules to allow for TPROXY (L7 proxy injection), iptables-based masquerading and compatibility with kube-proxy. |
installNoConntrackIptablesRules | bool | false |
Install Iptables rules to skip netfilter connection tracking on all pod traffic. This option is only effective when Cilium is running in direct routing and full KPR mode. Moreover, this option cannot be enabled when Cilium is running in a managed Kubernetes environment or in a chained CNI setup. |
ipMasqAgent | object | {"enabled":false} |
Configure the eBPF-based ip-masq-agent |
ipam.mode | string | "cluster-pool" |
Configure IP Address Management mode. ref: https://docs.cilium.io/en/stable/concepts/networking/ipam/ |
ipam.operator.clusterPoolIPv4MaskSize | int | 24 |
IPv4 CIDR mask size to delegate to individual nodes for IPAM. |
ipam.operator.clusterPoolIPv4PodCIDR | string | "10.0.0.0/8" |
Deprecated in favor of ipam.operator.clusterPoolIPv4PodCIDRList. IPv4 CIDR range to delegate to individual nodes for IPAM. |
ipam.operator.clusterPoolIPv4PodCIDRList | list | [] |
IPv4 CIDR list range to delegate to individual nodes for IPAM. |
ipam.operator.clusterPoolIPv6MaskSize | int | 120 |
IPv6 CIDR mask size to delegate to individual nodes for IPAM. |
ipam.operator.clusterPoolIPv6PodCIDR | string | "fd00::/104" |
Deprecated in favor of ipam.operator.clusterPoolIPv6PodCIDRList. IPv6 CIDR range to delegate to individual nodes for IPAM. |
ipam.operator.clusterPoolIPv6PodCIDRList | list | [] |
IPv6 CIDR list range to delegate to individual nodes for IPAM. |
ipv4.enabled | bool | true |
Enable IPv4 support. |
ipv6.enabled | bool | false |
Enable IPv6 support. |
ipvlan.enabled | bool | false |
Enable the IPVLAN datapath (deprecated) |
k8s | object | {} |
Configure Kubernetes specific configuration |
keepDeprecatedLabels | bool | false |
Keep the deprecated selector labels when deploying Cilium DaemonSet. |
keepDeprecatedProbes | bool | false |
Keep the deprecated probes when deploying Cilium DaemonSet |
kubeProxyReplacementHealthzBindAddr | string | "" |
healthz server bind address for the kube-proxy replacement. To enable set the value to '0.0.0.0:10256' for all ipv4 addresses and this '[::]:10256' for all ipv6 addresses. By default it is disabled. |
l2NeighDiscovery.enabled | bool | true |
Enable L2 neighbor discovery in the agent |
l2NeighDiscovery.refreshPeriod | string | "30s" |
Override the agent's default neighbor resolution refresh period. |
l7Proxy | bool | true |
Enable Layer 7 network policy. |
livenessProbe.failureThreshold | int | 10 |
failure threshold of liveness probe |
livenessProbe.periodSeconds | int | 30 |
interval between checks of the liveness probe |
localRedirectPolicy | bool | false |
Enable Local Redirect Policy. |
logSystemLoad | bool | false |
Enables periodic logging of system load |
maglev | object | {} |
Configure maglev consistent hashing |
monitor | object | {"enabled":false} |
cilium-monitor sidecar. |
monitor.enabled | bool | false |
Enable the cilium-monitor sidecar. |
name | string | "cilium" |
Agent container name. |
nodePort | object | {"autoProtectPortRange":true,"bindProtection":true,"enableHealthCheck":true,"enabled":false} |
Configure N-S k8s service loadbalancing |
nodePort.autoProtectPortRange | bool | true |
Append NodePort range to ip_local_reserved_ports if clash with ephemeral ports is detected. |
nodePort.bindProtection | bool | true |
Set to true to prevent applications binding to service ports. |
nodePort.enableHealthCheck | bool | true |
Enable healthcheck nodePort server for NodePort services |
nodePort.enabled | bool | false |
Enable the Cilium NodePort service implementation. |
nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node selector for cilium-agent. |
nodeinit.affinity | object | {} |
Affinity for cilium-nodeinit |
nodeinit.bootstrapFile | string | "/tmp/cilium-bootstrap.d/cilium-bootstrap-time" |
bootstrapFile is the location of the file where the bootstrap timestamp is written by the node-init DaemonSet |
nodeinit.enabled | bool | false |
Enable the node initialization DaemonSet |
nodeinit.extraEnv | list | [] |
Additional nodeinit environment variables. |
nodeinit.image | object | {"override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/startup-script","tag":"d69851597ea019af980891a4628fb36b7880ec26"} |
node-init image. |
nodeinit.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for nodeinit pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
nodeinit.podAnnotations | object | {} |
Annotations to be added to node-init pods. |
nodeinit.podLabels | object | {} |
Labels to be added to node-init pods. |
nodeinit.priorityClassName | string | "" |
The priority class to use for the nodeinit pod. |
nodeinit.resources | object | {"requests":{"cpu":"100m","memory":"100Mi"}} |
nodeinit resource limits & requests ref: https://kubernetes.io/docs/user-guide/compute-resources/ |
nodeinit.securityContext | object | {"capabilities":{"add":["SYS_MODULE","NET_ADMIN","SYS_ADMIN","SYS_CHROOT","SYS_PTRACE"]},"privileged":false,"seLinuxOptions":{"level":"s0","type":"spc_t"}} |
Security context to be added to nodeinit pods. |
nodeinit.tolerations | list | [{"operator":"Exists"}] |
Node tolerations for nodeinit scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
nodeinit.updateStrategy | object | {"type":"RollingUpdate"} |
node-init update strategy |
operator.affinity | object | {"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"io.cilium/app":"operator"}},"topologyKey":"kubernetes.io/hostname"}]}} |
Affinity for cilium-operator |
operator.dnsPolicy | string | "" |
DNS policy for Cilium operator pods. Ref: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy |
operator.enabled | bool | true |
Enable the cilium-operator component (required). |
operator.endpointGCInterval | string | "5m0s" |
Interval for endpoint garbage collection. |
operator.extraArgs | list | [] |
Additional cilium-operator container arguments. |
operator.extraEnv | list | [] |
Additional cilium-operator environment variables. |
operator.extraHostPathMounts | list | [] |
Additional cilium-operator hostPath mounts. |
operator.extraVolumeMounts | list | [] |
Additional cilium-operator volumeMounts. |
operator.extraVolumes | list | [] |
Additional cilium-operator volumes. |
operator.identityGCInterval | string | "15m0s" |
Interval for identity garbage collection. |
operator.identityHeartbeatTimeout | string | "30m0s" |
Timeout for identity heartbeats. |
operator.image | object | {"alibabacloudDigest":"sha256:93dddf88e92119a141a913b44ab9cb909f19b9a7bf01e30b98c1e8afeec51cd5","awsDigest":"sha256:cb73df18b03b4fc914c80045d0ddb6c9256972449382e3c4b294fd9c371ace22","azureDigest":"sha256:98ffa2c8ebff33d4e91762fb57d4c36f152bb044c4e2141e15362cf95ecc24ba","genericDigest":"sha256:bb2a42eda766e5d4a87ee8a5433f089db81b72dd04acf6b59fcbb445a95f9410","override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/operator","suffix":"","tag":"v1.12.0","useDigest":true} |
cilium-operator image. |
operator.nodeGCInterval | string | "5m0s" |
Interval for cilium node garbage collection. |
operator.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for cilium-operator pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
operator.podAnnotations | object | {} |
Annotations to be added to cilium-operator pods |
operator.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
operator.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
operator.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
operator.podLabels | object | {} |
Labels to be added to cilium-operator pods |
operator.priorityClassName | string | "" |
The priority class to use for cilium-operator |
operator.prometheus | object | {"enabled":false,"port":9963,"serviceMonitor":{"annotations":{},"enabled":false,"labels":{}}} |
Enable prometheus metrics for cilium-operator on the configured port at /metrics |
operator.prometheus.serviceMonitor.annotations | object | {} |
Annotations to add to ServiceMonitor cilium-operator |
operator.prometheus.serviceMonitor.enabled | bool | false |
Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/master/example/prometheus-operator-crd/monitoring.coreos.com_servicemonitors.yaml) |
operator.prometheus.serviceMonitor.labels | object | {} |
Labels to add to ServiceMonitor cilium-operator |
operator.removeNodeTaints | bool | true |
Remove Cilium node taint from Kubernetes nodes that have a healthy Cilium pod running. |
operator.replicas | int | 2 |
Number of replicas to run for the cilium-operator deployment |
operator.resources | object | {} |
cilium-operator resource limits & requests ref: https://kubernetes.io/docs/user-guide/compute-resources/ |
operator.rollOutPods | bool | false |
Roll out cilium-operator pods automatically when configmap is updated. |
operator.securityContext | object | {} |
Security context to be added to cilium-operator pods |
operator.setNodeNetworkStatus | bool | true |
Set Node condition NetworkUnavailable to 'false' with the reason 'CiliumIsUp' for nodes that have a healthy Cilium pod. |
operator.skipCRDCreation | bool | false |
Skip CRDs creation for cilium-operator |
operator.tolerations | list | [{"operator":"Exists"}] |
Node tolerations for cilium-operator scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
operator.unmanagedPodWatcher.intervalSeconds | int | 15 |
Interval, in seconds, to check if there are any pods that are not managed by Cilium. |
operator.unmanagedPodWatcher.restart | bool | true |
Restart any pod that are not managed by Cilium. |
operator.updateStrategy | object | {"rollingUpdate":{"maxSurge":1,"maxUnavailable":1},"type":"RollingUpdate"} |
cilium-operator update strategy |
podAnnotations | object | {} |
Annotations to be added to agent pods |
podLabels | object | {} |
Labels to be added to agent pods |
policyEnforcementMode | string | "default" |
The agent can be put into one of the three policy enforcement modes: default, always and never. ref: https://docs.cilium.io/en/stable/policy/intro/#policy-enforcement-modes |
pprof.enabled | bool | false |
Enable Go pprof debugging |
preflight.affinity | object | {"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}} |
Affinity for cilium-preflight |
preflight.enabled | bool | false |
Enable Cilium pre-flight resources (required for upgrade) |
preflight.extraEnv | list | [] |
Additional preflight environment variables. |
preflight.image | object | {"digest":"sha256:079baa4fa1b9fe638f96084f4e0297c84dd4fb215d29d2321dcbe54273f63ade","override":null,"pullPolicy":"IfNotPresent","repository":"quay.io/cilium/cilium","tag":"v1.12.0","useDigest":true} |
Cilium pre-flight image. |
preflight.nodeSelector | object | {"kubernetes.io/os":"linux"} |
Node labels for preflight pod assignment ref: https://kubernetes.io/docs/user-guide/node-selection/ |
preflight.podAnnotations | object | {} |
Annotations to be added to preflight pods |
preflight.podDisruptionBudget.enabled | bool | false |
enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ |
preflight.podDisruptionBudget.maxUnavailable | int | 1 |
Maximum number/percentage of pods that may be made unavailable |
preflight.podDisruptionBudget.minAvailable | string | nil |
Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by maxUnavailable: null |
preflight.podLabels | object | {} |
Labels to be added to the preflight pod. |
preflight.priorityClassName | string | "" |
The priority class to use for the preflight pod. |
preflight.resources | object | {} |
preflight resource limits & requests ref: https://kubernetes.io/docs/user-guide/compute-resources/ |
preflight.securityContext | object | {} |
Security context to be added to preflight pods |
preflight.terminationGracePeriodSeconds | int | 1 |
Configure termination grace period for preflight Deployment and DaemonSet. |
preflight.tofqdnsPreCache | string | "" |
Path to write the --tofqdns-pre-cache file to. |
preflight.tolerations | list | [{"effect":"NoSchedule","key":"node.kubernetes.io/not-ready"},{"effect":"NoSchedule","key":"node-role.kubernetes.io/master"},{"effect":"NoSchedule","key":"node.cloudprovider.kubernetes.io/uninitialized","value":"true"},{"key":"CriticalAddonsOnly","operator":"Exists"}] |
Node tolerations for preflight scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
preflight.updateStrategy | object | {"type":"RollingUpdate"} |
preflight update strategy |
preflight.validateCNPs | bool | true |
By default we should always validate the installed CNPs before upgrading Cilium. This will make sure the user will have the policies deployed in the cluster with the right schema. |
priorityClassName | string | "" |
The priority class to use for cilium-agent. |
prometheus | object | {"enabled":false,"metrics":null,"port":9962,"serviceMonitor":{"annotations":{},"enabled":false,"labels":{}}} |
Configure prometheus metrics on the configured port at /metrics |
prometheus.metrics | string | nil |
Metrics that should be enabled or disabled from the default metric list. (+metric_foo to enable metric_foo , -metric_bar to disable metric_bar). ref: https://docs.cilium.io/en/stable/operations/metrics/#exported-metrics |
prometheus.serviceMonitor.annotations | object | {} |
Annotations to add to ServiceMonitor cilium-agent |
prometheus.serviceMonitor.enabled | bool | false |
Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/master/example/prometheus-operator-crd/monitoring.coreos.com_servicemonitors.yaml) |
prometheus.serviceMonitor.labels | object | {} |
Labels to add to ServiceMonitor cilium-agent |
proxy | object | {"prometheus":{"enabled":true,"port":"9964"},"sidecarImageRegex":"cilium/istio_proxy"} |
Configure Istio proxy options. |
proxy.sidecarImageRegex | string | "cilium/istio_proxy" |
Regular expression matching compatible Istio sidecar istio-proxy container image names |
rbac.create | bool | true |
Enable creation of Resource-Based Access Control configuration. |
readinessProbe.failureThreshold | int | 3 |
failure threshold of readiness probe |
readinessProbe.periodSeconds | int | 30 |
interval between checks of the readiness probe |
remoteNodeIdentity | bool | true |
Enable use of the remote node identity. ref: https://docs.cilium.io/en/v1.7/install/upgrade/#configmap-remote-node-identity |
resourceQuotas | object | {"cilium":{"hard":{"pods":"10k"}},"enabled":false,"operator":{"hard":{"pods":"15"}}} |
Enable resource quotas for priority classes used in the cluster. |
resources | object | {} |
Agent resource limits & requests ref: https://kubernetes.io/docs/user-guide/compute-resources/ |
rollOutCiliumPods | bool | false |
Roll out cilium agent pods automatically when configmap is updated. |
securityContext | object | {"extraCapabilities":["DAC_OVERRIDE","FOWNER","SETGID","SETUID"],"privileged":false} |
Security context to be added to agent pods |
serviceAccounts | object | Component's fully qualified name. | Define serviceAccount names for components. |
serviceAccounts.clustermeshcertgen | object | {"annotations":{},"create":true,"name":"clustermesh-apiserver-generate-certs"} |
Clustermeshcertgen is used if clustermesh.apiserver.tls.auto.method=cronJob |
serviceAccounts.hubblecertgen | object | {"annotations":{},"create":true,"name":"hubble-generate-certs"} |
Hubblecertgen is used if hubble.tls.auto.method=cronJob |
sleepAfterInit | bool | false |
Do not run Cilium agent when running with clean mode. Useful to completely uninstall Cilium as it will stop Cilium from starting and create artifacts in the node. |
socketLB | object | {"enabled":false} |
Configure socket LB |
socketLB.enabled | bool | false |
Enable socket LB |
sockops | object | {"enabled":false} |
Configure BPF socket operations configuration |
startupProbe.failureThreshold | int | 105 |
failure threshold of startup probe. 105 x 2s translates to the old behaviour of the readiness probe (120s delay + 30 x 3s) |
startupProbe.periodSeconds | int | 2 |
interval between checks of the startup probe |
svcSourceRangeCheck | bool | true |
Enable check of service source ranges (currently, only for LoadBalancer). |
synchronizeK8sNodes | bool | true |
Synchronize Kubernetes nodes to kvstore and perform CNP GC. |
terminationGracePeriodSeconds | int | 1 |
Configure termination grace period for cilium-agent DaemonSet. |
tls | object | {"ca":{"cert":"","certValidityDuration":1095,"key":""},"secretsBackend":"local"} |
Configure TLS configuration in the agent. |
tls.ca | object | {"cert":"","certValidityDuration":1095,"key":""} |
Base64 encoded PEM values for the CA certificate and private key. This can be used as common CA to generate certificates used by hubble and clustermesh components |
tls.ca.cert | string | "" |
Optional CA cert. If it is provided, it will be used by cilium to generate all other certificates. Otherwise, an ephemeral CA is generated. |
tls.ca.certValidityDuration | int | 1095 |
Generated certificates validity duration in days. This will be used for auto generated CA. |
tls.ca.key | string | "" |
Optional CA private key. If it is provided, it will be used by cilium to generate all other certificates. Otherwise, an ephemeral CA is generated. |
tls.secretsBackend | string | "local" |
This configures how the Cilium agent loads the secrets used TLS-aware CiliumNetworkPolicies (namely the secrets referenced by terminatingTLS and originatingTLS). Possible values: - local - k8s |
tolerations | list | [{"operator":"Exists"}] |
Node tolerations for agent scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/ |
tunnel | string | "vxlan" |
Configure the encapsulation configuration for communication between nodes. Possible values: - disabled - vxlan (default) - geneve |
updateStrategy | object | {"rollingUpdate":{"maxUnavailable":2},"type":"RollingUpdate"} |
Cilium agent update strategy |
vtep.cidr | string | "" |
A space separated list of VTEP device CIDRs, for example "1.1.1.0/24 1.1.2.0/24" |
vtep.enabled | bool | false |
Enables VXLAN Tunnel Endpoint (VTEP) Integration (beta) to allow Cilium-managed pods to talk to third party VTEP devices over Cilium tunnel. |
vtep.endpoint | string | "" |
A space separated list of VTEP device endpoint IPs, for example "1.1.1.1 1.1.2.1" |
vtep.mac | string | "" |
A space separated list of VTEP device MAC addresses (VTEP MAC), for example "x❌x❌x:x y:y:y:y:y:y:y" |
vtep.mask | string | "" |
VTEP CIDRs Mask that applies to all VTEP CIDRs, for example "255.255.255.0" |
waitForKubeProxy | bool | false |
Wait for KUBE-PROXY-CANARY iptables rule to appear in "wait-for-kube-proxy" init container before launching cilium-agent. More context can be found in the commit message of below PR https://github.com/cilium/cilium/pull/20123 |
wellKnownIdentities.enabled | bool | false |
Enable the use of well-known identities. |