Kubernetes Concepts: Difference between revisions

From NovaOrdis Knowledge Base
Jump to navigation Jump to search
 
(241 intermediate revisions by the same user not shown)
Line 6: Line 6:


* [[Kubernetes#Subjects|Kubernetes]]
* [[Kubernetes#Subjects|Kubernetes]]
* [[Kubernetes_API_Resources_Concepts|API Resource Concepts]]
* [[Docker Concepts]]
* [[Docker Concepts]]
* [[OpenShift Concepts]]
* [[OpenShift Concepts]]


=TODO=
=Overview=
 
Kubernetes is an Apache 2.0 Open Source container orchestration platform, or a container orchestrator. 
 
A high level abstraction often used for Kubernetes is "data center OS". The primary use case for Kubernetes consists of containerized cloud-native applications, which are applications that are made from a set of small autonomous services (microservices) that communicate with each other. Kubernetes helps deploying, scaling up, scaling down, performing updates and rollbacks of these services, materialized as a set of containers. It does that at scale. It abstracts out details such as what specific compute nodes or physical storage volumes are allocated to applications.
 
<span id='Control_Plane'></span>Kubernetes instances are known as <span id='Cluster'></span>[[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Cluster|clusters]]. All management interactions with a Kubernetes cluster are performed by sending REST requests into an [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#API_Server|API Server]]. The API Server is responsible with managing and exposing the state of the cluster. The state of the cluster is internally stored by a [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Cluster_Store|cluster store]] control plane system service, which is currently implemented by [[etcd]]. The [[#Control_Loop|control loops]] essential to the [[#Declarative_versus_Imperative_Approach|declarative model]] implemented by Kubernetes are driven by various specialized controllers under the supervision of the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Controller_Manager|controller manager]]. The workloads are dispatched by the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Scheduler|scheduler]]. All these components - the API Server, cluster store, controllers, scheduler, cloud controller manager - are collectively known as the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Control_Plane|control plane]] and are executed on [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Master_Node|master nodes]].  Externally, the state can be accessed and modified with specialized tools, of which the most common is a command line client named [[kubectl]]. The control plane, the API server and API server-related details, such as [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Admission_Controllers|admission controllers]], are discussed in [[Kubernetes_Control_Plane_and_Data_Plane_Concepts|Control Plane and Data Plane Concepts]].
 
Application workloads are deployed as [[Kubernetes_Pod_and_Container_Concepts#Pod|pods]] on a set of [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Worker_Node|worker nodes]] that constitute the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Data_Plane|data plane]]. Each node runs a <span id='Container_Runtime'></span>[[Kubernetes_Container_Runtime_Concepts#Container_Runtime|container runtime]], usually Docker. However, support for other container runtimes is available, via [[Kubernetes_Container_Runtime_Concepts#Container_Runtime_Interface_.28CRI.29|Container Runtime Interface (CRI)]]. Container runtime details are discussed in [[Kubernetes_Container_Runtime_Concepts|Container Runtime Concepts]].
 
Worker nodes are used to run workloads, deployed as <span id='Pod'></span>[[Kubernetes Pod and Container Concepts|pods]]. Pods are scheduled to nodes and then they are closely monitored. A pod is a wrapper that allows one or more containers to run on Kubernetes and it is the atomic unit of deployment in Kubernetes. Pods come and go - if a pod dies, it is not resurrected, but its failure is detected by the lack of response from configured [[Kubernetes Container Probes|probes]] that test expected container behavior and, depending on configuration, another pod may be scheduled as replacement. In consequence, the IP address of an individual pod cannot be relied on. Pod, containers, probes and pod scheduling are discussed in [[Kubernetes Pod and Container Concepts#Overview|Pod and Container Concepts]].
 
Because the pods and their IPs are ephemeral, Kubernetes introduces an additional mechanism aimed at providing stable access point to a set of equivalent pods that belong to the same application: a <span id='Service'></span>[[Kubernetes Service Concepts#Service|service]]. A service can be thought of stable networking access for a continuously changing set of pods. A service's IP address and port can be relied on to not change for the life of the service. All live pods represented by a service at a moment in time are known as service "endpoints" - in fact, there is a Kubernetes resource representing a live pod within the context of a service, and it is called [[Kubernetes_Service_Concepts#Endpoints|endpoint]]. There are several types of services: ClusterIP, NodePort and LoadBalancer. The association between services and pods is loose - it is established logically by the service's selector, which is a label-based mechanism: a pod "belongs" to a service if the service's selector matches the pod's labels. Services are explained at length in the [[Kubernetes Service Concepts|Service Concepts]] section and selectors in [[Kubernetes_Selector_Concepts|Selector Concepts]]. A layer 7 complement to services, named Ingress, is available. Ingresses are discussed in [[Kubernetes Ingress Concepts|Ingress Concepts]].
 
A pod by itself has no built-in resilience: if it fails for any reason, it is gone. <span id='Higher_Level_Pod_Controller'></span>A higher level primitive - the [[Kubernetes_Workload_Resources#Deployment|deployment]] - is used to manage a set of pods from a high availability perspective: the deployment insures that a specific number of equivalent pods is always running, and if one of more pods fail, the deployment brings up replacement pods. The deployment relies on an intermediary concept - the [[Kubernetes_Workload_Resources#ReplicaSet|ReplicaSet]]. Deployments are used to implement rolling updates and rollbacks. There are higher-level pod controllers that manage sets of pods in different ways: [[Kubernetes_Workload_Resources#DaemonSet|DaemonSets]] and [[Kubernetes_Workload_Resources#StatefulSet|StatefulSets]]. Individual pods can be managed as [[Kubernetes_Workload_Resources#Job|Jobs]] or [[Kubernetes_Workload_Resources#CronJob|CronJobs]]. The pod controllers are discussed in [[Kubernetes_Workload_Resources#Overview|Kubernetes Workload Resources]].


Deplete [[Kubernetes Concepts TO DEPLETE]].
Most Kubernetes resources can be logically grouped in namespaces. Services, pods, secrets, any many others are all namespaced. New namespaces can be created administratively, and all Kubernetes clusters come with a "default" namespace. There are resource types that cannot be allocated to namespaces, and those are named "cluster-level" resources. More details about namespaces are available in [[Kubernetes Namespace Concepts#Overview|Namespace Concepts]].


=Overview=
All Kubernetes resources can be annotated with labels and annotations, which are aimed at facilitating creating loose associations between resources. For more details see [[Kubernetes Labels and Annotations|Kubernetes Labels and Annotations]].


Kubernetes is a container orchestration platform, or a container orchestrator. To understand how Kubernetes works is to understand a set of high level concepts, briefly mentioned here. More details on individual concepts are available on their respective pages.  
A Kubernetes cluster exposes external storage to pods with three API resources: [[Kubernetes_Storage_Concepts#PV|PersistentVolumes]], [[Kubernetes_Storage_Concepts#PVC|PersistentVolumeClaims]] and [[Kubernetes_Storage_Concepts#SC|StorageClasses]], which are part of the [[Kubernetes_Storage_Concepts#Persistent_Volume_Subsystem|persistent volume subsystem]]. The actual storage is made available to Kubernetes by [[Kubernetes_Storage_Concepts#Storage_Plugin|storage plugins]], also known as provisioners, which should abide by the [[Kubernetes_Storage_Concepts#Container_Storage_Interface_.28CSI.29|Container Storage Interface (CSI)]]. All these are explained at length in [[Kubernetes Storage Concepts|Storage Concepts]].


A good high level abstraction for Kubernetes is that is a data center OS. The primary target of Kubernetes are containerized cloud-native applications, which are applications that are made from a set of small autonomous services (micro services) that communicate with each other. Kubernetes help deploying, scaling up, scaling down, performing updates and rollbacks of these services, handled as containers.
Every pod in the Kubernetes cluster has its own IP address, which is directly routable to every other pod. The stable IP addresses provided by services are resolvable by the cluster's internal DNS service, as described in the [[Kubernetes_DNS_Concepts|DNS Concepts]]. This and other networking-related aspects are explained in [[Kubernetes Networking Concepts|Networking Concepts]].


Kubernetes instances are known as <span id='Cluster'></span>Clusters. All interactions with a Kubernetes Cluster are performed by sending REST requests into an API Server. The API Server is responsible with managing and exposing the state of the Cluster. The state of the cluster is internally stored by a Cluster Store component, which is currently etcd. Externally, the state can be accessed and modified most commonly via a command line client named [[kubectl]]. The Cluster consists in a set of Nodes. There are Master Nodes and Worker Nodes. Each Node runs a <span id='Container_Runtime'></span>[[Kubernetes Container Runtime Concepts|container runtime]], usually Docker. However, support for other container runtimes is available, via Container Runtime Interface (CRI).
Configuration can be exposed to pods specialized resources such as [[Kubernetes_Cluster_Configuration_Concepts#ConfigMap|ConfigMaps]] and [[Kubernetes_Cluster_Configuration_Concepts#Secret|Secrets]]. Kubernetes is exposing pod information to the containers running inside the pod through files, which are projected in the container by a mechanism known as the [[Kubernetes Downward API Concepts|Downward API]]. More details about configuration are available in [[Kubernetes Cluster Configuration Concepts|Configuration Concepts]].


Worker Nodes are used to run workloads, as <span id='Pod'></span>[[Kubernetes Pod Concepts|Pods]] - pods are scheduled to Nodes and closely monitored. A Pod is the atomic unit of deployment in Kubernetes, and contains one or more containers. Pods come and go - if one Pod dies, it is not resurrected, but another Pod might be scheduled as replacement. In consequence, the IP address of an individual Pod cannot be relied on. To provide a stable access point to a set of equivalent Pods - which is how applications are deployed on Kubernetes, Kubernetes uses the concept of <span id='Service'></span>[[Kubernetes Service Concepts#Service|Service]], which can be thought of as stable networking for a continuously changing set of Pods. A Service's IP address and port can be relied on to be stable for the life of the Service. All live Pods represented by a Service at a moment in time are known as the Service's endpoint. There are several types of Services: ClusterIP, NodePort and LoadBalancer. The association between Services and Pods they expose is loose, established logically by the Service's Selector, which is a label-based mechanism: a Pod "belongs" to a Service if the Service's Selector matches the Pod's Labels. A Pod by itself has no built-in resilience: if it fails for any reason, it is gone. A higher level concept - the Deployment - is used to manage a set of Pods from a high availability perspective: the Deployment insures that a specific number of equivalent Pods is always running, and if one of more Pods fail, the Deployment brings up replacement Pods. The deployment relies on an intermediary concept - the Replica Set. Deployments are used to implement rolling updates and rollbacks. There are other resources that manage sets of Pods in different ways: DaemonSets and StatefulSets. Individual Pods can be managed as Jobs or CronJobs.
Kubernetes security system ensures that the API sever is only accessed by authenticated identities, and the access is limited to resources that are supposed to be accessible to the authenticated identity, and also ensuring that the applications running in containers only access the node and network resources that are supposed to access, and nothing more. These aspects are discussed in [[Kubernetes Security Concepts#Overview|Security Concepts]].


Storage.
Resources managed by Kubernetes are subject to <span id='Kubernetes_Policies'></span>[[Kubernetes Policies|policies]]: [[Kubernetes_Resource_Management_Concepts#Limit_Ranges|Limit Ranges]], [[Kubernetes_Resource_Management_Concepts#Resource_Quotas|Resource Quotas]] and specifically for pods, [[Kubernetes Pod Security Policy Concepts|Pod Security Policies]]. More details on resource management are available in [[Kubernetes Resource Management Concepts|Resource Management Concepts]].


Networking.
Application health monitoring, resource consumption monitoring and scaling decisions require metrics to be collected and analyzed. Kubernetes facilitates metrics collection from containers, pods, services and other resources via metric pipelines. Metrics and metric pipelines are discussed in [[Metrics_in_Kubernetes#Overview|Metrics in Kubernetes]].


Security.
Pods can be automatically scaled up and down based on interpretation of their performance characteristic or resource consumption. Kubernetes provides a built-in autoscaling mechanism. For more details see [[Kubernetes_Autoscaling_Concepts#Overview|Autoscaling Concepts]].


=Subjects=
Kubernetes has built-in extension capabilities, allowing for [[Kubernetes_Custom_Resources|custom resources]] and registering multiple APIs servers via the [[Kubernetes Aggregation Layer#Overview|aggregation layer]]. A specific kind of extension are the [[Kubernetes_Operators_Concepts|operators]]. More details are available in [[Extending Kubernetes]].


* [[Kubernetes Container Runtime Concepts|Container Runtime Concepts]]
<span id='spec_and_status'></span>Spec and status: inputs and outputs.
* [[Kubernetes Pod Concepts|Pod Concepts]]
* [[Kubernetes Service Concepts|Service Concepts]]


=Declarative versus Imperative Approach=
=Declarative versus Imperative Approach=
The preferred style while operating Kubernetes is to use a declarative model: Kubernetes likes to manage its resources declaratively, where we describe how we want our application to look - the '''desired state''' - in a set of YAML files, named [[Kubernetes Manifests#Overview|manifests]], POST these files into the Kubernetes [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#API_Server|API Server]] with [[kubectl apply]] or other tools and wait for the changes to be applied. The [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Controller_Manager|controller manager]] and specialized controllers check whether the '''current state''' matches the desired state, and if the states do not match, they act to reconcile them, which usually happens after a short delay. <span id='Control_Loop'></span>This pattern is referred to as '''control loop'''. A "control loop" is a design pattern for distributed software that allows to define state declaratively and employ a controller to bring the current state to the desired state. It typically obtains the desired state, repeatedly observes the current state, determines differences and, if differences exist, reconciles differences. The terms "control loop" are used interchangeably with "watch loop" and "reconciliation loop".
Step-by-step, the declarative model works as follows:
# The desired state of the application is declared in the [[Kubernetes Manifests|manifest]] file.
# The manifest file is POSTed into the  [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#API_Server|API Server]], usually with [[kubectl]] command.
# The API server authenticates and authorizes the request and then validates the manifest.
# The API server stores the state - as desired state - in the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Cluster_Store|cluster store]].
# The API server identifies the [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Controller_Manager|controller]] responsible with enforcing and monitoring the state.
# The controller in charge implements the desired state, by adjusting the current state to match it.
# <span id='Current_State'></span>The [[Kubernetes_Control_Plane_and_Data_Plane_Concepts#Controller_Manager|controller manager]]'s control loops monitor the current state and make sure that it does not diverge from the desired state. The current state of a resource can be obtained from the cluster with [[Kubectl#Output_in_YAML_Format|CLI commands]].
# If the current state of the cluster diverges from the desired state, the cluster control plane will perform whatever tasks are necessary to bring those states in sync.
This model is the opposite of the traditional imperative model, where a precise sequence of specific commands are issued to explicitly adjust the state. In other words, in the declarative model we tell the cluster how things should look, as opposite to telling it how to adjust the state.
Also see: {{Internal|Infrastructure_as_Code_Concepts#Declarative_Infrastructure_Languages|Declarative Infrastructure Languages}}


=API Resources=
=Subjects=


The full list of API resources available to interact with in Kubernetes is generated by:
* [[Kubernetes_Control_Plane_and_Data_Plane_Concepts|Control Plane and Data Plane Concepts]]
kubectl api-resources
* <span id='Container_Runtime_Concepts'></span>[[Kubernetes Container Runtime Concepts|Container Runtime Concepts]]
* [[Kubernetes Pod and Container Concepts|Pod and Container Concepts]]
** [[Kubernetes_Scheduling,_Preemption_and_Eviction_Concepts|Kubernetes Scheduling, Preemption and Eviction Concepts]]
* [[Kubernetes Service Concepts|Service Concepts]]
* [[Kubernetes Ingress Concepts#Overview|Ingress Concepts]]
* [[Kubernetes Selector Concepts|Selector Concepts]]
* [[Kubernetes Workload Resources#Overview|Kubernetes Workload Resources]]
* [[Kubernetes Namespace Concepts|Namespace Concepts]]
* [[Kubernetes Labels and Annotations|Labels and Annotations Concepts]]
* [[Kubernetes Storage Concepts|Storage Concepts]]
* [[Kubernetes Networking Concepts|Networking Concepts]]
** [[Kubernetes DNS Concepts|DNS Concepts]]
* [[Kubernetes Cluster Configuration Concepts|Cluster Configuration Concepts]]
** [[Kubernetes_Downward_API_Concepts#Overview|Downward API Concepts]]
* [[Kubernetes Security Concepts|Security Concepts]]
* [[Kubernetes Resource Management Concepts|Resource Management Concepts]]
** [[Kubernetes Policies]]
* [[Metrics in Kubernetes|Metrics in Kubernetes]]
* [[Kubernetes Autoscaling Concepts|Autoscaling Concepts]]
* [[Extending Kubernetes]]
** [[Kubernetes Aggregation Layer|Aggregation Layer]]
** [[Kubernetes Custom Resources|Custom Resources]]
** [[Kubernetes Operators Concepts|Operators]]
* <span id='API_Resources'></span>[[Kubernetes API Resources Concepts|API Resources Concepts]]
** [[Kubernetes_API_Resources_Concepts#Names|Kuberentes Names]]
** [[Kubernetes Manifests|Resource Manifests]]
* [[Kubernetes Patterns|Kubernetes Patterns]]


'''NAME'''                              SHORTNAMES  APIGROUP                      NAMESPACED  '''KIND'''
=Kubernetes Flavors=
bindings                                                                      true        Binding
* [[Managed Kubernetes]]
componentstatuses                cs                                          false        ComponentStatus
* [[OpenShift]]
configmaps                        cm                                          true        ConfigMap
* [[kind]]
endpoints                        ep                                          true        Endpoints
events                            ev                                          true        Event
limitranges                      limits                                      true        LimitRange
namespaces                        ns                                          false        Namespace
nodes                            no                                          false        Node
persistentvolumeclaims            pvc                                        true        PersistentVolumeClaim
persistentvolumes                pv                                          false        PersistentVolume
pods                              po                                          true        Pod
podtemplates                                                                  true        PodTemplate
replicationcontrollers            rc                                          true        ReplicationController
resourcequotas                    quota                                      true        ResourceQuota
secrets                                                                      true        Secret
serviceaccounts                  sa                                          true        ServiceAccount
services                          svc                                        true        Service
mutatingwebhookconfigurations                  admissionregistration.k8s.io  false        MutatingWebhookConfiguration
validatingwebhookconfigurations                admissionregistration.k8s.io  false        ValidatingWebhookConfiguration
customresourcedefinitions        crd,crds    apiextensions.k8s.io          false        CustomResourceDefinition
apiservices                                    apiregistration.k8s.io        false        APIService
controllerrevisions                            apps                          true        ControllerRevision
daemonsets                        ds          apps                          true        DaemonSet
deployments                      deploy      apps                          true        Deployment
replicasets                      rs          apps                          true        ReplicaSet
statefulsets                      sts          apps                          true        StatefulSet
tokenreviews                                  authentication.k8s.io          false        TokenReview
localsubjectaccessreviews                      authorization.k8s.io          true        LocalSubjectAccessReview
selfsubjectaccessreviews                      authorization.k8s.io          false        SelfSubjectAccessReview
selfsubjectrulesreviews                        authorization.k8s.io          false        SelfSubjectRulesReview
subjectaccessreviews                          authorization.k8s.io          false        SubjectAccessReview
horizontalpodautoscalers          hpa          autoscaling                    true        HorizontalPodAutoscaler
cronjobs                          cj          batch                          true        CronJob
jobs                                          batch                          true        Job
certificatesigningrequests        csr          certificates.k8s.io            false        CertificateSigningRequest
stacks                                        compose.docker.com            true        Stack
leases                                        coordination.k8s.io            true        Lease
events                            ev          events.k8s.io                  true        Event
daemonsets                        ds          extensions                    true        DaemonSet
deployments                      deploy      extensions                    true        Deployment
ingresses                        ing          extensions                    true        Ingress
networkpolicies                  netpol      extensions                    true        NetworkPolicy
podsecuritypolicies              psp          extensions                    false        PodSecurityPolicy
replicasets                      rs          extensions                    true        ReplicaSet
ingresses                        ing          networking.k8s.io              true        Ingress
networkpolicies                  netpol      networking.k8s.io              true        NetworkPolicy
runtimeclasses                                node.k8s.io                    false        RuntimeClass
poddisruptionbudgets              pdb          policy                        true        PodDisruptionBudget
podsecuritypolicies              psp          policy                        false        PodSecurityPolicy
clusterrolebindings                            rbac.authorization.k8s.io      false        ClusterRoleBinding
clusterroles                                  rbac.authorization.k8s.io      false        ClusterRole
rolebindings                                  rbac.authorization.k8s.io      true        RoleBinding
roles                                          rbac.authorization.k8s.io      true        Role
priorityclasses                  pc          scheduling.k8s.io              false        PriorityClass
csidrivers                                    storage.k8s.io                false        CSIDriver
csinodes                                      storage.k8s.io                false        CSINode
storageclasses                    sc          storage.k8s.io                false        StorageClass
volumeattachments                              storage.k8s.io                false        VolumeAttachment

Latest revision as of 21:06, 22 March 2024

External

Internal

Overview

Kubernetes is an Apache 2.0 Open Source container orchestration platform, or a container orchestrator.

A high level abstraction often used for Kubernetes is "data center OS". The primary use case for Kubernetes consists of containerized cloud-native applications, which are applications that are made from a set of small autonomous services (microservices) that communicate with each other. Kubernetes helps deploying, scaling up, scaling down, performing updates and rollbacks of these services, materialized as a set of containers. It does that at scale. It abstracts out details such as what specific compute nodes or physical storage volumes are allocated to applications.

Kubernetes instances are known as clusters. All management interactions with a Kubernetes cluster are performed by sending REST requests into an API Server. The API Server is responsible with managing and exposing the state of the cluster. The state of the cluster is internally stored by a cluster store control plane system service, which is currently implemented by etcd. The control loops essential to the declarative model implemented by Kubernetes are driven by various specialized controllers under the supervision of the controller manager. The workloads are dispatched by the scheduler. All these components - the API Server, cluster store, controllers, scheduler, cloud controller manager - are collectively known as the control plane and are executed on master nodes. Externally, the state can be accessed and modified with specialized tools, of which the most common is a command line client named kubectl. The control plane, the API server and API server-related details, such as admission controllers, are discussed in Control Plane and Data Plane Concepts.

Application workloads are deployed as pods on a set of worker nodes that constitute the data plane. Each node runs a container runtime, usually Docker. However, support for other container runtimes is available, via Container Runtime Interface (CRI). Container runtime details are discussed in Container Runtime Concepts.

Worker nodes are used to run workloads, deployed as pods. Pods are scheduled to nodes and then they are closely monitored. A pod is a wrapper that allows one or more containers to run on Kubernetes and it is the atomic unit of deployment in Kubernetes. Pods come and go - if a pod dies, it is not resurrected, but its failure is detected by the lack of response from configured probes that test expected container behavior and, depending on configuration, another pod may be scheduled as replacement. In consequence, the IP address of an individual pod cannot be relied on. Pod, containers, probes and pod scheduling are discussed in Pod and Container Concepts.

Because the pods and their IPs are ephemeral, Kubernetes introduces an additional mechanism aimed at providing stable access point to a set of equivalent pods that belong to the same application: a service. A service can be thought of stable networking access for a continuously changing set of pods. A service's IP address and port can be relied on to not change for the life of the service. All live pods represented by a service at a moment in time are known as service "endpoints" - in fact, there is a Kubernetes resource representing a live pod within the context of a service, and it is called endpoint. There are several types of services: ClusterIP, NodePort and LoadBalancer. The association between services and pods is loose - it is established logically by the service's selector, which is a label-based mechanism: a pod "belongs" to a service if the service's selector matches the pod's labels. Services are explained at length in the Service Concepts section and selectors in Selector Concepts. A layer 7 complement to services, named Ingress, is available. Ingresses are discussed in Ingress Concepts.

A pod by itself has no built-in resilience: if it fails for any reason, it is gone. A higher level primitive - the deployment - is used to manage a set of pods from a high availability perspective: the deployment insures that a specific number of equivalent pods is always running, and if one of more pods fail, the deployment brings up replacement pods. The deployment relies on an intermediary concept - the ReplicaSet. Deployments are used to implement rolling updates and rollbacks. There are higher-level pod controllers that manage sets of pods in different ways: DaemonSets and StatefulSets. Individual pods can be managed as Jobs or CronJobs. The pod controllers are discussed in Kubernetes Workload Resources.

Most Kubernetes resources can be logically grouped in namespaces. Services, pods, secrets, any many others are all namespaced. New namespaces can be created administratively, and all Kubernetes clusters come with a "default" namespace. There are resource types that cannot be allocated to namespaces, and those are named "cluster-level" resources. More details about namespaces are available in Namespace Concepts.

All Kubernetes resources can be annotated with labels and annotations, which are aimed at facilitating creating loose associations between resources. For more details see Kubernetes Labels and Annotations.

A Kubernetes cluster exposes external storage to pods with three API resources: PersistentVolumes, PersistentVolumeClaims and StorageClasses, which are part of the persistent volume subsystem. The actual storage is made available to Kubernetes by storage plugins, also known as provisioners, which should abide by the Container Storage Interface (CSI). All these are explained at length in Storage Concepts.

Every pod in the Kubernetes cluster has its own IP address, which is directly routable to every other pod. The stable IP addresses provided by services are resolvable by the cluster's internal DNS service, as described in the DNS Concepts. This and other networking-related aspects are explained in Networking Concepts.

Configuration can be exposed to pods specialized resources such as ConfigMaps and Secrets. Kubernetes is exposing pod information to the containers running inside the pod through files, which are projected in the container by a mechanism known as the Downward API. More details about configuration are available in Configuration Concepts.

Kubernetes security system ensures that the API sever is only accessed by authenticated identities, and the access is limited to resources that are supposed to be accessible to the authenticated identity, and also ensuring that the applications running in containers only access the node and network resources that are supposed to access, and nothing more. These aspects are discussed in Security Concepts.

Resources managed by Kubernetes are subject to policies: Limit Ranges, Resource Quotas and specifically for pods, Pod Security Policies. More details on resource management are available in Resource Management Concepts.

Application health monitoring, resource consumption monitoring and scaling decisions require metrics to be collected and analyzed. Kubernetes facilitates metrics collection from containers, pods, services and other resources via metric pipelines. Metrics and metric pipelines are discussed in Metrics in Kubernetes.

Pods can be automatically scaled up and down based on interpretation of their performance characteristic or resource consumption. Kubernetes provides a built-in autoscaling mechanism. For more details see Autoscaling Concepts.

Kubernetes has built-in extension capabilities, allowing for custom resources and registering multiple APIs servers via the aggregation layer. A specific kind of extension are the operators. More details are available in Extending Kubernetes.

Spec and status: inputs and outputs.

Declarative versus Imperative Approach

The preferred style while operating Kubernetes is to use a declarative model: Kubernetes likes to manage its resources declaratively, where we describe how we want our application to look - the desired state - in a set of YAML files, named manifests, POST these files into the Kubernetes API Server with kubectl apply or other tools and wait for the changes to be applied. The controller manager and specialized controllers check whether the current state matches the desired state, and if the states do not match, they act to reconcile them, which usually happens after a short delay. This pattern is referred to as control loop. A "control loop" is a design pattern for distributed software that allows to define state declaratively and employ a controller to bring the current state to the desired state. It typically obtains the desired state, repeatedly observes the current state, determines differences and, if differences exist, reconciles differences. The terms "control loop" are used interchangeably with "watch loop" and "reconciliation loop".

Step-by-step, the declarative model works as follows:

  1. The desired state of the application is declared in the manifest file.
  2. The manifest file is POSTed into the API Server, usually with kubectl command.
  3. The API server authenticates and authorizes the request and then validates the manifest.
  4. The API server stores the state - as desired state - in the cluster store.
  5. The API server identifies the controller responsible with enforcing and monitoring the state.
  6. The controller in charge implements the desired state, by adjusting the current state to match it.
  7. The controller manager's control loops monitor the current state and make sure that it does not diverge from the desired state. The current state of a resource can be obtained from the cluster with CLI commands.
  8. If the current state of the cluster diverges from the desired state, the cluster control plane will perform whatever tasks are necessary to bring those states in sync.

This model is the opposite of the traditional imperative model, where a precise sequence of specific commands are issued to explicitly adjust the state. In other words, in the declarative model we tell the cluster how things should look, as opposite to telling it how to adjust the state.

Also see:

Declarative Infrastructure Languages

Subjects

Kubernetes Flavors