WO2024049636A1 - Optimized system design for deploying and managing containerized workloads at scale - Google Patents

Optimized system design for deploying and managing containerized workloads at scale Download PDF

Info

Publication number
WO2024049636A1
WO2024049636A1 PCT/US2023/030300 US2023030300W WO2024049636A1 WO 2024049636 A1 WO2024049636 A1 WO 2024049636A1 US 2023030300 W US2023030300 W US 2023030300W WO 2024049636 A1 WO2024049636 A1 WO 2024049636A1
Authority
WO
WIPO (PCT)
Prior art keywords
control plane
configuration
state
workload
deploying
Prior art date
Application number
PCT/US2023/030300
Other languages
French (fr)
Inventor
Erol Aygar
Nikhil Deshpande
Benjamin John Corrie
George Gregory Hicken
Elliott DAVIS
Original Assignee
Vmware, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US18/157,196 external-priority patent/US20240078127A1/en
Application filed by Vmware, Inc. filed Critical Vmware, Inc.
Publication of WO2024049636A1 publication Critical patent/WO2024049636A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/60Software deployment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0895Configuration of virtualised networks or elements, e.g. virtualised network function or OpenFlow elements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/084Configuration by using pre-existing information, e.g. using templates or copying from other elements
    • H04L41/0846Configuration by using pre-existing information, e.g. using templates or copying from other elements based on copy from other elements

Definitions

  • Modem applications are applications designed to take advantage of the benefits of modem computing platforms and infrastructure. For example, modern applications can be deployed in a multi-cloud or hybrid cloud fashion, such as, consuming both cloud services executing in a public cloud and local services executing in a private data center (e.g., a private cloud). Within the public cloud or private data center, modem applications can be deployed onto one or more virtual machines (VMs), containers, application services, and/or the like.
  • VMs virtual machines
  • a container is a package that relies on virtual isolation to deploy and run applications that access a shared operating system (OS) kernel.
  • Containerized applications also referred to as containerized workloads, can include a collection of one or more related applications packaged into one or more groups of containers, referred to as pods, that can be deployed based on the context defined in a manifest for the containerized workloads.
  • Each pod may include one or more containers.
  • Containerized workloads run on a container orchestration platform that enables the automation of much of the operational effort required to run containers having workloads and services.
  • This operational effort includes a wide range of things needed to manage a container’s lifecycle, including, but not limited to, provisioning, deployment, scaling (up and down), networking, and load balancing.
  • Kubernetes® (K8S)® software is an example open-source container orchestration platform that automates the operation of such containerized workloads.
  • Kubernetes software allows for distributed computing by running the pods of containerized workloads on a cluster of interconnected worker nodes (e.g., VMs or physical machines) that may scale vertically and/or horizontally over hybrid cloud topology.
  • containers are the building-blocks that enable a scalable environment
  • containers are not the only part of the software stack that needs to scale.
  • tools used to instantiate, manage, monitor, and/or secure containers may also need to be able to scale as seamlessly as the containers.
  • scalability of virtualization software and architecture for deploying and managing containerized workloads may also affect an ability of the environment to handle an increased or expanding workload.
  • a software-defined data center includes clusters of physical servers (e.g., hosts) that are virtualized and managed by virtualization management servers.
  • a host can include a virtualization layer (e.g., a hypervisor) that provides a software abstraction of the hardware platform of the physical server (e.g., central processing unit (CPU), random access memory (RAM), storage, network interface card (NIC), etc.) to allow multiple virtual computing instances (e.g., such as VMs) to run thereon.
  • a control plane for each cluster of hosts may support the deployment and management of applications (or services) on the cluster using containers. In some cases, the control plane deploys applications as pods of containers running on one or more worker nodes.
  • scalability of the environment may also depend on the operational effort required to initialize the control plane and/or worker nodes for container deployment, as well as an ability of the control plane to determine a context defined in a manifest for instantiation of workloads in these containers.
  • One or more embodiments provide a method of automatically deploying a containerized workload on a hypervisor based device.
  • the method generally includes booting the device running a hypervisor, in response to booting the device: automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload; deploying a control plane pod configured according to the control plane configuration; deploying one or more worker nodes based on the control plane configuration, and deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
  • FIG. 1 is a block diagram of an exemplary container-based cluster according to aspects of the present disclosure.
  • FIG. 2 is a block diagram of an example system for deploying and managing distributed containerized workloads at scale according to aspects of the present disclosure.
  • FIG. 3 is a flow diagram illustrating example operations for automatically deploying containerized workloads according to aspects of the present disclosure.
  • FIG. 4 is a flow diagram illustrating example operations for managing different states of a deployed system according to aspects of the present disclosure.
  • FIG. 5 shows a flow diagram depicting a method of automatically deploying a containerized workload on a hypervisor based device according to an embodiment.
  • aspects of the present disclosure provide an auto-deployment system for automatically deploying and/or managing containerized workloads at scale.
  • the containerized workloads may be distributed.
  • Automated deployment refers to the automation of steps, processes, and/or activities that are necessary to make a system and/or update available to its intended users.
  • automation of much of the operational effort required to set up and/or manage a system capable of supporting containerized workloads is based on an ability of the system to access one or more intended state configuration files made up of one or more manifests that declare intended system infrastructure and workloads to be deployed in the system.
  • the manifests are JavaScript Object Notation (JSON) and/or YAML files.
  • deployment of the system to a device may be triggered by a user booting a device, such as from a USB drive, flash drive, SD-RAM card, preboot execution environment (PXE), etc.
  • the device may execute an initialization script that causes a hypervisor to be installed on the device.
  • the hypervisor may be a type-1 hypervisor that includes and integrates operating system (OS) components, such as a kernel, security management, file, process, memory, and/or input/output (VO) device management.
  • OS operating system
  • VO input/output
  • the installed hypervisor may include an infrastructure supervisor (infravisor) layer that provides a runtime for services for a cluster of hosts.
  • the infravisor layer provides a cluster runtime for virtualization services and management, as further described herein.
  • the infravisor layer may be configured to provide a control plane service for managing the cluster of hosts, including configuring the cluster of hosts to run according to a defined intended state.
  • a control plane and one or more infravisor pods may be instantiated on the infravisor layer to provide a service for deploying and/or managing containerized workloads.
  • the control plane may be created to deploy and automatically manage clusters of containerized applications such that they align with their intended states as defined in one or more intended state configuration files. Accordingly, when deployed, the control plane may be configured to (1) determine one or more intended infrastructure configuration parameters by accessing one or more of the intended state configuration files at an accessible external datastore and (2) build the infrastructure based on these determined parameters.
  • the intended infrastructure configuration parameters may declare infrastructure necessary for the deployment of one or more applications, for example, a number of worker node virtual machines (VMs) necessary for deployment inside the infravisor layer, virtual network configurations, storage configurations, etc.
  • VMs virtual machines
  • FIG. 1 the intended infrastructure configuration parameters may declare infrastructure necessary for the deployment of one or more applications, for example, a number of worker node virtual machines (VMs) necessary for deployment inside the infravisor layer, virtual network configurations, storage configurations, etc.
  • VMs virtual machines
  • VCIs virtual machines
  • hardware nodes may similarly be used as worker nodes.
  • building the infrastructure involves deploying one or more infravisor pods on the infravisor layer. At least one of the pods may be used to (1) determine one or more workloads to be instantiated on the created infrastructure, by accessing one or more of the intended state configuration files at an accessible external datastore and (2) instantiate these workloads. In some cases, these workloads are instantiated on one or more worker node VMs running on the hypervisor.
  • the system described herein when booted, is designed to automatically pull intended configuration parameters (e.g., from an external datastore) and apply these parameters to automatically configure an underlying virtualization infrastructure of the system and execute containerized workloads on the configured infrastructure. Accordingly, human interaction and/or intervention with the system may not be necessary, thereby resulting in a more streamlined deployment of containerized workloads which may help to decrease errors, increase speed of delivery, boost quality, minimize costs, and overall, simplify the deployment process. Further, the auto-deployed system described herein may be able to scale.
  • intended configuration parameters e.g., from an external datastore
  • introducing automation into deployment of the containerized workloads and underlying infrastructures allows for a same system to be instantiated multiple times, and/or a much larger system (including multiple clusters, containers, applications, etc.) to be instantiated in the same way (e.g., without human intervention).
  • the one or more intended state configuration files are stored in a repository on a version control system.
  • GitHub made commercially available by GitHub, Inc.
  • GitHub is a git repository hosting service which provides version control software capable of managing and tracking changes to software code. Accordingly, intended configuration parameters stored in the one or more intended sate configuration files may be easily updated and tracked when stored on the version control system. Any updates made to the intended state configuration files may be automatically pulled and used by the system for adjustment to help guarantee consistent realization of the intended state defined by the intended state configuration files.
  • the one or more intended state configuration files include at least two intended state configuration files.
  • a first file (referred to as the “infrastructure manifest”) may include one or more intended infrastructure configuration parameters
  • the second file (referred to as the “workload manifest”) may include one or more intended workload definitions.
  • patching e.g., identifying, acquiring, testing and installing patches, or code changes, that are intended to resolve functionality issues, improve security, and/or add features
  • any number of intended state configuration files may be used, such as a single file including both intended infrastructure configuration parameters and intended workload definitions.
  • control plane deployed to automatically deploy and manage containerized workloads may be isolated from such workloads.
  • the control plane may be isolated from a worker plane where worker node VMs are instantiated for running the containerized workloads.
  • a state of the control plane may be managed separate from a state of the worker node VMs in the work plane.
  • the state of the control plane is managed by a control plane pod within the control plane.
  • the state of the worker node VMs in the worker plane are managed by a worker pod in the extension plane, which is separated from the control plane.
  • the system described herein offers logical separation of the state of the control plane from the state of the worker node VMs, as well as separate management of each of the states. Isolation of the control plane from the workloads may result in improved redundancy, security, and/or scalability.
  • FIG. 1 a block diagram of an exemplary container-based cluster is illustrated. It should be noted that the block diagram of Fig. 1 is a logical representation of a container-based cluster, and does not show where the various components are implemented and run on physical systems.
  • Fig. 2 provides an example system for implementing and running a container-based cluster of Fig. 1. While the example container-based cluster shown in Fig.
  • the container-based cluster may be another type of container-based cluster based on container technology, such as Docker® clusters.
  • a cluster such as K8S cluster 100 illustrated in FIG. 1, is formed from a combination of worker nodes 104 and a control plane 102. Worker nodes 104 are managed by control plane 102, which manages the computation, storage, and memory resources to run all worker nodes 104.
  • Each worker node 104 or worker compute machine, includes a kubelet 106, which is an agent that ensures that one or more pods 110 run in the worker node 104 according to a defined specification for the pods, such as defined in a workload definition manifest.
  • Each pod 110 may include one or more containers 112.
  • the worker nodes 104 can be used to execute various applications and software processes using container 112.
  • each worker node 104 includes a kube proxy 108.
  • Kube proxy 108 is a Kubemetes network proxy that maintains network rules on worker nodes 104. These network rules allow for network communication to pods 110 from network sessions inside and/or outside of K8S cluster 100.
  • Control plane 102 includes components such as an application programming interface (API) server 114, a cluster store (etcd) 116, a controller 118, and a scheduler 120.
  • API application programming interface
  • Control plane 102 ’s components make global decisions about K8S cluster 100 (e.g., scheduling), as well as detect and respond to cluster events (e.g., starting up a new pod 110 when a workload deployment's replicas field is unsatisfied).
  • API server 114 operates as a gateway to K8S cluster 100. As such, a command line interface, web user interface, users, and/or services communicate with K8S cluster 100 through API server 114.
  • a Kubemetes API server 114 is kube-apiserver.
  • kube- apiserver is designed to scale horizontally — that is, this component scales by deploying more instances. Several instances of kube-apiserver may be run, and traffic may be balanced between those instances.
  • Cluster store (etcd) 116 is a data store, such as a consistent and highly-available key value store, used as a backing store for all K8S cluster 100 data.
  • Controller 118 is a control plane 102 component that runs and manages controller processes in K8S cluster 100.
  • control plane 102 may have (e.g., four) control loops called controller processes, which watch the state of cluster 100 and try to modify the current state of cluster 100 to match a intended state of cluster 100.
  • controller processes of controller 118 are configured to monitor external storage for changes to the state of cluster 100.
  • Scheduler 120 is a control plane 102 component configured to allocate new pods 110 to worker nodes 104. Additionally, scheduler 118 may be configured to distribute resources and/or workloads across worker nodes 104. Resources may refer to processor resources, memory resources, networking resources, and/or the like. Schedule 118 may watch worker nodes 104 for how well each worker node 104 is handling their workload, and match available resources to the worker nodes 104. Scheduler 118 may then schedule newly created containers 112 to one or more of the worker nodes 104.
  • control plane 102 manages and controls every component of the cluster 100.
  • Control plane 102 handles most, if not all, operations within cluster 100, and its components define and control cluster 100’s configuration and state data.
  • Control plane 102 configures and runs the deployment, management, and maintenance of the containerized applications. As such, ensuring high availability of the control plane may be critical to container deployment and management. High availability is a characteristic of a component or system that is capable of operating continuously without failing.
  • control plane 102 may operate as a high availability (HA) control plane. Additional details of HA control planes are disclosed in U.S. Application Serial No. 63/347,815, filed on June 1, 2022, and titled “AUTONOMOUS CLUSTERS IN A VIRTUALIZATION COMPUTING ENVIRONMENT,” which is hereby incorporated by reference herein in its entirety.
  • containers are not the only part of the software stack that needs to be automated for scaling.
  • deployment of a control plane e.g., such as control plane 102 illustrated in Fig. 1 may be necessary for the deployment and management of such containerized workloads.
  • Architecture of the auto-deployed system described herein includes an improved system designed for deploying and managing distributed containerized workloads at scale.
  • the system described herein includes one or more components configured to automatically configure and deploy a control plane for managing containerized workloads.
  • the system further includes one or more components to determine a number of worker nodes needed for deployment of intended workloads, deploy the determined number of worker nodes, and instantiate the intended workloads on the deployed worker nodes.
  • the system design described herein offers physical separation of the state of a control plane used to manage deployed workloads from the state of worker nodes where the workloads are deployed.
  • Fig. 2 is a block diagram of an auto-deployed system 200 for deploying and managing distributed containerized workloads at scale according to aspects of the present disclosure.
  • the architecture of system 200 described in Fig. 2 may allow for implementation of a container-based cluster on a hypervisor.
  • a logical construct of K8S cluster 100 illustrated in Fig. 1 may be implemented by the architecture described below with reference to Fig. 2. Further, such architecture may bring along with it a new management model, as described in more detail below.
  • a hypervisor is a type of virtualization software that supports the creation and management of virtual endpoints by separating a physical machine’s software from its hardware. In other words, hypervisors translate requests between physical and virtual resources, thereby making virtualization possible.
  • bare-metal hypervisor When a hypervisor is installed directly on the hardware of a physical machine, as opposed to on top of an operating system (OS) of the machine, the hypervisor is referred to as bare-metal hypervisor.
  • hypervisor 201 illustrated in Fig. 2 is a bare-metal hypervisor.
  • a host 203 runs hypervisor 201.
  • Host 203 is a physical computing device, such as a server-grade computer, that includes hardware components, such as a memory, processor, storage, networking card, and/or the like, for running components described herein.
  • a user interface may be provided to enable users to interact with hypervisor 201, such as to check on system status, update configuration, etc.
  • the user interface may be accessible by directly accessing host 203, or by accessing host 203 over a network, such as via machine web browser or API client.
  • hypervisor 201 may include a host daemon 230 running as a background process, which in part allows connection to hypervisor 201 for monitoring hypervisor 201.
  • hypervisor 201 is a multi-layer entity, where each layer is provided a different level of privilege.
  • hypervisor 201 architecture may include underlying OS features, referred to as a kernel, and processes that run on top of the kernel.
  • the kernel may be a microkernel that provides functions such as process creation, process control, process threads, signals, file system, etc.
  • a process running on or above the kernel may be referred to as a "user world" process.
  • a user world process may run in a limited environment.
  • a privilege level of the kernel may be greater than a privilege level of a user world process.
  • Hypervisor 201 may include an infravisor daemon 228 running as a background process.
  • infravisor daemon 228 is an infravisor watchdog running on hypervisor 201.
  • the infravisor daemon 228 is configured to monitor individual infravisor services (e.g., including an infravisor runtime pod 226, described in detail below) running in a cluster of hosts to help guarantee that a minimum number of individual services are continuously running in the cluster.
  • infravisor daemon 228 monitors an API server (e.g., such as API server 114 illustrated in Fig. 1) to determine whether a minimum number of individual services are running.
  • API server e.g., such as API server 114 illustrated in Fig.
  • Hypervisor 201 may further include an infravisor runtime pod 226, which may be a pod of containers running on the hypervisor 201 that execute control plane entities, such as API server 114, cluster store (etcd) 116, controller 118, and scheduler 120 illustrated in Fig. 1, for a cluster of hosts. These components may run as separate or consolidated pods for isolation or footprint reduction, respectively.
  • the infravisor runtime pod 226 may access the cluster store (etcd) 116 to store a cluster’s runtime state.
  • the infravisor runtime pod 226 is bootstrapped on a host in a cluster when the infravisor daemon 228 detects the absence of a functional infravisor runtime pod. It should be noted that such control plane functionality provided by infravisor runtime pod 226 may be separate from the control plane 206 described herein for worker nodes. In particular, while infravisor runtime pod 226 and a control plane pod 212 of control plane 206 may execute similar control plane/runtime entities, in certain aspects, the infravisor runtime pod 226 runs in a higher privilege level than control plane pod 212 of control plane 206.
  • control plane pod 212 of control plane 206 manages the runtime state of such containerized workloads, and the infrastructure necessary for implementation.
  • Hypervisor 201 provides resources of host 203 to run one or more pods or services, collectively referred to as a Keswick node 204, which is a logical abstraction of the one or more pods or services.
  • a Keswick node 204 is a logical abstraction of the one or more pods or services.
  • the pods and services of the Keswick node 204 are logically separated by function into a control plane 206, an extension plane 208, and a worker plane 210, which are used to provide services for deploying and/or managing containerized workloads.
  • Control plane 206 includes a control plane pod 212, which may be a pod of containers running on the hypervisor 201 that execute control plane entities, such as API server 114, cluster store (etcd) 116, controller 118, and scheduler 120 illustrated in Fig. 1, for worker nodes 224.
  • the control plane pod 212 runs an infrastructure state controller 214 configured to manage the state of control plane 206, such as a number of worker nodes 224 to run, networking configuration of such worker nodes 224, etc.
  • control plane 206 is configured based on infrastructure manifest 244 stored in storage 240.
  • Infrastructure manifest 244 provides information about intended system infrastructure to be deployed on host 203.
  • infrastructure manifest 244 may define the infrastructure on which containerized workloads are expected to run. This may include information about a number of worker VMs 224 to instantiate, assignment of hardware resources to worker VMs 224, software configuration (e.g., a version of Kubemetes an application/workload uses), and/or network infrastructure (e.g., a software defined network).
  • the infrastructure manifest 244 may indicate a number of worker node VMs to deploy on hypervisor 201 and, in some cases, images to use for instantiating each of these worker node VMs.
  • the number of worker node VMs indicated in infrastructure manifest 244 may be a number of worker node VMs needed to run particular workloads defined in a workloads manifest 242.
  • infrastructure manifest 244 is included in an intended state configuration file.
  • the intended state configuration file may include one or more other manifests (e.g., such as workloads manifest 242).
  • the intended state configuration file may be stored in storage 240, which may be an external storage that is accessible by hypervisor 201. Storage 240 may further be accessible by infrastructure state controller 214 of control plane 206 after the control plane is instantiated, such as to monitor for updates to the infrastructure manifest 244 and automatically update the configuration of control plane 206, accordingly.
  • storage 240 is a repository on a version control system. As mentioned previously, one example version control system that may be configured and used in aspects described herein is GitHub made commercially available by GitHub, Inc.
  • hypervisor 201 may be configured to pull information from infrastructure manifest 244 and use this information to instantiate and configure control plane 206, such as by instantiating and configuring control plane pod 212. In certain aspects, this involves instantiating worker plane 210 by deploying one or more worker node VMs 224 in worker plane 210. A number of worker node VMs 224 deployed in worker plane 210 may be based, at least in part, on a number of work node VMs indicated for deployment in infrastructure manifest 244.
  • Infrastructure state controller 214 on control plane 206 is configured to manage a state of the infrastructure.
  • infrastructure state controller 214 accepts an “intended state” (also referred to as “desired state” or “declared state”) from a human operator (e.g., via infrastructure manifest 244), observes the state of the infrastructure, and dynamically configures the infrastructure such that the infrastructure matches the “intended state.”
  • infrastructure state controller 214 may also be configured to interact with infrastructure manifest 244 stored in storage 240.
  • infrastructure state controller 214 monitors storage 240 for changes/updates to infrastructure manifest 244.
  • Infrastructure state controller 214 may be configured to dynamically update the infrastructure such that the infrastructure matches a new “intended state” defined by infrastructure manifest 244, for example, when infrastructure state controller 214 determines infrastructure manifest 244 has been updated.
  • Worker node VMs 224 deployed in worker plane 210 are compute resources that use software to run programs and deploy applications/workloads. More specifically, worker node VMs 224 may be used to deploy containerized workloads on hypervisor 201. Worker node VMs 224 deployed in worker plane 210 may each include a cluster agent 222. A cluster agent 222 may be a container or a pod within a worker node VM 224. In certain aspects, cluster agent 222 is configured to monitor the health of a container-based cluster supported via worker node VM 224. Further, in certain aspects, cluster agent 222 is configured to collect metrics and metadata for a container-based cluster deployed on worker node VM 224, including each node and namespace down to the container level.
  • worker plane 210 is isolated from control plane 206; thus, worker node VMs 224 and their corresponding cluster agents 222 are similarly isolated from control plane 206.
  • Worker plane 210 describes resources managed by hypervisor 201 dedicated for running VMs. These resources may be distinct from resources used to run control plane 206 and extension plane 208.
  • Extension plane 208 includes a runtime controller for worker nodes 216 and an admin worker pod 220 which includes GitOps agents 218.
  • GitOps agents 218 are configured to interact with workloads manifest 242 stored in storage 240.
  • Workloads manifest 242 provides information about intended workloads to be deployed in hypervisor 201.
  • workloads manifest 242 may outline details of one or more workloads to be deployed in worker node VMs 224 in worker plane 210 on hypervisor 201.
  • workloads manifest 242 includes an identifier of a binary to be loaded.
  • workloads manifest 242 includes information about resources to be deployed, workload parameters associated with these resources, and/or protected resources for one or more workloads.
  • the workload parameters may include a workload name, a workload ID, a service name, an associated organization ID, and/or the like.
  • workloads manifest 242 is included in an intended state configuration file.
  • the intended state configuration file may include one or more other manifests (e.g., such as infrastructure manifest 244).
  • the intended state configuration file may be stored in storage 240 which is external storage that is accessible by GitOps agents 218.
  • GitOps agents 218 may be configured to pull information from workloads manifest 242 and use this information to instantiate workloads on worker node VMs 224 running in worker plane 210 (e.g., previously deployed by control plane 206).
  • Runtime controller for worker nodes 216 is configured to manage a state of the worker node VMs 224.
  • runtime controller for worker nodes 216 accepts a “intended state” (also referred to as “intended stated” or “declared state”) from a human operator (e.g., via workloads manifest 242), observes the state of the worker node VMs 224, and dynamically configures the worker node VMs 224 such that their behavior matches the “intended state.”
  • runtime controller for worker nodes 216 monitors storage 240 for changes/updates to workloads manifest 242.
  • Runtime controller for worker nodes 216 may be configured to dynamically update the state of the worker node VMs 224 to match a new “intended state” defined by workloads manifest 244, for example, when runtime controller for work nodes 216 determines workloads manifest 244 has been updated.
  • privilege becomes diluted when moving from bottom to top layers of hypervisor 201.
  • the infravisor layer of hypervisor 201 is at a lower, more privileged level of hypervisor 201, while control plane 206 and extension plane 208 are at a lesser-privileged level in hypervisor 201.
  • the worker node VMs 224 running in worker plane 210 may be on top of hypervisor 201, as this is where the deployed workloads are expected to run.
  • defined management levels may be assigned to different entities. For example, in certain aspects, worker node VMs 224 are managed by control plane pod 212 of control plane 206, and the control plane pod 212 is managed by the infravisor layer of hypervisor 201.
  • Keswick Node 204 is a logical abstraction that represents the control plane 206, extension plane 208, and worker plane 210. Though certain example implementations are described herein of how each of the control plane 206, extension plane 208, and worker plane 210 are implemented (e.g., as pods, VMs, etc.) and where they run (e.g., in hypervisor 201, on top of hypervisor 201, etc.), it should be noted that other implementations may be possible, such as having certain components run in different privilege levels, layers, within hypervisor 201, outside hypervisor 201, etc.
  • hypervisor 201 includes an initialization script 232.
  • Initialization script 232 is a sequence of instructions that are interpreted or carried out during startup of hypervisor 201. Initialization script 232 helps to automate the deployment of hypervisor 201 and Keswick node 204. In other words, initialization script 232 helps to automate and streamline the deployment of containerized workloads on hypervisor 201.
  • initi lization script 2 2 may interact with container registry 246 available in storage 240.
  • Container registry' 246 may be a repository-, or a collection of repositories, used to store and access container images.
  • container registry 246 is illustrated as being stored in storage 240 with workloads manifest 242 and infrastructure manifest 244, in certain other aspects, container registry may be stored separately from one or both of these manifests.
  • Fig. 3 is a flow diagram illustrating example operations 300 for automatically deploying containerized workloads according to aspects of the present disclosure.
  • operations 300 provide a method which eliminates the need for human intervention when deploying containerized workloads, for example, on one or more worker node VMs deployed on a hypervisor.
  • Operations 300 may be performed by components illustrated in Fig. 2.
  • operations 300 begin at block 302, by booting a device, such as host 203.
  • the device may be booted from a universal serial bus (USB) thumb drive, a synchronous dynamic random access memory (SD-RAM) car, a preboot execution environment (PXE), or the like to load hypervisor 201.
  • USB universal serial bus
  • SD-RAM synchronous dynamic random access memory
  • PXE preboot execution environment
  • operations 300 proceed with running an initialization script.
  • instructions included in initialization script 232 illustrated in Fig. 2 are executed to initiate automatic deployment of containerized workloads and their necessary infrastructure on the device.
  • operations 300 proceed with determining one or more intended infrastructure configuration parameters based on infrastructure manifest 244.
  • the hypervisor 201 based on instructions in initialization script 232, pulls information from infrastructure manifest 244 stored on storage 240 to configure control plane 206, as well as other infrastructure necessary to run containerized workloads on the device.
  • operations 300 proceed with determining a number of worker node VMs to deploy based, at least in part, on the one or more intended infrastructure configuration parameters.
  • the infrastructure state controller 214 may indicate a number of worker node VMs 224 needed to run particular workloads.
  • control plane pod 212 is deployed based, at least in part, on the one or more intended infrastructure configuration parameters. As such, control plane pod 212 is brought up in a configured state. For example, as illustrated in Fig. 2, control plane pod 212 having infrastructure state controller 214 for managing a state of the infrastructure is deployed. In certain other aspects, control plane pod 212 is brought up in a default state and then later configured.
  • operations 300 proceed with deploying one or more worker node VMs 224 in worker plane 210 based on the determined number of worker node VMs 224 to deploy.
  • operations 300 proceed with deploying an extension plane based, at least in part, on the one or more intended infrastructure configuration parameters.
  • extension plane 208 including a runtime controller for worker nodes 216, an admin worker pod 220, and GitOps agents 218 may be deployed.
  • operations 300 proceed with determining one or more workloads to be instantiated on the one or more work node VMs 224 based on workloads manifest 242.
  • GitOps agents 218 pull information from workloads manifest 242 stored on storage 240.
  • operations 300 proceed with instantiating the one or more workloads on the one or more work node VMs 224.
  • the deployed system may be enabled to run and manage containerized workloads.
  • the deployed system may manage a state of the infrastructure such that it aligns with an intended state of the infrastructure.
  • the deployed system may manage the containerized workloads such that they align with their intended states. Management of each of these states may be separated.
  • Fig. 4 is a flow diagram illustrating example operations 400 for managing different states of a deployed system according to aspects of the present disclosure. Operations 400 may be performed by one or more components illustrated in Fig. 2, and more specifically, infrastructure state controller 214 in control plane 206 and runtime controller for worker nodes 216 in extension plane 208.
  • operations 400 include, at block 402, an infrastructure state controller (e.g., infrastructure state controller 214) in the control plane managing a state of the control plane.
  • operations 400 further include a runtime controller (e.g., runtime controller for worker nodes 216) in the extension plane managing a state of one or more worker node VMs.
  • operations of block 404 are illustrated as occurring subsequent to operations of block 402, in certain aspects, operations of blocks 402 and 404 may be performed concurrently such that the state of the entire system (e.g., including both the state of the infrastructure/control plane and the state of the worker node VMs) is continuously being managed.
  • the state of the entire system e.g., including both the state of the infrastructure/control plane and the state of the worker node VMs
  • FIG. 5 is a flow diagram illustrating example operations 500 for automatically deploying a containerized workload on a hypervisor based device according to aspects of the present disclosure.
  • operations 500 begin at block 502, by booting the device running a hypervisor.
  • operations 500 proceed with, in response to booting the device, automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload.
  • the control plane configuration for providing services for deploying and managing the containerized workload is a control plane configuration for Keswick Node 204 illustrated in Fig. 2.
  • operations 500 proceed with deploying a control plane pod configured according to the control plane configuration.
  • operations 500 proceed with deploying one or more worker nodes based on the control plane configuration.
  • operations 500 proceed with deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
  • One or more embodiments of the invention also relate to a device or an apparatus for performing these operations.
  • the apparatus may be specially constructed for required purposes, or the apparatus may be a general-purpose computer selectively activated or configured by a computer program stored in the computer.
  • Various general -purpose machines may be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
  • One or more embodiments of the present invention may be implemented as one or more computer programs or as one or more computer program modules embodied in computer readable media.
  • the term computer readable medium refers to any data storage device that can store data which can thereafter be input to a computer system.
  • Computer readable media may be based on any existing or subsequently developed technology that embodies computer programs in a manner that enables a computer to read the programs. Examples of computer readable media are hard drives, NAS systems, read-only memory (ROM), RAM, compact disks (CDs), digital versatile disks (DVDs), magnetic tapes, and other optical and non-optical data storage devices.
  • a computer readable medium can also be distributed over a network-coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
  • Virtualization systems in accordance with the various embodiments may be implemented as hosted embodiments, non-hosted embodiments, or as embodiments that blur distinctions between the two.
  • various virtualization operations may be wholly or partially implemented in hardware.
  • a hardware implementation may employ a look-up table for modification of storage access requests to secure non-disk data.
  • the virtualization software can therefore include components of a host, console, or guest OS that perform virtualization functions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Stored Programmes (AREA)

Abstract

An example method of automatically deploying a containerized workload on a hypervisor based device is provided. The method generally includes booting the device running a hypervisor, in response to booting the device: automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload; deploying a control plane pod configured according to the control plane configuration; deploying one or more worker nodes based on the control plane configuration, and deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.

Description

OPTIMIZED SYSTEM DESIGN FOR DEPLOYING AND MANAGING CONTAINERIZED WORKLOADS AT SCALE
Inventors: Erol Aygar, Nikhil Deshpande, Benjamin John Corrie, George Gregory Hicken, Elliott Davis
Cross-Reference to Related Applications
[0001] This application claims priority to U.S. Provisional Patent Application No. 63/403,267, filed September 1, 2022, entitled “OPTIMIZED SYSTEM DESIGN FOR DEPLOYING AND MANAGING CONTAINERIZED WORKLOADS AT SCALE,” and U.S. Non-Provisional Patent Application No. 18/157,196, filed January 20, 2023, entitled “OPTIMIZED SYSTEM DESIGN FOR DEPLOYING AND MANAGING CONTAINERIZED WORKLOADS AT SCALE,” both assigned to the assignee hereof, the contents of which are hereby incorporated by reference in their entirety.
Background
[0002] Modem applications are applications designed to take advantage of the benefits of modem computing platforms and infrastructure. For example, modern applications can be deployed in a multi-cloud or hybrid cloud fashion, such as, consuming both cloud services executing in a public cloud and local services executing in a private data center (e.g., a private cloud). Within the public cloud or private data center, modem applications can be deployed onto one or more virtual machines (VMs), containers, application services, and/or the like.
[0003] A container is a package that relies on virtual isolation to deploy and run applications that access a shared operating system (OS) kernel. Containerized applications, also referred to as containerized workloads, can include a collection of one or more related applications packaged into one or more groups of containers, referred to as pods, that can be deployed based on the context defined in a manifest for the containerized workloads. Each pod may include one or more containers.
[0004] Containerized workloads run on a container orchestration platform that enables the automation of much of the operational effort required to run containers having workloads and services. This operational effort includes a wide range of things needed to manage a container’s lifecycle, including, but not limited to, provisioning, deployment, scaling (up and down), networking, and load balancing.
[0005] Kubernetes® (K8S)® software is an example open-source container orchestration platform that automates the operation of such containerized workloads. Kubernetes software allows for distributed computing by running the pods of containerized workloads on a cluster of interconnected worker nodes (e.g., VMs or physical machines) that may scale vertically and/or horizontally over hybrid cloud topology.
[0006] While containers are the building-blocks that enable a scalable environment, containers are not the only part of the software stack that needs to scale. In particular, tools used to instantiate, manage, monitor, and/or secure containers may also need to be able to scale as seamlessly as the containers. In other words, scalability of virtualization software and architecture for deploying and managing containerized workloads may also affect an ability of the environment to handle an increased or expanding workload.
[0007] For example, a software-defined data center (SDDC) includes clusters of physical servers (e.g., hosts) that are virtualized and managed by virtualization management servers. A host can include a virtualization layer (e.g., a hypervisor) that provides a software abstraction of the hardware platform of the physical server (e.g., central processing unit (CPU), random access memory (RAM), storage, network interface card (NIC), etc.) to allow multiple virtual computing instances (e.g., such as VMs) to run thereon. A control plane for each cluster of hosts may support the deployment and management of applications (or services) on the cluster using containers. In some cases, the control plane deploys applications as pods of containers running on one or more worker nodes. Accordingly, scalability of the environment may also depend on the operational effort required to initialize the control plane and/or worker nodes for container deployment, as well as an ability of the control plane to determine a context defined in a manifest for instantiation of workloads in these containers.
[0008] It should be noted that the information included in the Background section herein is simply meant to provide a reference for the discussion of certain embodiments in the Detailed Description. None of the information included in this Background should be considered as an admission of prior art.
Summary
[0009] One or more embodiments provide a method of automatically deploying a containerized workload on a hypervisor based device. The method generally includes booting the device running a hypervisor, in response to booting the device: automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload; deploying a control plane pod configured according to the control plane configuration; deploying one or more worker nodes based on the control plane configuration, and deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
[0010] Further embodiments include a non-transitory computer-readable storage medium comprising instructions that cause a computer system to carry out the above methods, as well as a computer system configured to carry out the above methods.
Brief Description of the Drawings
[0011] Fig. 1 is a block diagram of an exemplary container-based cluster according to aspects of the present disclosure.
[0012] Fig. 2 is a block diagram of an example system for deploying and managing distributed containerized workloads at scale according to aspects of the present disclosure.
[0013] Fig. 3 is a flow diagram illustrating example operations for automatically deploying containerized workloads according to aspects of the present disclosure.
[0014] Fig. 4 is a flow diagram illustrating example operations for managing different states of a deployed system according to aspects of the present disclosure.
[0015] Fig. 5 shows a flow diagram depicting a method of automatically deploying a containerized workload on a hypervisor based device according to an embodiment.
Detailed Description
[0016] Aspects of the present disclosure provide an auto-deployment system for automatically deploying and/or managing containerized workloads at scale. In certain aspects, the containerized workloads may be distributed. Automated deployment refers to the automation of steps, processes, and/or activities that are necessary to make a system and/or update available to its intended users. As described in more detail herein, automation of much of the operational effort required to set up and/or manage a system capable of supporting containerized workloads is based on an ability of the system to access one or more intended state configuration files made up of one or more manifests that declare intended system infrastructure and workloads to be deployed in the system. In certain aspects, the manifests are JavaScript Object Notation (JSON) and/or YAML files.
[0017] For example, deployment of the system to a device may be triggered by a user booting a device, such as from a USB drive, flash drive, SD-RAM card, preboot execution environment (PXE), etc. The device may execute an initialization script that causes a hypervisor to be installed on the device. The hypervisor may be a type-1 hypervisor that includes and integrates operating system (OS) components, such as a kernel, security management, file, process, memory, and/or input/output (VO) device management. [0018] According to aspects described herein, the installed hypervisor may include an infrastructure supervisor (infravisor) layer that provides a runtime for services for a cluster of hosts. In particular, the infravisor layer provides a cluster runtime for virtualization services and management, as further described herein. For example, the infravisor layer may be configured to provide a control plane service for managing the cluster of hosts, including configuring the cluster of hosts to run according to a defined intended state.
[0019] For example, a control plane and one or more infravisor pods may be instantiated on the infravisor layer to provide a service for deploying and/or managing containerized workloads. The control plane may be created to deploy and automatically manage clusters of containerized applications such that they align with their intended states as defined in one or more intended state configuration files. Accordingly, when deployed, the control plane may be configured to (1) determine one or more intended infrastructure configuration parameters by accessing one or more of the intended state configuration files at an accessible external datastore and (2) build the infrastructure based on these determined parameters. As mentioned, the intended infrastructure configuration parameters may declare infrastructure necessary for the deployment of one or more applications, for example, a number of worker node virtual machines (VMs) necessary for deployment inside the infravisor layer, virtual network configurations, storage configurations, etc. Though certain aspects are described herein with respect to VMs as worker nodes, other suitable types of VCIs, and/or hardware nodes may similarly be used as worker nodes.
[0020] Further, in some cases, building the infrastructure involves deploying one or more infravisor pods on the infravisor layer. At least one of the pods may be used to (1) determine one or more workloads to be instantiated on the created infrastructure, by accessing one or more of the intended state configuration files at an accessible external datastore and (2) instantiate these workloads. In some cases, these workloads are instantiated on one or more worker node VMs running on the hypervisor.
[0021] As such, when booted, the system described herein is designed to automatically pull intended configuration parameters (e.g., from an external datastore) and apply these parameters to automatically configure an underlying virtualization infrastructure of the system and execute containerized workloads on the configured infrastructure. Accordingly, human interaction and/or intervention with the system may not be necessary, thereby resulting in a more streamlined deployment of containerized workloads which may help to decrease errors, increase speed of delivery, boost quality, minimize costs, and overall, simplify the deployment process. Further, the auto-deployed system described herein may be able to scale. For example, introducing automation into deployment of the containerized workloads and underlying infrastructures allows for a same system to be instantiated multiple times, and/or a much larger system (including multiple clusters, containers, applications, etc.) to be instantiated in the same way (e.g., without human intervention).
[0022] In certain aspects, the one or more intended state configuration files are stored in a repository on a version control system. One example version control system that may be configured and used in aspects described herein is GitHub made commercially available by GitHub, Inc. GitHub is a git repository hosting service which provides version control software capable of managing and tracking changes to software code. Accordingly, intended configuration parameters stored in the one or more intended sate configuration files may be easily updated and tracked when stored on the version control system. Any updates made to the intended state configuration files may be automatically pulled and used by the system for adjustment to help guarantee consistent realization of the intended state defined by the intended state configuration files.
[0023] In certain aspects, the one or more intended state configuration files include at least two intended state configuration files. A first file (referred to as the “infrastructure manifest”) may include one or more intended infrastructure configuration parameters, while the second file (referred to as the “workload manifest”) may include one or more intended workload definitions. By keeping the infrastructure manifest separate from the workload manifest, patching (e.g., identifying, acquiring, testing and installing patches, or code changes, that are intended to resolve functionality issues, improve security, and/or add features) of these manifests may be less complex and easier to manage, as changes to infrastructure configuration parameters are independent of changes to intended workload definitions due to the file separation. It should be noted that any number of intended state configuration files may be used, such as a single file including both intended infrastructure configuration parameters and intended workload definitions.
[0024] According to aspects described herein, the control plane deployed to automatically deploy and manage containerized workloads may be isolated from such workloads. In particular, in terms of infrastructure, the control plane may be isolated from a worker plane where worker node VMs are instantiated for running the containerized workloads. Further, a state of the control plane may be managed separate from a state of the worker node VMs in the work plane. For example, the state of the control plane is managed by a control plane pod within the control plane. The state of the worker node VMs in the worker plane are managed by a worker pod in the extension plane, which is separated from the control plane. Thus, the system described herein offers logical separation of the state of the control plane from the state of the worker node VMs, as well as separate management of each of the states. Isolation of the control plane from the workloads may result in improved redundancy, security, and/or scalability.
[0025] Turning now to Fig. 1, a block diagram of an exemplary container-based cluster is illustrated. It should be noted that the block diagram of Fig. 1 is a logical representation of a container-based cluster, and does not show where the various components are implemented and run on physical systems. Fig. 2 provides an example system for implementing and running a container-based cluster of Fig. 1. While the example container-based cluster shown in Fig.
1 is a Kubemetes cluster 100, in other examples, the container-based cluster may be another type of container-based cluster based on container technology, such as Docker® clusters.
[0026] When Kubernetes is used to deploy applications, a cluster, such as K8S cluster 100 illustrated in FIG. 1, is formed from a combination of worker nodes 104 and a control plane 102. Worker nodes 104 are managed by control plane 102, which manages the computation, storage, and memory resources to run all worker nodes 104.
[0027] Each worker node 104, or worker compute machine, includes a kubelet 106, which is an agent that ensures that one or more pods 110 run in the worker node 104 according to a defined specification for the pods, such as defined in a workload definition manifest. Each pod 110 may include one or more containers 112. The worker nodes 104 can be used to execute various applications and software processes using container 112. Further each worker node 104 includes a kube proxy 108. Kube proxy 108 is a Kubemetes network proxy that maintains network rules on worker nodes 104. These network rules allow for network communication to pods 110 from network sessions inside and/or outside of K8S cluster 100.
[0028] Control plane 102 includes components such as an application programming interface (API) server 114, a cluster store (etcd) 116, a controller 118, and a scheduler 120. Control plane 102’s components make global decisions about K8S cluster 100 (e.g., scheduling), as well as detect and respond to cluster events (e.g., starting up a new pod 110 when a workload deployment's replicas field is unsatisfied).
[0029] API server 114 operates as a gateway to K8S cluster 100. As such, a command line interface, web user interface, users, and/or services communicate with K8S cluster 100 through API server 114. One example of a Kubemetes API server 114 is kube-apiserver. kube- apiserver is designed to scale horizontally — that is, this component scales by deploying more instances. Several instances of kube-apiserver may be run, and traffic may be balanced between those instances. [0030] Cluster store (etcd) 116 is a data store, such as a consistent and highly-available key value store, used as a backing store for all K8S cluster 100 data.
[0031] Controller 118 is a control plane 102 component that runs and manages controller processes in K8S cluster 100. For example, control plane 102 may have (e.g., four) control loops called controller processes, which watch the state of cluster 100 and try to modify the current state of cluster 100 to match a intended state of cluster 100. In certain aspects, controller processes of controller 118 are configured to monitor external storage for changes to the state of cluster 100.
[0032] Scheduler 120 is a control plane 102 component configured to allocate new pods 110 to worker nodes 104. Additionally, scheduler 118 may be configured to distribute resources and/or workloads across worker nodes 104. Resources may refer to processor resources, memory resources, networking resources, and/or the like. Schedule 118 may watch worker nodes 104 for how well each worker node 104 is handling their workload, and match available resources to the worker nodes 104. Scheduler 118 may then schedule newly created containers 112 to one or more of the worker nodes 104.
[0033] In other words, control plane 102 manages and controls every component of the cluster 100. Control plane 102 handles most, if not all, operations within cluster 100, and its components define and control cluster 100’s configuration and state data. Control plane 102 configures and runs the deployment, management, and maintenance of the containerized applications. As such, ensuring high availability of the control plane may be critical to container deployment and management. High availability is a characteristic of a component or system that is capable of operating continuously without failing.
[0034] Accordingly, in certain aspects, control plane 102 may operate as a high availability (HA) control plane. Additional details of HA control planes are disclosed in U.S. Application Serial No. 63/347,815, filed on June 1, 2022, and titled “AUTONOMOUS CLUSTERS IN A VIRTUALIZATION COMPUTING ENVIRONMENT,” which is hereby incorporated by reference herein in its entirety.
[0035] As mentioned, while container orchestration platforms, such as Kubernetes, provide automation to deploy and run clusters of containerized applications (e.g., such as K8S cluster 100 illustrated in Fig. 1), thereby allowing for easy scalability of containers based on application requirements, containers are not the only part of the software stack that needs to be automated for scaling. For example, deployment of a control plane (e.g., such as control plane 102 illustrated in Fig. 1) may be necessary for the deployment and management of such containerized workloads. [0036] Accordingly, as discussed, aspects of the present disclosure provide a system for fully automatically deploying and/or managing distributed containerized workloads at scale. Architecture of the auto-deployed system described herein includes an improved system designed for deploying and managing distributed containerized workloads at scale. In particular, the system described herein includes one or more components configured to automatically configure and deploy a control plane for managing containerized workloads. The system further includes one or more components to determine a number of worker nodes needed for deployment of intended workloads, deploy the determined number of worker nodes, and instantiate the intended workloads on the deployed worker nodes. Further, the system design described herein offers physical separation of the state of a control plane used to manage deployed workloads from the state of worker nodes where the workloads are deployed.
[0037] Fig. 2 is a block diagram of an auto-deployed system 200 for deploying and managing distributed containerized workloads at scale according to aspects of the present disclosure. The architecture of system 200 described in Fig. 2 may allow for implementation of a container-based cluster on a hypervisor. For example, a logical construct of K8S cluster 100 illustrated in Fig. 1 may be implemented by the architecture described below with reference to Fig. 2. Further, such architecture may bring along with it a new management model, as described in more detail below.
[0038] As mentioned, a hypervisor is a type of virtualization software that supports the creation and management of virtual endpoints by separating a physical machine’s software from its hardware. In other words, hypervisors translate requests between physical and virtual resources, thereby making virtualization possible. When a hypervisor is installed directly on the hardware of a physical machine, as opposed to on top of an operating system (OS) of the machine, the hypervisor is referred to as bare-metal hypervisor. In certain aspects, hypervisor 201 illustrated in Fig. 2 is a bare-metal hypervisor. As shown, a host 203 runs hypervisor 201. Host 203 is a physical computing device, such as a server-grade computer, that includes hardware components, such as a memory, processor, storage, networking card, and/or the like, for running components described herein.
[0039] In certain aspects, a user interface (not shown) may be provided to enable users to interact with hypervisor 201, such as to check on system status, update configuration, etc. The user interface may be accessible by directly accessing host 203, or by accessing host 203 over a network, such as via machine web browser or API client. For example, hypervisor 201 may include a host daemon 230 running as a background process, which in part allows connection to hypervisor 201 for monitoring hypervisor 201. [0040] In certain aspects, hypervisor 201 is a multi-layer entity, where each layer is provided a different level of privilege. In particular, hypervisor 201 architecture may include underlying OS features, referred to as a kernel, and processes that run on top of the kernel. The kernel may be a microkernel that provides functions such as process creation, process control, process threads, signals, file system, etc. A process running on or above the kernel may be referred to as a "user world" process. A user world process may run in a limited environment. A privilege level of the kernel may be greater than a privilege level of a user world process.
[0041] Hypervisor 201, as part of an infravisor layer, may include an infravisor daemon 228 running as a background process. In certain aspects, infravisor daemon 228 is an infravisor watchdog running on hypervisor 201. The infravisor daemon 228 is configured to monitor individual infravisor services (e.g., including an infravisor runtime pod 226, described in detail below) running in a cluster of hosts to help guarantee that a minimum number of individual services are continuously running in the cluster. In certain aspects, infravisor daemon 228 monitors an API server (e.g., such as API server 114 illustrated in Fig. 1) to determine whether a minimum number of individual services are running.
[0042] Hypervisor 201, as part of an infravisor layer, may further include an infravisor runtime pod 226, which may be a pod of containers running on the hypervisor 201 that execute control plane entities, such as API server 114, cluster store (etcd) 116, controller 118, and scheduler 120 illustrated in Fig. 1, for a cluster of hosts. These components may run as separate or consolidated pods for isolation or footprint reduction, respectively. The infravisor runtime pod 226 may access the cluster store (etcd) 116 to store a cluster’s runtime state. In certain aspects, the infravisor runtime pod 226 is bootstrapped on a host in a cluster when the infravisor daemon 228 detects the absence of a functional infravisor runtime pod. It should be noted that such control plane functionality provided by infravisor runtime pod 226 may be separate from the control plane 206 described herein for worker nodes. In particular, while infravisor runtime pod 226 and a control plane pod 212 of control plane 206 may execute similar control plane/runtime entities, in certain aspects, the infravisor runtime pod 226 runs in a higher privilege level than control plane pod 212 of control plane 206. Further, while infravisor runtime pod 226 may manage at least part of the lifecycle of pods or services for running containerized workloads, control plane pod 212 of control plane 206 manages the runtime state of such containerized workloads, and the infrastructure necessary for implementation.
[0043] Hypervisor 201 provides resources of host 203 to run one or more pods or services, collectively referred to as a Keswick node 204, which is a logical abstraction of the one or more pods or services. (The term, “Keswick” is an arbitrary name given to the abstraction for purpose of easy reference.) The pods and services of the Keswick node 204 are logically separated by function into a control plane 206, an extension plane 208, and a worker plane 210, which are used to provide services for deploying and/or managing containerized workloads.
[0044] Control plane 206 includes a control plane pod 212, which may be a pod of containers running on the hypervisor 201 that execute control plane entities, such as API server 114, cluster store (etcd) 116, controller 118, and scheduler 120 illustrated in Fig. 1, for worker nodes 224. The control plane pod 212 runs an infrastructure state controller 214 configured to manage the state of control plane 206, such as a number of worker nodes 224 to run, networking configuration of such worker nodes 224, etc. In certain aspects, control plane 206 is configured based on infrastructure manifest 244 stored in storage 240.
[0045] Infrastructure manifest 244 provides information about intended system infrastructure to be deployed on host 203. For example, infrastructure manifest 244 may define the infrastructure on which containerized workloads are expected to run. This may include information about a number of worker VMs 224 to instantiate, assignment of hardware resources to worker VMs 224, software configuration (e.g., a version of Kubemetes an application/workload uses), and/or network infrastructure (e.g., a software defined network). As an illustrative example, the infrastructure manifest 244 may indicate a number of worker node VMs to deploy on hypervisor 201 and, in some cases, images to use for instantiating each of these worker node VMs. The number of worker node VMs indicated in infrastructure manifest 244 may be a number of worker node VMs needed to run particular workloads defined in a workloads manifest 242.
[0046] In certain aspects, infrastructure manifest 244 is included in an intended state configuration file. In certain aspects, the intended state configuration file may include one or more other manifests (e.g., such as workloads manifest 242). The intended state configuration file may be stored in storage 240, which may be an external storage that is accessible by hypervisor 201. Storage 240 may further be accessible by infrastructure state controller 214 of control plane 206 after the control plane is instantiated, such as to monitor for updates to the infrastructure manifest 244 and automatically update the configuration of control plane 206, accordingly. In certain aspects, storage 240 is a repository on a version control system. As mentioned previously, one example version control system that may be configured and used in aspects described herein is GitHub made commercially available by GitHub, Inc.
[0047] As such, hypervisor 201 may be configured to pull information from infrastructure manifest 244 and use this information to instantiate and configure control plane 206, such as by instantiating and configuring control plane pod 212. In certain aspects, this involves instantiating worker plane 210 by deploying one or more worker node VMs 224 in worker plane 210. A number of worker node VMs 224 deployed in worker plane 210 may be based, at least in part, on a number of work node VMs indicated for deployment in infrastructure manifest 244.
[0048] Infrastructure state controller 214 on control plane 206 is configured to manage a state of the infrastructure. In other words, infrastructure state controller 214 accepts an “intended state” (also referred to as “desired state” or “declared state”) from a human operator (e.g., via infrastructure manifest 244), observes the state of the infrastructure, and dynamically configures the infrastructure such that the infrastructure matches the “intended state.” Accordingly, infrastructure state controller 214 may also be configured to interact with infrastructure manifest 244 stored in storage 240.
[0049] Further, in certain aspects, infrastructure state controller 214 monitors storage 240 for changes/updates to infrastructure manifest 244. Infrastructure state controller 214 may be configured to dynamically update the infrastructure such that the infrastructure matches a new “intended state” defined by infrastructure manifest 244, for example, when infrastructure state controller 214 determines infrastructure manifest 244 has been updated.
[0050] Worker node VMs 224 deployed in worker plane 210 are compute resources that use software to run programs and deploy applications/workloads. More specifically, worker node VMs 224 may be used to deploy containerized workloads on hypervisor 201. Worker node VMs 224 deployed in worker plane 210 may each include a cluster agent 222. A cluster agent 222 may be a container or a pod within a worker node VM 224. In certain aspects, cluster agent 222 is configured to monitor the health of a container-based cluster supported via worker node VM 224. Further, in certain aspects, cluster agent 222 is configured to collect metrics and metadata for a container-based cluster deployed on worker node VM 224, including each node and namespace down to the container level. As shown in Fig. 2, worker plane 210 is isolated from control plane 206; thus, worker node VMs 224 and their corresponding cluster agents 222 are similarly isolated from control plane 206. Worker plane 210 describes resources managed by hypervisor 201 dedicated for running VMs. These resources may be distinct from resources used to run control plane 206 and extension plane 208.
[0051] Extension plane 208 includes a runtime controller for worker nodes 216 and an admin worker pod 220 which includes GitOps agents 218. In certain aspects, GitOps agents 218 are configured to interact with workloads manifest 242 stored in storage 240. [0052] Workloads manifest 242 provides information about intended workloads to be deployed in hypervisor 201. For example, workloads manifest 242 may outline details of one or more workloads to be deployed in worker node VMs 224 in worker plane 210 on hypervisor 201. In particular, in certain aspects, workloads manifest 242 includes an identifier of a binary to be loaded. In certain aspects, workloads manifest 242 includes information about resources to be deployed, workload parameters associated with these resources, and/or protected resources for one or more workloads. The workload parameters may include a workload name, a workload ID, a service name, an associated organization ID, and/or the like.
[0053] In certain aspects, workloads manifest 242 is included in an intended state configuration file. In some cases, the intended state configuration file may include one or more other manifests (e.g., such as infrastructure manifest 244). The intended state configuration file may be stored in storage 240 which is external storage that is accessible by GitOps agents 218.
[0054] As such, GitOps agents 218 may be configured to pull information from workloads manifest 242 and use this information to instantiate workloads on worker node VMs 224 running in worker plane 210 (e.g., previously deployed by control plane 206).
[0055] Runtime controller for worker nodes 216 is configured to manage a state of the worker node VMs 224. In other words, runtime controller for worker nodes 216 accepts a “intended state” (also referred to as “intended stated” or “declared state”) from a human operator (e.g., via workloads manifest 242), observes the state of the worker node VMs 224, and dynamically configures the worker node VMs 224 such that their behavior matches the “intended state.”
[0056] Further, in certain aspects, runtime controller for worker nodes 216 monitors storage 240 for changes/updates to workloads manifest 242. Runtime controller for worker nodes 216 may be configured to dynamically update the state of the worker node VMs 224 to match a new “intended state” defined by workloads manifest 244, for example, when runtime controller for work nodes 216 determines workloads manifest 244 has been updated.
[0057] As mentioned, in certain aspects, privilege becomes diluted when moving from bottom to top layers of hypervisor 201. As such, in certain aspects, the infravisor layer of hypervisor 201 is at a lower, more privileged level of hypervisor 201, while control plane 206 and extension plane 208 are at a lesser-privileged level in hypervisor 201. Additionally, the worker node VMs 224 running in worker plane 210 may be on top of hypervisor 201, as this is where the deployed workloads are expected to run. [0058] Further, in addition or alternative to different privilege levels, defined management levels may be assigned to different entities. For example, in certain aspects, worker node VMs 224 are managed by control plane pod 212 of control plane 206, and the control plane pod 212 is managed by the infravisor layer of hypervisor 201.
[0059] It should be noted that Keswick Node 204 is a logical abstraction that represents the control plane 206, extension plane 208, and worker plane 210. Though certain example implementations are described herein of how each of the control plane 206, extension plane 208, and worker plane 210 are implemented (e.g., as pods, VMs, etc.) and where they run (e.g., in hypervisor 201, on top of hypervisor 201, etc.), it should be noted that other implementations may be possible, such as having certain components run in different privilege levels, layers, within hypervisor 201, outside hypervisor 201, etc.
[0060] Further, as shown in Fig. 2, hypervisor 201 includes an initialization script 232. Initialization script 232 is a sequence of instructions that are interpreted or carried out during startup of hypervisor 201. Initialization script 232 helps to automate the deployment of hypervisor 201 and Keswick node 204. In other words, initialization script 232 helps to automate and streamline the deployment of containerized workloads on hypervisor 201.
[0061] In certain aspects, initi lization script 2 2 may interact with container registry 246 available in storage 240. Container registry' 246 may be a repository-, or a collection of repositories, used to store and access container images. Although container registry 246 is illustrated as being stored in storage 240 with workloads manifest 242 and infrastructure manifest 244, in certain other aspects, container registry may be stored separately from one or both of these manifests.
[0062] Fig. 3 is a flow diagram illustrating example operations 300 for automatically deploying containerized workloads according to aspects of the present disclosure. In particular, operations 300 provide a method which eliminates the need for human intervention when deploying containerized workloads, for example, on one or more worker node VMs deployed on a hypervisor. Operations 300 may be performed by components illustrated in Fig. 2.
[0063] As illustrated, operations 300 begin at block 302, by booting a device, such as host 203. The device may be booted from a universal serial bus (USB) thumb drive, a synchronous dynamic random access memory (SD-RAM) car, a preboot execution environment (PXE), or the like to load hypervisor 201.
[0064] At block 304, operations 300 proceed with running an initialization script. For example, instructions included in initialization script 232 illustrated in Fig. 2 are executed to initiate automatic deployment of containerized workloads and their necessary infrastructure on the device.
[0065] At block 306, operations 300 proceed with determining one or more intended infrastructure configuration parameters based on infrastructure manifest 244. For example, the hypervisor 201, based on instructions in initialization script 232, pulls information from infrastructure manifest 244 stored on storage 240 to configure control plane 206, as well as other infrastructure necessary to run containerized workloads on the device.
[0066] At block 308, operations 300 proceed with determining a number of worker node VMs to deploy based, at least in part, on the one or more intended infrastructure configuration parameters. In particular, the infrastructure state controller 214 may indicate a number of worker node VMs 224 needed to run particular workloads.
[0067] At block 310, operations 300 proceed with deploying the control plane pod 212. In certain aspects, control plane pod 212 is deployed based, at least in part, on the one or more intended infrastructure configuration parameters. As such, control plane pod 212 is brought up in a configured state. For example, as illustrated in Fig. 2, control plane pod 212 having infrastructure state controller 214 for managing a state of the infrastructure is deployed. In certain other aspects, control plane pod 212 is brought up in a default state and then later configured.
[0068] At block 312, operations 300 proceed with deploying one or more worker node VMs 224 in worker plane 210 based on the determined number of worker node VMs 224 to deploy.
[0069] At block 314, operations 300 proceed with deploying an extension plane based, at least in part, on the one or more intended infrastructure configuration parameters. For example, as illustrated in Fig. 2, extension plane 208 including a runtime controller for worker nodes 216, an admin worker pod 220, and GitOps agents 218 may be deployed.
[0070] At block 316, operations 300 proceed with determining one or more workloads to be instantiated on the one or more work node VMs 224 based on workloads manifest 242. For example, GitOps agents 218 pull information from workloads manifest 242 stored on storage 240.
[0071] At block 318, operations 300 proceed with instantiating the one or more workloads on the one or more work node VMs 224.
[0072] Subsequent to block 320, the deployed system may be enabled to run and manage containerized workloads. In particular, the deployed system may manage a state of the infrastructure such that it aligns with an intended state of the infrastructure. Further, the deployed system may manage the containerized workloads such that they align with their intended states. Management of each of these states may be separated.
[0073] Fig. 4 is a flow diagram illustrating example operations 400 for managing different states of a deployed system according to aspects of the present disclosure. Operations 400 may be performed by one or more components illustrated in Fig. 2, and more specifically, infrastructure state controller 214 in control plane 206 and runtime controller for worker nodes 216 in extension plane 208.
[0074] As illustrated, operations 400 include, at block 402, an infrastructure state controller (e.g., infrastructure state controller 214) in the control plane managing a state of the control plane. At block 404, operations 400 further include a runtime controller (e.g., runtime controller for worker nodes 216) in the extension plane managing a state of one or more worker node VMs.
[0075] Although operations of block 404 are illustrated as occurring subsequent to operations of block 402, in certain aspects, operations of blocks 402 and 404 may be performed concurrently such that the state of the entire system (e.g., including both the state of the infrastructure/control plane and the state of the worker node VMs) is continuously being managed.
[0076] Fig. 5 is a flow diagram illustrating example operations 500 for automatically deploying a containerized workload on a hypervisor based device according to aspects of the present disclosure.
[0077] As illustrated, operations 500 begin at block 502, by booting the device running a hypervisor.
[0078] At block 504, operations 500 proceed with, in response to booting the device, automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload. In certain aspects, the control plane configuration for providing services for deploying and managing the containerized workload is a control plane configuration for Keswick Node 204 illustrated in Fig. 2.
[0079] At block 506, operations 500 proceed with deploying a control plane pod configured according to the control plane configuration.
[0080] At block 508, operations 500 proceed with deploying one or more worker nodes based on the control plane configuration. [0081] At block 510, operations 500 proceed with deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
[0082] One or more embodiments of the invention also relate to a device or an apparatus for performing these operations. The apparatus may be specially constructed for required purposes, or the apparatus may be a general-purpose computer selectively activated or configured by a computer program stored in the computer. Various general -purpose machines may be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
[0083] The embodiments described herein may be practiced with other computer system configurations including hand-held devices, microprocessor systems, microprocessorbased or programmable consumer electronics, minicomputers, mainframe computers, etc.
[0084] One or more embodiments of the present invention may be implemented as one or more computer programs or as one or more computer program modules embodied in computer readable media. The term computer readable medium refers to any data storage device that can store data which can thereafter be input to a computer system. Computer readable media may be based on any existing or subsequently developed technology that embodies computer programs in a manner that enables a computer to read the programs. Examples of computer readable media are hard drives, NAS systems, read-only memory (ROM), RAM, compact disks (CDs), digital versatile disks (DVDs), magnetic tapes, and other optical and non-optical data storage devices. A computer readable medium can also be distributed over a network-coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
[0085] Although one or more embodiments of the present invention have been described in some detail for clarity of understanding, certain changes may be made within the scope of the claims. Accordingly, the described embodiments are to be considered as illustrative and not restrictive, and the scope of the claims is not to be limited to details given herein but may be modified within the scope and equivalents of the claims. In the claims, elements and/or steps do not imply any particular order of operation unless explicitly stated in the claims.
[0086] Virtualization systems in accordance with the various embodiments may be implemented as hosted embodiments, non-hosted embodiments, or as embodiments that blur distinctions between the two. Furthermore, various virtualization operations may be wholly or partially implemented in hardware. For example, a hardware implementation may employ a look-up table for modification of storage access requests to secure non-disk data.
[0087] Many variations, additions, and improvements are possible, regardless of the degree of virtualization. The virtualization software can therefore include components of a host, console, or guest OS that perform virtualization functions.
[0088] Plural instances may be provided for components, operations, or structures described herein as a single instance. Boundaries between components, operations, and data stores are somewhat arbitrary, and particular operations are illustrated in the context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within the scope of the invention. In general, structures and functionalities presented as separate components in exemplary configurations may be implemented as a combined structure or component. Similarly, structures and functionalities presented as a single component may be implemented as separate components. These and other variations, additions, and improvements may fall within the scope of the appended claims.

Claims

What is claimed is:
1. A method of automatically deploying a containerized workload on a hypervisor based device, the method comprising: booting the device running a hypervisor; in response to booting the device: automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload; deploying a control plane pod configured according to the control plane configuration; deploying one or more worker nodes based on the control plane configuration; and deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
2. The method of claim 1, wherein: the control plane pod is deployed in a control plane; the one or more worker nodes are deployed in a worker plane; and the control plane and the worker plane are logically separated planes.
3. The method of claim 2, further comprising: deploying an infrastructure controller configured to manage a state of the control plane; and deploying a runtime controller for the one or more worker nodes configured to manage a state of the one or more worker nodes, wherein: the infrastructure controller is deployed in the control plane, the runtime controller is deployed in an extension plane, and the control plane and the extension plane are logically separated planes.
4. The method of claim 3, wherein the infrastructure controller is configured to manage the state of the control plane based on: monitoring for changes to the control plane configuration; and updating the state of the control plane based on detecting a change to the control plane configuration when monitoring for the changes to the control plane configuration.
5. The method of claim 3, wherein the runtime controller is configured to manage the state of the one or more worker nodes based on: monitoring for changes to the workload configuration parameters; and updating the state of the one or more worker nodes based on detecting a change to the workload configuration parameters when monitoring for the changes to the workload configuration parameters.
6. The method of claim 1, wherein: the device obtains two intended state configuration files, a first intended state configuration file of the two intended state configuration files defining the control plane configuration, and a second intended state configuration file of the two intended state configuration files defining the workload configuration parameters.
7. The method of claim 1, wherein the one or more worker nodes comprise virtual machines (VMs) deployed on the hypervisor.
8. A system comprising: one or more processors; and at least one memory, the one or more processors and the at least one memory configured to: boot a device running a hypervisor; in response to booting the device: automatically obtain, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing a containerized workload and workload configuration parameters for the containerized workload; deploy a control plane pod configured according to the control plane configuration; deploy one or more worker nodes based on the control plane configuration; and deploy one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
9. The system of claim 8, wherein: the control plane pod is deployed in a control plane; the one or more worker nodes are deployed in a worker plane; and the control plane and the worker plane are logically separated planes.
10. The system of claim 9, wherein the one or more processors and the at least one memory are further configured to: deploy an infrastructure controller configured to manage a state of the control plane; and deploy a runtime controller for the one or more worker nodes configured to manage a state of the one or more worker nodes, wherein: the infrastructure controller is deployed in the control plane, the runtime controller is deployed in an extension plane, and the control plane and the extension plane are logically separated planes.
11. The system of claim 10, wherein the infrastructure controller is configured to manage the state of the control plane based on: monitoring for changes to the control plane configuration; and updating the state of the control plane based on detecting a change to the control plane configuration when monitoring for the changes to the control plane configuration.
12. The system of claim 10, wherein the runtime controller is configured to manage the state of the one or more worker nodes based on: monitoring for changes to the workload configuration parameters; and updating the state of the one or more worker nodes based on detecting a change to the workload configuration parameters when monitoring for the changes to the workload configuration parameters.
13. The system of claim 8, wherein: the device obtains two intended state configuration files, a first intended state configuration file of the two intended state configuration files defining the control plane configuration, and a second intended state configuration file of the two intended state configuration files defining the workload configuration parameters.
14. The system of claim 8, wherein the one or more worker nodes comprise virtual machines (VMs) deployed on the hypervisor.
15. A non-transitory computer-readable medium comprising instructions that, when executed by one or more processors of a computing system, cause the computing system to perform operations for automatically deploying a containerized workload on a hypervisor based device, the operations comprising: booting the device running a hypervisor; in response to booting the device: automatically obtaining, by the device, one or more intended state configuration files from a server external to the device, the one or more intended state configuration files defining a control plane configuration for providing services for at least deploying and managing the containerized workload and workload configuration parameters for the containerized workload; deploying a control plane pod configured according to the control plane configuration; deploying one or more worker nodes based on the control plane configuration; and deploying one or more workloads identified by the workload configuration parameters on the one or more worker nodes.
16. The non-transitory computer-readable medium of claim 15, wherein: the control plane pod is deployed in a control plane; the one or more worker nodes are deployed in a worker plane; and the control plane and the worker plane are logically separated planes.
17. The non-transitory computer-readable medium of claim 16, wherein the operations further comprise: deploying an infrastructure controller configured to manage a state of the control plane; and deploying a runtime controller for the one or more worker nodes configured to manage a state of the one or more worker nodes, wherein: the infrastructure controller is deployed in the control plane, the runtime controller is deployed in an extension plane, and the control plane and the extension plane are logically separated planes.
18. The non-transitory computer-readable medium of claim 17, wherein the infrastructure controller is configured to manage the state of the control plane based on: monitoring for changes to the control plane configuration; and updating the state of the control plane based on detecting a change to the control plane configuration when monitoring for the changes to the control plane configuration.
19. The non-transitory computer-readable medium of claim 17, wherein the runtime controller is configured to manage the state of the one or more worker nodes based on: monitoring for changes to the workload configuration parameters; and updating the state of the one or more worker nodes based on detecting a change to the workload configuration parameters when monitoring for the changes to the workload configuration parameters.
20. The non-transitory computer-readable medium of claim 15, wherein: the device obtains two intended state configuration files, a first intended state configuration file of the two intended state configuration files defining the control plane configuration, and a second intended state configuration file of the two intended state configuration files defining the workload configuration parameters.
PCT/US2023/030300 2022-09-01 2023-08-15 Optimized system design for deploying and managing containerized workloads at scale WO2024049636A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202263403267P 2022-09-01 2022-09-01
US63/403,267 2022-09-01
US18/157,196 US20240078127A1 (en) 2022-09-01 2023-01-20 Optimized system design for deploying and managing containerized workloads at scale
US18/157,196 2023-01-20

Publications (1)

Publication Number Publication Date
WO2024049636A1 true WO2024049636A1 (en) 2024-03-07

Family

ID=87974301

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2023/030300 WO2024049636A1 (en) 2022-09-01 2023-08-15 Optimized system design for deploying and managing containerized workloads at scale

Country Status (1)

Country Link
WO (1) WO2024049636A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9983891B1 (en) * 2017-03-21 2018-05-29 Veritas Technologies Llc Systems and methods for distributing configuration templates with application containers
US10901966B2 (en) * 2016-08-29 2021-01-26 Vmware, Inc. Synchronizing configurations for container hosted applications
US20210271521A1 (en) * 2018-07-19 2021-09-02 Namu Tech Co., Ltd. Method for provisioning and managing multi-cluster on cloud platform

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10901966B2 (en) * 2016-08-29 2021-01-26 Vmware, Inc. Synchronizing configurations for container hosted applications
US9983891B1 (en) * 2017-03-21 2018-05-29 Veritas Technologies Llc Systems and methods for distributing configuration templates with application containers
US20210271521A1 (en) * 2018-07-19 2021-09-02 Namu Tech Co., Ltd. Method for provisioning and managing multi-cluster on cloud platform

Similar Documents

Publication Publication Date Title
US11822949B2 (en) Guest cluster deployed as virtual extension of management cluster in a virtualized computing system
KR102125260B1 (en) Integrated management system of distributed intelligence module
US11907742B2 (en) Software-defined network orchestration in a virtualized computer system
US10474488B2 (en) Configuration of a cluster of hosts in virtualized computing environments
US11556372B2 (en) Paravirtual storage layer for a container orchestrator in a virtualized computing system
US11604672B2 (en) Operational health of an integrated application orchestration and virtualized computing system
US20220004417A1 (en) Logical network platform install and upgrade in a virtualized computer system
US20220237049A1 (en) Affinity and anti-affinity with constraints for sets of resources and sets of domains in a virtualized and clustered computer system
US20230153145A1 (en) Pod deployment in a guest cluster executing as a virtual extension of management cluster in a virtualized computing system
US11720382B2 (en) Declarative VM management for a container orchestrator in a virtualized computing system
US20220237048A1 (en) Affinity and anti-affinity for sets of resources and sets of domains in a virtualized and clustered computer system
US20230021723A1 (en) Automated migration of monolithic applications to container platforms
US11900141B2 (en) Direct access storage for persistent services in a distributed storage system
US11263041B2 (en) Managing configuration and sensitive data for workloads in a virtualized computing system
US20240176639A1 (en) Diagnosing remote sites of a distributed container orchestration system
US20240028370A1 (en) Diagnosing remote sites of a distributed container orchestration system
US20230229483A1 (en) Fault-handling for autonomous cluster control plane in a virtualized computing system
US20230229482A1 (en) Autonomous cluster control plane in a virtualized computing system
US20220197684A1 (en) Monitoring for workloads managed by a container orchestrator in a virtualized computing system
US11593095B2 (en) Upgrade of a distributed service in a virtualized computing system
US20230229476A1 (en) Flexible infrastructure for provisioning virtual computing instances
US20240078127A1 (en) Optimized system design for deploying and managing containerized workloads at scale
US12106135B2 (en) Distributed autonomous lifecycle management of hypervisors in a virtualized computing system
US11720271B2 (en) Direct access storage for persistent services in a virtualized computing system
US20220229686A1 (en) Scheduling workloads in a container orchestrator of a virtualized computer system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23768054

Country of ref document: EP

Kind code of ref document: A1