<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>LWS</title><link>/</link><description>Recent content on LWS</description><generator>Hugo</generator><language>en</language><atom:link href="/index.xml" rel="self" type="application/rss+xml"/><item><title>Configure external cert-manager</title><link>/docs/manage/cert_manager/</link><pubDate>Mon, 28 Apr 2025 00:00:00 +0000</pubDate><guid>/docs/manage/cert_manager/</guid><description>&lt;p&gt;This page shows how you can a third party certificate authority solution like
Cert Manager.&lt;/p&gt;
&lt;h2 id="before-you-begin"&gt;Before you begin&lt;/h2&gt;
&lt;p&gt;Make sure the following conditions are set:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A Kubernetes cluster is running.&lt;/li&gt;
&lt;li&gt;The kubectl command-line tool has communication with your cluster.&lt;/li&gt;
&lt;li&gt;Cert Manager is &lt;a href="https://cert-manager.io/docs/installation/"&gt;installed&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;LWS supports either Kustomize or installation via a Helm chart.&lt;/p&gt;
&lt;h3 id="internal-certificate-management"&gt;Internal Certificate management&lt;/h3&gt;
&lt;p&gt;In all cases, LWS&amp;rsquo;s internal certificate management must be turned off
if one wants to use CertManager.&lt;/p&gt;</description></item><item><title>vLLM</title><link>/docs/examples/vllm/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/vllm/</guid><description>&lt;p&gt;In this example, we will use LeaderWorkerSet to deploy a distributed inference service with vLLM on GPUs.
&lt;a href="https://docs.vllm.ai/en/latest/index.html"&gt;vLLM&lt;/a&gt; supports distributed tensor-parallel inference and serving. Currently, it supports Megatron-LM’s tensor parallel algorithm. It manages the distributed runtime with &lt;a href="https://docs.ray.io/en/latest/index.html"&gt;Ray&lt;/a&gt;. See the doc &lt;a href="https://docs.vllm.ai/en/latest/serving/distributed_serving.html"&gt;vLLM Distributed Inference and Serving&lt;/a&gt; for more details.&lt;/p&gt;
&lt;h2 id="deploy-leaderworkerset-of-vllm"&gt;Deploy LeaderWorkerSet of vLLM&lt;/h2&gt;
&lt;p&gt;We use LeaderWorkerSet to deploy two vLLM model replicas. We have two flavors of the deployment:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;GPU: Each vLLM replica has 2 pods (&lt;code&gt;pipeline_parallel_size=2&lt;/code&gt;) and 8 GPUs per pod (&lt;code&gt;tensor_parallel_size=8&lt;/code&gt;).&lt;/li&gt;
&lt;li&gt;TPU: The example assumes that you have a GKE cluster with two TPU v5e-16 slices. You can view how to create a cluster with multiple TPU slices &lt;a href="https://cloud.google.com/kubernetes-engine/docs/how-to/tpus"&gt;here&lt;/a&gt;. Each TPU slice has 4 hosts, and each host has 4 TPUs. The vLLM server is deployed on the TPU slice with &lt;code&gt;pipeline_parallel_size=2&lt;/code&gt; and &lt;code&gt;tensor_parallel_size=16&lt;/code&gt;.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;In both examples, Ray uses the leader pod as the head node and the worker pods as the worker nodes. The leader pod runs the vLLM server, with a ClusterIP Service exposing the port.&lt;/p&gt;</description></item><item><title>Configure Prometheus</title><link>/docs/manage/prometheus/</link><pubDate>Mon, 28 Apr 2025 00:00:00 +0000</pubDate><guid>/docs/manage/prometheus/</guid><description>&lt;p&gt;This page shows how you configure LWS to use prometheus metrics.&lt;/p&gt;
&lt;h2 id="before-you-begin"&gt;Before you begin&lt;/h2&gt;
&lt;p&gt;Make sure you the following conditions are set:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A Kubernetes cluster is running.&lt;/li&gt;
&lt;li&gt;The kubectl command-line tool has communication with your cluster.&lt;/li&gt;
&lt;li&gt;Prometheus is &lt;a href="https://prometheus-operator.dev/docs/getting-started/installation/"&gt;installed&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;Cert Manager can be optionally &lt;a href="https://cert-manager.io/docs/installation/"&gt;installed&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;LWS supports either Kustomize or installation via a Helm chart.&lt;/p&gt;
&lt;h3 id="kustomize-installation"&gt;Kustomize Installation&lt;/h3&gt;
&lt;ol&gt;
&lt;li&gt;Enable &lt;code&gt;prometheus&lt;/code&gt; in &lt;code&gt;config/default/kustomization.yaml&lt;/code&gt; and uncomment all sections with &amp;lsquo;PROMETHEUS&amp;rsquo;.&lt;/li&gt;
&lt;/ol&gt;
&lt;h4 id="kustomize-prometheus-with-certificates"&gt;Kustomize Prometheus with certificates&lt;/h4&gt;
&lt;p&gt;If you want to enable TLS verification for the metrics endpoint, follow the directions below.&lt;/p&gt;</description></item><item><title>TensorRT-LLM</title><link>/docs/examples/tensorrt-llm/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/tensorrt-llm/</guid><description>&lt;p&gt;In this example, we will use LeaderWorkerSet to deploy a distributed inference service with Triton TensorRT-LLM on GPUs.
&lt;a href="https://nvidia.github.io/TensorRT-LLM/"&gt;TensorRT-LLM&lt;/a&gt; supports multinode serving using tensor and pipeline parallelism. It manages the distributed runtime with &lt;a href="https://www.open-mpi.org/"&gt;MPI&lt;/a&gt;.&lt;/p&gt;
&lt;h2 id="build-the-triton-tensorrt-llm-image"&gt;Build the Triton TensorRT-LLM image&lt;/h2&gt;
&lt;p&gt;We provide a &lt;a href="https://github.com/kubernetes-sigs/lws/blob/main/docs/examples/tensorrt-llm/build/Dockerfile"&gt;Dockerfile&lt;/a&gt; to build the image. The Dockerfile contains an installation script to download any Llama model from hugging face and prepare it to be used by TensorRT-LLM. It also has a python script to initialize MPI and start the server.&lt;/p&gt;</description></item><item><title>llama.cpp</title><link>/docs/examples/llamacpp/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/llamacpp/</guid><description>&lt;h2 id="deploy-distributed-inference-service-with-llamacpp"&gt;Deploy Distributed Inference Service with llama.cpp&lt;/h2&gt;
&lt;p&gt;In this example, we will use LeaderWorkerSet to deploy a distributed
inference service using &lt;a href="https://github.com/ggerganov/llama.cpp"&gt;llama.cpp&lt;/a&gt;.
llama.cpp began as a project to support CPU-only inference on a single node, but has
since expanded to support accelerators and distributed inference.&lt;/p&gt;
&lt;h3 id="deploy-leaderworkerset-of-llamacpp"&gt;Deploy LeaderWorkerSet of llama.cpp&lt;/h3&gt;
&lt;p&gt;We use LeaderWorkerSet to deploy a llama.cpp leader and two llama.cpp workers.
The leader pod loads the model and distributes layers to the workers; the workers
perform the majority of the computation.&lt;/p&gt;</description></item><item><title>SGLang</title><link>/docs/examples/sglang/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/sglang/</guid><description>&lt;h2 id="deploy-distributed-inference-service-with-sglang-and-lws-on-gpus"&gt;Deploy Distributed Inference Service with SGLang and LWS on GPUs&lt;/h2&gt;
&lt;p&gt;In this example, we demonstrate how to deploy a distributed inference service using LeaderWorkerSet (LWS) with &lt;a href="https://docs.sglang.ai/"&gt;SGLang&lt;/a&gt; on GPU clusters.&lt;/p&gt;
&lt;p&gt;SGLang provides native support for distributed tensor-parallel inference and serving, enabling efficient deployment of large language models (LLMs) such as DeepSeek-R1 671B and Llama-3.1-405B across multiple nodes. This example uses the &lt;a href="https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct"&gt;meta-llama/Meta-Llama-3.1-8B-Instruct&lt;/a&gt; model to demonstrate multi-node serving capabilities. For implementation details on distributed execution, see the SGLang docs &lt;a href="https://docs.sglang.ai/references/multi_node_deployment/multi_node.html"&gt;Run Multi-Node Inference&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>Horizontal Pod Autoscaler (HPA)</title><link>/docs/examples/hpa/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/hpa/</guid><description>&lt;p&gt;LeaderWorkerSet supports Horizontal Pod Autoscaler (HPA) through its scale subresource. This allows you to automatically scale the number of replica groups based on resource utilization metrics like CPU or memory.&lt;/p&gt;
&lt;h2 id="how-hpa-works-with-leaderworkerset"&gt;How HPA Works with LeaderWorkerSet&lt;/h2&gt;
&lt;p&gt;When using HPA with LeaderWorkerSet:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;HPA monitors &lt;strong&gt;leader pods&lt;/strong&gt; only (not worker pods)&lt;/li&gt;
&lt;li&gt;The &lt;code&gt;hpaPodSelector&lt;/code&gt; in LeaderWorkerSet status helps HPA identify which pods to monitor&lt;/li&gt;
&lt;li&gt;Scaling affects the number of &lt;strong&gt;replica groups&lt;/strong&gt;, not individual pods within a group&lt;/li&gt;
&lt;li&gt;Each replica group (leader + workers) is scaled as a unit&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id="prerequisites"&gt;Prerequisites&lt;/h2&gt;
&lt;p&gt;Before setting up HPA, ensure you have:&lt;/p&gt;</description></item><item><title>Topology Aware Scheduling with Kueue</title><link>/docs/examples/tas/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/examples/tas/</guid><description>&lt;p&gt;AI Inference workloads require constant Pod-to-Pod communication. This makes the network bandwidth an important requirement of
running workloads efficiently. The bandwidth between the Pods depends on the placement of the Nodes in the data center. Topology Aware Scheduling (TAS), looks to place the pods as closely as possible to maximize the network bandwidth. To learn more about TAS, visit the page in the &lt;a href="https://kueue.sigs.k8s.io/docs/concepts/topology_aware_scheduling/"&gt;Kueue website&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;This example will cover how to deploy a vLLM multi-host workload using TAS.&lt;/p&gt;</description></item><item><title>Labels, Annotations and Environment Variables</title><link>/docs/reference/labels-annotations-and-environment-variables/</link><pubDate>Fri, 14 Mar 2025 00:00:00 +0000</pubDate><guid>/docs/reference/labels-annotations-and-environment-variables/</guid><description>&lt;h1 id="labels"&gt;Labels&lt;/h1&gt;
&lt;table&gt;
 &lt;thead&gt;
 &lt;tr&gt;
 &lt;th&gt;Key&lt;/th&gt;
 &lt;th&gt;Description&lt;/th&gt;
 &lt;th&gt;Example&lt;/th&gt;
 &lt;th&gt;Applies to&lt;/th&gt;
 &lt;/tr&gt;
 &lt;/thead&gt;
 &lt;tbody&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/name&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The name of the LeaderWorkerSet object to which these resources belong.&lt;/td&gt;
 &lt;td&gt;leaderworkerset-multi-template&lt;/td&gt;
 &lt;td&gt;Pod, StatefulSet, Service&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/template-revision-hash&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Hash used to track the controller revision that matches a LeaderWorkerSet object.&lt;/td&gt;
 &lt;td&gt;5c5fcdfb44&lt;/td&gt;
 &lt;td&gt;Pod, StatefulSet&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/group-index&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The group to which it belongs.&lt;/td&gt;
 &lt;td&gt;0&lt;/td&gt;
 &lt;td&gt;Pod, StatefulSet (only worker)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/group-key&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Unique key identifying the group.&lt;/td&gt;
 &lt;td&gt;689ce1b5&amp;hellip;b07&lt;/td&gt;
 &lt;td&gt;Pod, StatefulSet (only worker)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/worker-index&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The index or identity of the pod within the group.&lt;/td&gt;
 &lt;td&gt;0&lt;/td&gt;
 &lt;td&gt;Pod&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/subgroup-index&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Tracks which subgroup the pod is part of.&lt;/td&gt;
 &lt;td&gt;0&lt;/td&gt;
 &lt;td&gt;Pod (only if SubGroup is set)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/subgroup-key&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Pods that are part of the same subgroup will have the same unique hash value.&lt;/td&gt;
 &lt;td&gt;92904e74&amp;hellip;801&lt;/td&gt;
 &lt;td&gt;Pod (only if SubGroup is set)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;/tbody&gt;
&lt;/table&gt;
&lt;h1 id="annotations"&gt;Annotations&lt;/h1&gt;
&lt;table&gt;
 &lt;thead&gt;
 &lt;tr&gt;
 &lt;th&gt;Key&lt;/th&gt;
 &lt;th&gt;Description&lt;/th&gt;
 &lt;th&gt;Example&lt;/th&gt;
 &lt;th&gt;Applies to&lt;/th&gt;
 &lt;/tr&gt;
 &lt;/thead&gt;
 &lt;tbody&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/size&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The total number of pods in each group.&lt;/td&gt;
 &lt;td&gt;4&lt;/td&gt;
 &lt;td&gt;Pod&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/replicas&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Replicas Number of leader-workers groups.&lt;/td&gt;
 &lt;td&gt;3&lt;/td&gt;
 &lt;td&gt;StatefulSet (only leader)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/leader-name&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The name of the leader pod.&lt;/td&gt;
 &lt;td&gt;leaderworkerset-multi-template-0&lt;/td&gt;
 &lt;td&gt;Pod (only worker)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/exclusive-topology&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Specifies the topology for exclusive 1:1 scheduling.&lt;/td&gt;
 &lt;td&gt;cloud.google.com/gke-nodepool&lt;/td&gt;
 &lt;td&gt;LeaderWorkerSet, Pod (only if exclusive-topology is used)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/subdomainPolicy&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Determines what type of domain will be injected.&lt;/td&gt;
 &lt;td&gt;UniquePerReplica&lt;/td&gt;
 &lt;td&gt;Pod (only if leader and subdomainPolicy set to UniquePerReplica)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/subgroup-size&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The number of pods per subgroup.&lt;/td&gt;
 &lt;td&gt;2&lt;/td&gt;
 &lt;td&gt;Pod (only if SubGroup is set)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/subgroup-exclusive-topology&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Specifies the topology for exclusive 1:1 scheduling within a subgroup.&lt;/td&gt;
 &lt;td&gt;topologyKey&lt;/td&gt;
 &lt;td&gt;LeaderWorkerSet, Pod (only if SubGroup is set and subgroup-exclusive-topology is used)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;leaderworkerset.sigs.k8s.io/leader-requests-tpus&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Indicates if the leader pod requests TPU.&lt;/td&gt;
 &lt;td&gt;true&lt;/td&gt;
 &lt;td&gt;Pod (only if leader pod requests TPU)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;/tbody&gt;
&lt;/table&gt;
&lt;h1 id="environment-variables"&gt;Environment Variables&lt;/h1&gt;
&lt;table&gt;
 &lt;thead&gt;
 &lt;tr&gt;
 &lt;th&gt;Key&lt;/th&gt;
 &lt;th&gt;Description&lt;/th&gt;
 &lt;th&gt;Example&lt;/th&gt;
 &lt;th&gt;Applies to&lt;/th&gt;
 &lt;/tr&gt;
 &lt;/thead&gt;
 &lt;tbody&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;LWS_LEADER_ADDRESS&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The address of the leader via the headless service.&lt;/td&gt;
 &lt;td&gt;leaderworkerset-multi-template-0.leaderworkerset-multi-template.default&lt;/td&gt;
 &lt;td&gt;Pod&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;LWS_GROUP_SIZE&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Tracks the size of the LWS group.&lt;/td&gt;
 &lt;td&gt;4&lt;/td&gt;
 &lt;td&gt;Pod&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;LWS_WORKER_INDEX&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;The index or identity of the pod within the group.&lt;/td&gt;
 &lt;td&gt;2&lt;/td&gt;
 &lt;td&gt;Pod&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;TPU_WORKER_HOSTNAMES&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Hostnames of TPU workers only in the same subgroup.&lt;/td&gt;
 &lt;td&gt;test-sample-1-5.default,test-sample-1-6.default,test-sample-1-7.default,test-sample-1-8.default&lt;/td&gt;
 &lt;td&gt;Pod (only if TPU enabled)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;TPU_WORKER_ID&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;ID of the TPU worker.&lt;/td&gt;
 &lt;td&gt;0&lt;/td&gt;
 &lt;td&gt;Pod (only if TPU enabled)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;tr&gt;
 &lt;td&gt;&lt;code&gt;TPU_NAME&lt;/code&gt;&lt;/td&gt;
 &lt;td&gt;Name of the TPU.&lt;/td&gt;
 &lt;td&gt;test-sample-1&lt;/td&gt;
 &lt;td&gt;Pod (only if TPU enabled)&lt;/td&gt;
 &lt;/tr&gt;
 &lt;/tbody&gt;
&lt;/table&gt;
&lt;p&gt;If you want to use more environment variables, they are available in the labels or annotations but not listed in the Environment Variables section.
We can obtain the index by using the &lt;a href="https://kubernetes.io/docs/concepts/workloads/pods/downward-api/"&gt;Downward API&lt;/a&gt; to pass the Pod&amp;rsquo;s label as an environment variable to the container.&lt;/p&gt;</description></item><item><title>LeaderWorkerSet API</title><link>/docs/reference/leaderworkerset.v1/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/docs/reference/leaderworkerset.v1/</guid><description>&lt;h2 id="resource-types"&gt;Resource Types&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href="/docs/reference/leaderworkerset.v1/#leaderworkerset-x-k8s-io-v1-LeaderWorkerSet"&gt;LeaderWorkerSet&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;h2 id="leaderworkerset-x-k8s-io-v1-LeaderWorkerSet"&gt;&lt;code&gt;LeaderWorkerSet&lt;/code&gt; &lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Appears in:&lt;/strong&gt;&lt;/p&gt;
&lt;p&gt;LeaderWorkerSet is the Schema for the leaderworkersets API&lt;/p&gt;
&lt;table class="table"&gt;
&lt;thead&gt;&lt;tr&gt;&lt;th width="30%"&gt;Field&lt;/th&gt;&lt;th&gt;Description&lt;/th&gt;&lt;/tr&gt;&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;&lt;td&gt;&lt;code&gt;apiVersion&lt;/code&gt;&lt;br/&gt;string&lt;/td&gt;&lt;td&gt;&lt;code&gt;leaderworkerset.x-k8s.io/v1&lt;/code&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td&gt;&lt;code&gt;kind&lt;/code&gt;&lt;br/&gt;string&lt;/td&gt;&lt;td&gt;&lt;code&gt;LeaderWorkerSet&lt;/code&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td&gt;&lt;code&gt;spec&lt;/code&gt; &lt;B&gt;[Required]&lt;/B&gt;&lt;br/&gt;
&lt;a href="#leaderworkerset-x-k8s-io-v1-LeaderWorkerSetSpec"&gt;&lt;code&gt;LeaderWorkerSetSpec&lt;/code&gt;&lt;/a&gt;
&lt;/td&gt;
&lt;td&gt;
 &lt;p&gt;spec defines the desired behavior of LeaderWorkerSet.&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;&lt;td&gt;&lt;code&gt;status&lt;/code&gt; &lt;B&gt;[Required]&lt;/B&gt;&lt;br/&gt;
&lt;a href="#leaderworkerset-x-k8s-io-v1-LeaderWorkerSetStatus"&gt;&lt;code&gt;LeaderWorkerSetStatus&lt;/code&gt;&lt;/a&gt;
&lt;/td&gt;
&lt;td&gt;
 &lt;p&gt;status represents the current status of LeaderWorkerSet.&lt;/p&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;
&lt;h2 id="leaderworkerset-x-k8s-io-v1-LeaderWorkerSetSpec"&gt;&lt;code&gt;LeaderWorkerSetSpec&lt;/code&gt; &lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;Appears in:&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href="/docs/reference/leaderworkerset.v1/#leaderworkerset-x-k8s-io-v1-LeaderWorkerSet"&gt;LeaderWorkerSet&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;One group consists of a single leader and M workers, and the total number of pods in a group is M+1.
LeaderWorkerSet will create N replicas of leader-worker pod groups (hereinafter referred to as group).&lt;/p&gt;</description></item><item><title>Search Results</title><link>/search/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/search/</guid><description/></item></channel></rss>