CloneSet
This controller provides advanced features to efficiently manage stateless applications in large-scale scenarios that
do not have instance order requirement during scaling and rollout.
Analogously, CloneSet can be recognized as an enhanced version of upstream Deployment
workload, but it does many more.
As name suggests, CloneSet is a Set -suffix controller which manages Pods directly. A sample CloneSet yaml looks like following:
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
metadata:
labels:
app: sample
name: sample
spec:
replicas: 5
selector:
matchLabels:
app: sample
template:
metadata:
labels:
app: sample
spec:
containers:
- name: nginx
image: nginx:alpine
Scale featuresโ
Support PVCsโ
CloneSet allows user to define PVC templates volumeClaimTemplates
in CloneSetSpec
, which can support PVCs per Pod.
This cannot be done with Deployment
. If not specified, CloneSet will only create Pods without PVCs.
A few reminders:
- Each PVC created by CloneSet has an owner reference. So when a CloneSet has been deleted, its PVCs will be cascading deleted.
- Each Pod and PVC created by CloneSet has a "apps.kruise.io/cloneset-instance-id" label key. The associated Pod and PVC will have the same instance-id. They use the same string as label value which is composed of a unique instance-id as suffix of the CloneSet name.
- When a Pod has been deleted by CloneSet controller, all PVCs related to it will be deleted together.
- When a Pod has been deleted manually, all PVCs related to the Pod are preserved, and CloneSet controller will create a new Pod with the same instance-id and reuse the PVCs.
- When a Pod is updated using recreate policy, all PVCs related to it will be deleted together.
- When a Pod is updated using in-place policy, all PVCs related to it are preserved.
The following shows a sample CloneSet yaml file which contains PVC templates.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
metadata:
labels:
app: sample
name: sample-data
spec:
replicas: 5
selector:
matchLabels:
app: sample
template:
metadata:
labels:
app: sample
spec:
containers:
- name: nginx
image: nginx:alpine
volumeMounts:
- name: data-vol
mountPath: /usr/share/nginx/html
volumeClaimTemplates:
- metadata:
name: data-vol
spec:
accessModes: [ "ReadWriteOnce" ]
resources:
requests:
storage: 20Gi
FEATURE STATE: Kruise v1.4.0
When a Pod has been deleted manually, all PVCs related to the Pod are preserved, and CloneSet controller will create a new Pod with the same instance-id and reuse the PVCs.
However, if the Node where the Pod is located experiences an exception, reusing the PVCs may result in the failure of the new Pod to start. For more details, please refer to issue 1099. To address this issue, you can set the DisablePVCReuse=true field, and the PVCs associated with the Pod will be automatically deleted and no longer reused.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
...
replicas: 4
scaleStrategy:
disablePVCReuse: true
Selective Pod deletionโ
When a CloneSet is scaled down, sometimes user has preference to deleting specific Pods.
This cannot be done using StatefulSet
or Deployment
, because StatefulSet
always delete Pod
in order and Deployment
/ReplicaSet
only delete Pod by its own sorted sequence.
CloneSet allows user to specify to-be-deleted Pod names when scaling down replicas
. Take the following
sample as an example,
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
replicas: 4
scaleStrategy:
podsToDelete:
- sample-9m4hp
when controller receives above update request, it ensures the number of replicas to be 4. If some Pods needs to be
deleted, the Pods listed in podsToDelete
will be deleted first.
Controller will clear podsToDelete
automatically once the listed Pods are deleted. Note that:
If one just adds a Pod name to podsToDelete
and do not modify replicas
, controller will delete this Pod, and create a new Pod.
If one is unable to change CloneSet directly, an alternative way is to add a label apps.kruise.io/specified-delete: true
onto the Pod waiting to delete.
Comparing to delete the Pod directly, using podsToDelete
or apps.kruise.io/specified-delete: true
will have CloneSet protection by maxUnavailable
/maxSurge
and lifecycle PreparingDelete
triggering (See below).
Deletion Sequenceโ
- Node unassigned < assigned
- PodPending < PodUnknown < PodRunning
- Not ready < ready
- Lower pod-deletion cost < higher pod-deletion-cost
- Higher spread rank < lower spread rank
- Been ready for empty time < less time < more time
- Pods with containers with higher restart counts < lower restart counts
- Empty creation time pods < newer pods < older pods
Pod deletion costโ
FEATURE STATE: Kruise v0.9.0
The controller.kubernetes.io/pod-deletion-cost annotation
is defined in Kubernetes since v1.21
, Deployment/ReplicaSet will remove pods according to this cost when downscaling.
And CloneSet has also supported it since Kruise v0.9.0
.
The annotation should be set on the pod, the range is [-2147483647, 2147483647]. It represents the cost of deleting a pod compared to other pods belonging to the same CloneSet. Pods with lower deletion cost are preferred to be deleted before pods with higher deletion cost.
The implicit value for this annotation for pods that don't set it is 0; negative values are permitted.
Deletion by Spread Constraintsโ
FEATURE STATE: Kruise v0.10.0
The original proposal(design doc) is here.
Currently, it supports deletion by same node spread and deletion by pod topolocy spread constraints.
If there are Pod Topology Spread Constraints defined in CloneSet template, controller will choose pods according to spread constraints when the cloneset needs to scale down. Otherwise, controller will choose pods by same node spread by default when scaling down.
Short hash labelโ
FEATURE STATE: Kruise v0.9.0
By default, CloneSet set the controller-revision-hash
in Pod label to the full name of ControllerRevision, such as:
apiVersion: v1
kind: Pod
metadata:
labels:
controller-revision-hash: demo-cloneset-956df7994
It is joined by the name of CloneSet and the hash of the ControllerRevision. Length of the hash is usually 8~10 characters, and the label value in Kubernetes can not be more than 63 characters. So the name of CloneSet should be less than 52 characters.
A new feature-gate named CloneSetShortHash
has been introduced.
If it is enabled, CloneSet will only set the controller-revision-hash
to the real hash, such as 956df7994
.
So there will be no limit to CloneSet name.
Don't worry. Even if you enable the CloneSetShortHash
, CloneSet will still recognize and manage the old Pods with full revision label.
Since Kruise v1.1.0, CloneSet will add another pod-template-hash
label into Pods, which will always be the short hash.
Scale up with rate limitโ
FEATURE STATE: Kruise v1.0.0
Users can specify ScaleStrategy.MaxUnavailable
to limit the step size of CloneSet Scaling Up, so as to minimize the impact on application services.
This value can be an absolute number (e.g., 5) or a percentage of desired number of Pods (e.g., 10%). Default value is nil
(i.e., empty pointer), which indicates non-limitation.
ScaleStrategy.MaxUnavailable
field can cooperate with 'Spec.MinReadySeconds' field to work, for example:
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
minReadySeconds: 60
scaleStrategy:
maxUnavailable: 1
The effect of the above configuration is that during scaling up, CloneSet will not create the next pod until the previous pod has been ready for more than one minute.
Update featuresโ
Update typesโ
CloneSet provides three update types, defaults to ReCreate
.
ReCreate
: controller will delete old Pods and PVCs and create new ones.InPlaceIfPossible
: controller will try to in-place update Pod instead of recreating them if possible. Please read the concept doc below.InPlaceOnly
: controller will in-place update Pod instead of recreating them. WithInPlaceOnly
policy, user cannot modify any fields other than the fields that supported to in-place update.
You may need to read the concept doc for more details of in-place update.
We also bring graceful period into in-place update. CloneSet has supported gracePeriodSeconds
, which is a period
duration between controller update pod status and update pod images.
So that endpoints-controller could have enough time to remove this Pod from endpoints.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
type: InPlaceIfPossible
inPlaceUpdateStrategy:
gracePeriodSeconds: 10
Template and revisionโ
spec.template
defines the latest pod template in the CloneSet.
Controller will calculate a revision hash for each version of spec.template
when it has been initialized or modified.
For example, when we create a sample CloneSet, controller will calculate the revision hash sample-744d4796cc
and
present the hash in CloneSet Status.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
metadata:
generation: 1
# ...
spec:
replicas: 5
# ...
status:
observedGeneration: 1
readyReplicas: 5
replicas: 5
currentRevision: sample-d4d4fb5bd
updateRevision: sample-d4d4fb5bd
updatedReadyReplicas: 5
updatedReplicas: 5
# ...
Here are the explanations for the counters presented in CloneSet status:
status.replicas
: Number of podsstatus.readyReplicas
: Number of ready podsstatus.availableReplicas
: Number of ready and available pods (satisfied withminReadySeconds
and pod lifecycle state isNormal
)status.currentRevision
: Latest revision hash that has used to be updated to all Podsstatus.updateRevision
: Latest revision hash of this CloneSetstatus.updatedReplicas
: Number of pods with the latest revisionstatus.updatedReadyReplicas
: Number of ready pods with the latest revision
FEATURE STATE: Kruise v1.2.0
status.expectedUpdatedReplicas
: Number of the pods that were updated or will be updated with the latest revision under the currentpartition
settings.
Partitionโ
Partition is the desired number or percent of Pods in old revisions, defaults to 0
. This field does NOT imply any update order.
When partition
is set during update:
- If it is a number:
(replicas - partition)
number of pods will be updated with the new version. - If it is a percent:
(replicas * (100% - partition))
number of pods will be updated with the new version.
FEATURE STATE: Kruise v1.2.0
- If
partition
is a percent, andpartition < 100% && replicas > 1
, CloneSet will ensure at least one pod will be updated with the new version. - One can use the condition
.status.updatedReplicas >= .status.expectedUpdatedReplicas
to decide whether workload had finish rolling out new revision under partition restriction.
For example, when we update sample CloneSet's container image to nginx:mainline
and set partition=3
, after a while, the sample CloneSet yaml looks like the following:
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
metadata:
# ...
generation: 2
spec:
replicas: 5
template:
metadata:
labels:
app: sample
spec:
containers:
- image: nginx:mainline
imagePullPolicy: Always
name: nginx
updateStrategy:
partition: 3
# ...
status:
observedGeneration: 2
readyReplicas: 5
replicas: 5
currentRevision: sample-d4d4fb5bd
updateRevision: sample-56dfb978d4
updatedReadyReplicas: 2
updatedReplicas: 2
Note that status.updateRevision
has been updated to sample-56dfb978d4
, a new hash.
Since we set partition=3
, controller only updates two Pods to the latest revision.
$ kubectl get pod -L controller-revision-hash
NAME READY STATUS RESTARTS AGE CONTROLLER-REVISION-HASH
sample-chvnr 1/1 Running 0 6m46s sample-d4d4fb5bd
sample-j6c4s 1/1 Running 0 6m46s sample-d4d4fb5bd
sample-ns85c 1/1 Running 0 6m46s sample-d4d4fb5bd
sample-jnjdp 1/1 Running 0 10s sample-56dfb978d4
sample-qqglp 1/1 Running 0 18s sample-56dfb978d4
Rollback by partitionโ
FEATURE STATE: Kruise v0.9.0
By default, partition
can only control Pods updating to the status.updateRevision
.
Which means for this CloneSet, when changes partition 5 -> 3
, CloneSet will update 2 Pods to status.updateRevision
.
Then changes partition 3 -> 5
back, CloneSet will do nothing.
But if you have enabled CloneSetPartitionRollback
feature-gate, in this case,
CloneSet will update the 2 Pods in status.updateRevision
back to status.currentRevision
.
MaxUnavailableโ
MaxUnavailable is the maximum number of Pods that can be unavailable. Value can be an absolute number (e.g., 5) or a percentage of desired number of Pods (e.g., 10%). Default value is 20%.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
maxUnavailable: 20%
Since Kruise v0.9.0
, maxUnavailable
not only controls Pods update, but also affect Pods specified deletion.
Which means if you declare to delete a Pod via podsToDelete
or apps.kruise.io/specified-delete: true
,
CloneSet will delete it only if the number of unavailable Pods (comparing to the replicas number) is less than maxUnavailable
.
MaxSurgeโ
MaxSurge is the maximum number of pods that can be scheduled above the desired replicas. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Defaults to 0.
If maxSurge is set somewhere, cloneset-controller will create maxSurge
number of Pods above the replicas
,
when it finds multiple active revisions of Pods which means the CloneSet is in the update stage.
After all Pods except partition
number have been updated to the latest revision, maxSurge
number Pods will be deleted,
and the number of Pods will be equal to the replica
number.
What's more, maxSurge is forbidden to use with InPlaceOnly
policy.
When maxSurge is used with InPlaceIfPossible
, controller will create additional Pods with latest revision first,
and then update the rest Pods with old revisions,
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
maxSurge: 3
Since Kruise v0.9.0
, maxSurge
not only controls Pods update, but also affect Pods specified deletion.
Which means if you declare to delete a Pod via podsToDelete
or apps.kruise.io/specified-delete: true
,
CloneSet may create new a Pod, wait it to be ready, and them delete the old one.
It depends on maxUnavailable
and the current number of unavailable Pods.
For example:
- For a CloneSet
maxUnavailable=2, maxSurge=1
and currently only one unavailable Pods ispod-a
, if you patchapps.kruise.io/specified-delete: true
ontopod-b
or put the Pod name intopodsToDelete
, CloneSet will delete it directly. - For a CloneSet
maxUnavailable=1, maxSurge=1
and currently only one unavailable Pods ispod-a
, if you patchapps.kruise.io/specified-delete: true
ontopod-b
or put the Pod name intopodsToDelete
, CloneSet will create a new Pod, waiting it to be ready, and finally deletepod-b
. - For a CloneSet
maxUnavailable=1, maxSurge=1
and currently only one unavailable Pods ispod-a
, if you patchapps.kruise.io/specified-delete: true
ontopod-a
or put the Pod name intopodsToDelete
, CloneSet will delete it directly. - ...
Update sequenceโ
When controller chooses Pods to update, it has default sort logic based on Pod phase and conditions:
unscheduled < scheduled, pending < unknown < running, not-ready < ready.
In addition, CloneSet also supports advanced priority
and scatter
strategies to allow users to specify the update order.
priorityโ
This strategy defines rules for calculating the priority of updating pods.
All update candidates will be applied with the priority terms.
priority
can be calculated either by weight or by order.
weight
: Priority is determined by the sum of weights for terms that match selector. For example,
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
priorityStrategy:
weightPriority:
- weight: 50
matchSelector:
matchLabels:
test-key: foo
- weight: 30
matchSelector:
matchLabels:
test-key: bar
order
: Priority will be determined by the value of the orderKey. The update candidates are sorted based on the "int" part of the value string. For example, 5 in string "5" and 10 in string "sts-10".
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
priorityStrategy:
orderPriority:
- orderedKey: some-label-key
scatterโ
This strategy defines rules to make certain Pods be scattered during update.
For example, if a CloneSet has replica=10
, and we add foo=bar
label in 3 Pods and specify the following scatter rule. These 3 Pods will
be the 1st, 6th and 10th updated Pods.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
scatterStrategy:
- key: foo
value: bar
Note that:
- Although
priority
strategy andscatter
strategy can be applied together, we strongly suggest to just use one of them to avoid confusion. - If
scatter
strategy is used, we suggest to just use one term. Otherwise, the update order can be hard to understand.
Last but not the least, the above advanced update strategies require independent Pod labeling mechanisms, which are not provided by CloneSet.
Paused updateโ
paused
indicates that Pods updating is paused, controller will not update Pods but just maintain the number of replicas.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# ...
updateStrategy:
paused: true
Pre-download image for in-place updateโ
FEATURE STATE: Kruise v0.9.0
If you have enabled the PreDownloadImageForInPlaceUpdate
feature-gate during Kruise installation or upgrade,
CloneSet controller will automatically pre-download the image you want to update to the nodes of all old Pods.
It is quite useful to accelerate the progress of applications upgrade.
The parallelism of each new image pre-downloading by CloneSet is 1
, which means the image is downloaded on nodes one by one.
You can change the parallelism using apps.kruise.io/image-predownload-parallelism
annotation on CloneSet according to the capability of image registry,
for registries with more bandwidth and P2P image downloading ability, a larger parallelism can speed up the pre-download process.
Since Kruise v1.1.0, you can use apps.kruise.io/image-predownload-min-updated-ready-pods
to make sure the new image starting pre-download after a few new Pods have been updated ready. Its value can be absolute number or percentage.
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
metadata:
annotations:
apps.kruise.io/image-predownload-parallelism: "10"
apps.kruise.io/image-predownload-min-updated-ready-pods: "3"
Note that to avoid most unnecessary image downloading, now controller will only pre-download images for CloneSet with replicas > 3
.
Lifecycle hookโ
Each Pod managed by CloneSet has a clear state defined in lifecycle.apps.kruise.io/state
label:
- Normal
- PreparingUpdate
- Updating
- Updated
- PreparingDelete
Lifecycle hook allows users to do something (for example remove pod from service endpoints) during Pod deleting and before/after in-place update.
type LifecycleStateType string
// Lifecycle contains the hooks for Pod lifecycle.
type Lifecycle struct
// PreDelete is the hook before Pod to be deleted.
PreDelete *LifecycleHook `json:"preDelete,omitempty"`
// InPlaceUpdate is the hook before Pod to update and after Pod has been updated.
InPlaceUpdate *LifecycleHook `json:"inPlaceUpdate,omitempty"`
// PreNormal is the hook after Pod to be created and ready to be Normal.
PreNormal *LifecycleHook `json:"preNormal,omitempty"`
}
type LifecycleHook struct {
LabelsHandler map[string]string `json:"labelsHandler,omitempty"`
FinalizersHandler []string `json:"finalizersHandler,omitempty"`
/********************** FEATURE STATE: 1.2.0 ************************/
// MarkPodNotReady = true means:
// - Pod will be set to 'NotReady' at preparingDelete/preparingUpdate state.
// - Pod will be restored to 'Ready' at Updated state if it was set to 'NotReady' at preparingUpdate state.
// Currently, MarkPodNotReady only takes effect on InPlaceUpdate & PreDelete hook.
// Default to false.
MarkPodNotReady bool `json:"markPodNotReady,omitempty"`
/*********************************************************************/
}
Examples:
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
# define with finalizer
lifecycle:
preNormal:
finalizersHandler:
- example.io/unready-blocker
preDelete:
finalizersHandler:
- example.io/unready-blocker
inPlaceUpdate:
finalizersHandler:
- example.io/unready-blocker
# or define with label
# lifecycle:
# inPlaceUpdate:
# labelsHandler:
# example.io/block-unready: "true"
MarkPodNotReadyโ
FEATURE STATE: Kruise v1.2.0
lifecycle:
preDelete:
markPodNotReady: true
finalizersHandler:
- example.io/unready-blocker
inPlaceUpdate:
markPodNotReady: true
finalizersHandler:
- example.io/unready-blocker
- If you set
markPodNotReady=true
forpreDelete
:- Kruise will try to set
KruisePodReady
condition toFalse
when Pods enterPreparingDelete
lifecycle state, and Pods will be set to NotReady, but containers stillRunning
.
- Kruise will try to set
- If you set
markPodNotReady=true
forinPlaceUpdate
:- Kruise will try to set
KruisePodReady
condition toFalse
when Pods enterPreparingUpdate
lifecycle state, and Pods will be set to NotReady, but containers stillRunning
. - Kruise will try to set
KruisePodReady
condition toTrue
when Pods enterUpdated
lifecycle state.
- Kruise will try to set
One can use this markPodNotReady
feature to drain service traffic before terminating containers.
Note: this feature only works when pod has KruisePodReady
ReadinessGate.
State circulationโ
- When CloneSet create a Pod (including scaling up and recreate update):
- The Pod will be regarded as
Available
only after PreNormal hook is satisfied. PreNormal
hook can be used for post-checks after pod creation. For example, one can check if the pod have been added as the SLB backends successfully. Without preNormal hook, one may encounter traffic loss during rolling upgrades if Operator(e.g., CCM) fails to add new pods to the SLB backends.
- The Pod will be regarded as
- When CloneSet delete a Pod (including scale in and recreate update):
- Delete it directly if no lifecycle hook definition or Pod not matched preDelete hook
- Otherwise, CloneSet will firstly update Pod to
PreparingDelete
state and wait for user controller to remove the label/finalizer and Pod not matched preDelete hook - Note that Pods in
PreparingDelete
state will not be updated
- When CloneSet update a Pod in-place:
- If lifecycle hook defined and Pod matched inPlaceUpdate hook, CloneSet will update Pod to
PreparingUpdate
state - After user controller remove the label/finalizer (thus Pod not matched inPlaceUpdate hook), CloneSet will update it to
Updating
state and start updating - After in-place update completed, CloneSet will update Pod to
Updated
state if lifecycle hook defined and Pod not matched inPlaceUpdate hook - When user controller add label/finalizer into Pod and it matched inPlaceUpdate hook, CloneSet will finally update it to
Normal
state
- If lifecycle hook defined and Pod matched inPlaceUpdate hook, CloneSet will update Pod to
Besides, although our design supports to change a Pod from PreparingDelete
back to Normal
(through cancel specified delete), but it is not recommended. Because Pods in PreparingDelete
state will not be updated by CloneSet, it might be updating immediately if comes back to Normal
. This case is hard for user controller to handle.
Example for user controller logicโ
Same as yaml example above, we can firstly define๏ผ
example.io/unready-blocker
finalizer as hook
Add these fields into CloneSet template:
apiVersion: apps.kruise.io/v1alpha1
kind: CloneSet
spec:
lifecycle:
preNormal:
finalizersHandler:
- example.io/unready-blocker
preDelete:
finalizersHandler:
- example.io/unready-blocker
inPlaceUpdate:
finalizersHandler:
- example.io/unready-blocker
template:
# ...
User controller logic:
- For Pod in
PreparingNormal
state, if there is noexample.io/unready-blocker
, patch the finalizer to Pod, and Pod will be available for CloneSet, and will enterNormal
state. - For Pod in
PreparingDelete
andPreparingUpdate
states, delete it from endpoints and removeexample.io/unready-blocker
finalizer - For Pod in
Updated
state, add it to endpoints and addexample.io/unready-blocker
finalizer
Scaling with PreparingDeleteโ
FEATURE STATE: Kruise v1.3.0
CloneSet considers Pods in PreparingDelete
state as normal by default, which means these Pods will still be calculated in the replicas
number.
In this situation:
- if you scale down
replicas
fromN
toN-1
, when the Pod to be deleted is still inPreparingDelete
, you scale upreplicas
toN
, the CloneSet will move the Pod back toNormal
. - if you scale down
replicas
fromN
toN-1
and put a Pod intopodsToDelete
, when the specific Pod is still inPreparingDelete
, you scale upreplicas
toN
, the CloneSet will not create a new Pod until the specific Pod goes into terminating. - if you specifically delete a Pod without
replicas
changed, when the specific Pod is still inPreparingDelete
, the CloneSet will not create a new Pod until the specific Pod goes into terminating.
Since Kruise v1.3.0, you can put a apps.kruise.io/cloneset-scaling-exclude-preparing-delete: "true"
label into CloneSet, which indicates Pods in PreparingDelete
will not be calculated in the replicas
number.
In this situation:
- if you scale down
replicas
fromN
toN-1
, when the Pod to be deleted is still inPreparingDelete
, you scale upreplicas
toN
, the CloneSet will move the Pod back toNormal
. - if you scale down
replicas
fromN
toN-1
and put a Pod intopodsToDelete
, even if the specific Pod is still inPreparingDelete
, you scale upreplicas
toN
, the CloneSet will create a new Pod immediately. - if you specifically delete a Pod without
replicas
changed, even if the specific Pod is still inPreparingDelete
, the CloneSet will create a new Pod immediately.
Performance optimizationโ
FEATURE STATE: Kruise v1.4.0
Currently, both status and metadata changes of Pods will trigger the reconcile of CloneSet Controller. CloneSet Reconcile is configured by default with three workers, it works normally in small-scale clusters.
However, for larger clusters or scenarios with frequent Pod update events, these unnecessary reconciles will block the real CloneSet reconciles, resulting in delayed rolling updates and other changes. To solve this problem, you can enable the feature-gate CloneSetEventHandlerOptimization to reduce some unnecessary reconcile enqueues.