그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그 그
7 minute read
This document provides an overview of the ShootStatus.
The Shoot status consists of a set of conditions. A Condition has the following fields:
|Name of the condition.|
|Indicates whether the condition is applicable, with possible values |
|Timestamp for when the condition last transitioned from one status to another.|
|Timestamp for when the condition was updated. Usually changes when |
|Machine-readable, UpperCamelCase text indicating the reason for the condition’s last transition.|
|Human-readable message indicating details about the last status transition.|
|Well-defined error codes in case the condition reports a problem.|
Currently, the available Shoot condition types are:
The Shoot conditions are maintained by the shoot care reconciler of the gardenlet. Find more information in the gardelent documentation.
The condition checks are executed periodically at an interval which is configurable in the
.controllers.shootCare.syncPeriod, defaults to
GardenletConfiguration also allows configuring condition thresholds (
controllers.shootCare.conditionThresholds). A condition threshold is the amount of time to consider a condition as
Processing on condition status changes.
Let’s check the following example to get a better understanding. Let’s say that the
APIServerAvailable condition of our Shoot is with status
True. If the next condition check fails (for example kube-apiserver becomes unreachable), then the condition first goes to
Processing state. Only if this state remains for condition threshold amount of time, then the condition is finally updated to
Constraints represent conditions of a Shoot’s current state that constraint some operations on it. The current constraints are:
This constraint indicates whether a Shoot is allowed to be hibernated.
The rationale behind this constraint is that a Shoot can have
MutatingWebhookConfigurations acting on resources that are critical for waking up a cluster.
For example, if a webhook has rules for
CREATE/UPDATE Pods or Nodes and
failurePolicy=Fail, the webhook will block joining
Nodes and creating critical system component Pods and thus block the entire wakeup operation, because the server backing the webhook is not running.
Even if the
failurePolicy is set to
Ignore, high timeouts (
>15s) can lead to blocking requests of control plane components.
That’s because most control-plane API calls are made with a client-side timeout of
30s, so if a webhook has
the overall request might still fail as there is overhead in communication with the API server and potential other webhooks.
Generally, it’s best practice to specify low timeouts in WebhookConfigs.
As an effort to correct this common problem, the webhook remediator has been created. This is enabled by setting
.controllers.shootCare.webhookRemediatorEnabled=true in the
gardenlet’s configuration. This feature simply checks whether webhook configurations in shoot clusters match a set of rules described here. If at least one of the rules matches, it will change set
status=False for the
.status.constraints of type
MaintenancePreconditionsSatisfied in the
Shoot resource. In addition, the
failurePolicy in the affected webhook configurations will be set from
Ignore. Gardenlet will also add an annotation to make it visible to end-users that their webhook configurations were mutated and should be fixed/adapted according to the rules and best practices.
In most cases, you can avoid this by simply excluding the
kube-system namespace from your webhook via the
apiVersion: admissionregistration.k8s.io/v1 kind: MutatingWebhookConfiguration webhooks: - name: my-webhook.example.com namespaceSelector: matchExpressions: - key: gardener.cloud/purpose operator: NotIn values: - kube-system rules: - operations: ["*"] apiGroups: [""] apiVersions: ["v1"] resources: ["pods"] scope: "Namespaced"
However, some other resources (some of them cluster-scoped) might still trigger the remediator, namely:
If one of the above resources triggers the remediator, the preferred solution is to remove that particular resource from your webhook’s
rules. You can also use the
objectSelector to reduce the scope of webhook’s
rules. However, in special cases where a webhook is absolutely needed for the workload, it is possible to add the
remediation.webhook.shoot.gardener.cloud/exclude=true label to your webhook so that the remediator ignores it. This label should not be used to silence an alert, but rather to confirm that a webhook won’t cause problems. Note that all of this is no perfect solution and just done on a best effort basis, and only the owner of the webhook can know whether it indeed is problematic and configured correctly.
You can also find more help from the Kubernetes documentation
This constraint indicates whether all preconditions for a safe maintenance operation are satisfied (see Shoot Maintenance for more information about what happens during a shoot maintenance).
As of today, the same checks as in the
HibernationPossible constraint are being performed (user-deployed webhooks that might interfere with potential rolling updates of shoot worker nodes).
There is no further action being performed on this constraint’s status (maintenance is still being performed).
It is meant to make the user aware of potential problems that might occur due to his configurations.
This constraints indicates that there is at least one CA certificate which expires in less than
It will not be added to the
.status.constraints if there is no such CA certificate.
However, if it’s visible, then a credentials rotation operation should be considered.
The Shoot status holds information about the last operation that is performed on the Shoot. The last operation field reflects overall progress and the tasks that are currently being executed. Allowed operation types are
Restore. Allowed operation states are
Aborted. An operation in
Error state is an operation that will be retried for a configurable amount of time (
controllers.shoot.retryDuration field in
GardenletConfiguration, defaults to
12h). If the operation cannot complete successfully for the configured retry duration, it will be marked as
Failed. An operation in
Failed state is an operation that won’t be retried automatically (to retry such an operation, see Retry failed operation).
The Shoot status also contains information about the last occurred error(s) (if any) during an operation. A LastError consists of identifier of the task returned error, human-readable message of the error and error codes (if any) associated with the error.
Known error codes are:
ERR_INFRA_UNAUTHENTICATED- Indicates that the last error occurred due to the client request not being completed because it lacks valid authentication credentials for the requested resource. It is classified as a non-retryable error code.
ERR_INFRA_UNAUTHORIZED- Indicates that the last error occurred due to the server understanding the request but refusing to authorize it. It is classified as a non-retryable error code.
ERR_INFRA_QUOTA_EXCEEDED- Indicates that the last error occurred due to infrastructure quota limits. It is classified as a non-retryable error code.
ERR_INFRA_RATE_LIMITS_EXCEEDED- Indicates that the last error occurred due to exceeded infrastructure request rate limits.
ERR_INFRA_DEPENDENCIES- Indicates that the last error occurred due to dependent objects on the infrastructure level. It is classified as a non-retryable error code.
ERR_RETRYABLE_INFRA_DEPENDENCIES- Indicates that the last error occurred due to dependent objects on the infrastructure level, but the operation should be retried.
ERR_INFRA_RESOURCES_DEPLETED- Indicates that the last error occurred due to depleted resource in the infrastructure.
ERR_CLEANUP_CLUSTER_RESOURCES- Indicates that the last error occurred due to resources in the cluster that are stuck in deletion.
ERR_CONFIGURATION_PROBLEM- Indicates that the last error occurred due to a configuration problem. It is classified as a non-retryable error code.
ERR_RETRYABLE_CONFIGURATION_PROBLEM- Indicates that the last error occurred due to a retryable configuration problem. “Retryable” means that the occurred error is likely to be resolved in a ungraceful manner after given period of time.
ERR_PROBLEMATIC_WEBHOOK- Indicates that the last error occurred due to a webhook not following the Kubernetes best practices.
Shoots will be automatically labeled with the
Its value might either be
unknown depending on the
status.lastErrors of the
This can be used as an easy filter method to find shoots based on their “health” status.