CAST AI v7.73.2 published on Wednesday, Oct 29, 2025 by CAST AI
castai.WorkloadScalingPolicy
Start a Neo task
Explain and create a castai.WorkloadScalingPolicy resource
Create WorkloadScalingPolicy Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new WorkloadScalingPolicy(name: string, args: WorkloadScalingPolicyArgs, opts?: CustomResourceOptions);@overload
def WorkloadScalingPolicy(resource_name: str,
args: WorkloadScalingPolicyArgs,
opts: Optional[ResourceOptions] = None)
@overload
def WorkloadScalingPolicy(resource_name: str,
opts: Optional[ResourceOptions] = None,
cpu: Optional[_workload.WorkloadScalingPolicyCpuArgs] = None,
apply_type: Optional[str] = None,
memory: Optional[_workload.WorkloadScalingPolicyMemoryArgs] = None,
cluster_id: Optional[str] = None,
management_option: Optional[str] = None,
downscaling: Optional[_workload.WorkloadScalingPolicyDownscalingArgs] = None,
anti_affinity: Optional[_workload.WorkloadScalingPolicyAntiAffinityArgs] = None,
confidence: Optional[_workload.WorkloadScalingPolicyConfidenceArgs] = None,
assignment_rules: Optional[Sequence[_workload.WorkloadScalingPolicyAssignmentRuleArgs]] = None,
memory_event: Optional[_workload.WorkloadScalingPolicyMemoryEventArgs] = None,
name: Optional[str] = None,
predictive_scaling: Optional[_workload.WorkloadScalingPolicyPredictiveScalingArgs] = None,
rollout_behavior: Optional[_workload.WorkloadScalingPolicyRolloutBehaviorArgs] = None,
startup: Optional[_workload.WorkloadScalingPolicyStartupArgs] = None)func NewWorkloadScalingPolicy(ctx *Context, name string, args WorkloadScalingPolicyArgs, opts ...ResourceOption) (*WorkloadScalingPolicy, error)public WorkloadScalingPolicy(string name, WorkloadScalingPolicyArgs args, CustomResourceOptions? opts = null)
public WorkloadScalingPolicy(String name, WorkloadScalingPolicyArgs args)
public WorkloadScalingPolicy(String name, WorkloadScalingPolicyArgs args, CustomResourceOptions options)
type: castai:workload:WorkloadScalingPolicy
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args WorkloadScalingPolicyArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args WorkloadScalingPolicyArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args WorkloadScalingPolicyArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args WorkloadScalingPolicyArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args WorkloadScalingPolicyArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
WorkloadScalingPolicy Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The WorkloadScalingPolicy resource accepts the following input properties:
- Apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Cluster
Id string - CAST AI cluster id
- Cpu
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Cpu - Management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- Memory
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Memory - Anti
Affinity Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Anti Affinity - Assignment
Rules List<Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Assignment Rule> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- Confidence
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- Downscaling
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Downscaling - Memory
Event Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Memory Event - Name string
- Scaling policy name
- Predictive
Scaling Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Predictive Scaling - Rollout
Behavior Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- Startup
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Startup
- Apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Cluster
Id string - CAST AI cluster id
- Cpu
Workload
Scaling Policy Cpu Args - Management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- Memory
Workload
Scaling Policy Memory Args - Anti
Affinity WorkloadScaling Policy Anti Affinity Args - Assignment
Rules WorkloadScaling Policy Assignment Rule Args - Allows defining conditions for automatically assigning workloads to this scaling policy.
- Confidence
Workload
Scaling Policy Confidence Args - Defines the confidence settings for applying recommendations.
- Downscaling
Workload
Scaling Policy Downscaling Args - Memory
Event WorkloadScaling Policy Memory Event Args - Name string
- Scaling policy name
- Predictive
Scaling WorkloadScaling Policy Predictive Scaling Args - Rollout
Behavior WorkloadScaling Policy Rollout Behavior Args - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- Startup
Workload
Scaling Policy Startup Args
- apply
Type String - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- cluster
Id String - CAST AI cluster id
- cpu
Workload
Scaling Policy Cpu - management
Option String - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
Workload
Scaling Policy Memory - anti
Affinity WorkloadScaling Policy Anti Affinity - assignment
Rules List<WorkloadScaling Policy Assignment Rule> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- confidence
Workload
Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- downscaling
Workload
Scaling Policy Downscaling - memory
Event WorkloadScaling Policy Memory Event - name String
- Scaling policy name
- predictive
Scaling WorkloadScaling Policy Predictive Scaling - rollout
Behavior WorkloadScaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
Workload
Scaling Policy Startup
- apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- cluster
Id string - CAST AI cluster id
- cpu
workload
Workload Scaling Policy Cpu - management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
workload
Workload Scaling Policy Memory - anti
Affinity workloadWorkload Scaling Policy Anti Affinity - assignment
Rules workloadWorkload Scaling Policy Assignment Rule[] - Allows defining conditions for automatically assigning workloads to this scaling policy.
- confidence
workload
Workload Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- downscaling
workload
Workload Scaling Policy Downscaling - memory
Event workloadWorkload Scaling Policy Memory Event - name string
- Scaling policy name
- predictive
Scaling workloadWorkload Scaling Policy Predictive Scaling - rollout
Behavior workloadWorkload Scaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
workload
Workload Scaling Policy Startup
- apply_
type str - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- cluster_
id str - CAST AI cluster id
- cpu
workload.
Workload Scaling Policy Cpu Args - management_
option str - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
workload.
Workload Scaling Policy Memory Args - anti_
affinity workload.Workload Scaling Policy Anti Affinity Args - assignment_
rules Sequence[workload.Workload Scaling Policy Assignment Rule Args] - Allows defining conditions for automatically assigning workloads to this scaling policy.
- confidence
workload.
Workload Scaling Policy Confidence Args - Defines the confidence settings for applying recommendations.
- downscaling
workload.
Workload Scaling Policy Downscaling Args - memory_
event workload.Workload Scaling Policy Memory Event Args - name str
- Scaling policy name
- predictive_
scaling workload.Workload Scaling Policy Predictive Scaling Args - rollout_
behavior workload.Workload Scaling Policy Rollout Behavior Args - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
workload.
Workload Scaling Policy Startup Args
- apply
Type String - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- cluster
Id String - CAST AI cluster id
- cpu Property Map
- management
Option String - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory Property Map
- anti
Affinity Property Map - assignment
Rules List<Property Map> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- confidence Property Map
- Defines the confidence settings for applying recommendations.
- downscaling Property Map
- memory
Event Property Map - name String
- Scaling policy name
- predictive
Scaling Property Map - rollout
Behavior Property Map - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup Property Map
Outputs
All input properties are implicitly available as output properties. Additionally, the WorkloadScalingPolicy resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing WorkloadScalingPolicy Resource
Get an existing WorkloadScalingPolicy resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: WorkloadScalingPolicyState, opts?: CustomResourceOptions): WorkloadScalingPolicy@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
anti_affinity: Optional[_workload.WorkloadScalingPolicyAntiAffinityArgs] = None,
apply_type: Optional[str] = None,
assignment_rules: Optional[Sequence[_workload.WorkloadScalingPolicyAssignmentRuleArgs]] = None,
cluster_id: Optional[str] = None,
confidence: Optional[_workload.WorkloadScalingPolicyConfidenceArgs] = None,
cpu: Optional[_workload.WorkloadScalingPolicyCpuArgs] = None,
downscaling: Optional[_workload.WorkloadScalingPolicyDownscalingArgs] = None,
management_option: Optional[str] = None,
memory: Optional[_workload.WorkloadScalingPolicyMemoryArgs] = None,
memory_event: Optional[_workload.WorkloadScalingPolicyMemoryEventArgs] = None,
name: Optional[str] = None,
predictive_scaling: Optional[_workload.WorkloadScalingPolicyPredictiveScalingArgs] = None,
rollout_behavior: Optional[_workload.WorkloadScalingPolicyRolloutBehaviorArgs] = None,
startup: Optional[_workload.WorkloadScalingPolicyStartupArgs] = None) -> WorkloadScalingPolicyfunc GetWorkloadScalingPolicy(ctx *Context, name string, id IDInput, state *WorkloadScalingPolicyState, opts ...ResourceOption) (*WorkloadScalingPolicy, error)public static WorkloadScalingPolicy Get(string name, Input<string> id, WorkloadScalingPolicyState? state, CustomResourceOptions? opts = null)public static WorkloadScalingPolicy get(String name, Output<String> id, WorkloadScalingPolicyState state, CustomResourceOptions options)resources: _: type: castai:workload:WorkloadScalingPolicy get: id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Anti
Affinity Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Anti Affinity - Apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Assignment
Rules List<Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Assignment Rule> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- Cluster
Id string - CAST AI cluster id
- Confidence
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- Cpu
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Cpu - Downscaling
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Downscaling - Management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- Memory
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Memory - Memory
Event Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Memory Event - Name string
- Scaling policy name
- Predictive
Scaling Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Predictive Scaling - Rollout
Behavior Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- Startup
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Startup
- Anti
Affinity WorkloadScaling Policy Anti Affinity Args - Apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Assignment
Rules WorkloadScaling Policy Assignment Rule Args - Allows defining conditions for automatically assigning workloads to this scaling policy.
- Cluster
Id string - CAST AI cluster id
- Confidence
Workload
Scaling Policy Confidence Args - Defines the confidence settings for applying recommendations.
- Cpu
Workload
Scaling Policy Cpu Args - Downscaling
Workload
Scaling Policy Downscaling Args - Management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- Memory
Workload
Scaling Policy Memory Args - Memory
Event WorkloadScaling Policy Memory Event Args - Name string
- Scaling policy name
- Predictive
Scaling WorkloadScaling Policy Predictive Scaling Args - Rollout
Behavior WorkloadScaling Policy Rollout Behavior Args - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- Startup
Workload
Scaling Policy Startup Args
- anti
Affinity WorkloadScaling Policy Anti Affinity - apply
Type String - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- assignment
Rules List<WorkloadScaling Policy Assignment Rule> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- cluster
Id String - CAST AI cluster id
- confidence
Workload
Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- cpu
Workload
Scaling Policy Cpu - downscaling
Workload
Scaling Policy Downscaling - management
Option String - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
Workload
Scaling Policy Memory - memory
Event WorkloadScaling Policy Memory Event - name String
- Scaling policy name
- predictive
Scaling WorkloadScaling Policy Predictive Scaling - rollout
Behavior WorkloadScaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
Workload
Scaling Policy Startup
- anti
Affinity workloadWorkload Scaling Policy Anti Affinity - apply
Type string - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- assignment
Rules workloadWorkload Scaling Policy Assignment Rule[] - Allows defining conditions for automatically assigning workloads to this scaling policy.
- cluster
Id string - CAST AI cluster id
- confidence
workload
Workload Scaling Policy Confidence - Defines the confidence settings for applying recommendations.
- cpu
workload
Workload Scaling Policy Cpu - downscaling
workload
Workload Scaling Policy Downscaling - management
Option string - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
workload
Workload Scaling Policy Memory - memory
Event workloadWorkload Scaling Policy Memory Event - name string
- Scaling policy name
- predictive
Scaling workloadWorkload Scaling Policy Predictive Scaling - rollout
Behavior workloadWorkload Scaling Policy Rollout Behavior - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
workload
Workload Scaling Policy Startup
- anti_
affinity workload.Workload Scaling Policy Anti Affinity Args - apply_
type str - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- assignment_
rules Sequence[workload.Workload Scaling Policy Assignment Rule Args] - Allows defining conditions for automatically assigning workloads to this scaling policy.
- cluster_
id str - CAST AI cluster id
- confidence
workload.
Workload Scaling Policy Confidence Args - Defines the confidence settings for applying recommendations.
- cpu
workload.
Workload Scaling Policy Cpu Args - downscaling
workload.
Workload Scaling Policy Downscaling Args - management_
option str - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory
workload.
Workload Scaling Policy Memory Args - memory_
event workload.Workload Scaling Policy Memory Event Args - name str
- Scaling policy name
- predictive_
scaling workload.Workload Scaling Policy Predictive Scaling Args - rollout_
behavior workload.Workload Scaling Policy Rollout Behavior Args - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup
workload.
Workload Scaling Policy Startup Args
- anti
Affinity Property Map - apply
Type String - Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- assignment
Rules List<Property Map> - Allows defining conditions for automatically assigning workloads to this scaling policy.
- cluster
Id String - CAST AI cluster id
- confidence Property Map
- Defines the confidence settings for applying recommendations.
- cpu Property Map
- downscaling Property Map
- management
Option String - Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
- memory Property Map
- memory
Event Property Map - name String
- Scaling policy name
- predictive
Scaling Property Map - rollout
Behavior Property Map - Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
- startup Property Map
Supporting Types
WorkloadScalingPolicyAntiAffinity, WorkloadScalingPolicyAntiAffinityArgs
- Consider
Anti boolAffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
- Consider
Anti boolAffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
- consider
Anti BooleanAffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
- consider
Anti booleanAffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
- consider_
anti_ boolaffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
- consider
Anti BooleanAffinity - Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
WorkloadScalingPolicyAssignmentRule, WorkloadScalingPolicyAssignmentRuleArgs
WorkloadScalingPolicyAssignmentRuleRule, WorkloadScalingPolicyAssignmentRuleRuleArgs
- Namespace
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Assignment Rule Rule Namespace - Allows assigning a scaling policy based on the workload's namespace.
- Workload
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Assignment Rule Rule Workload - Allows assigning a scaling policy based on the workload's metadata.
- Namespace
Workload
Scaling Policy Assignment Rule Rule Namespace - Allows assigning a scaling policy based on the workload's namespace.
- Workload
Workload
Scaling Policy Assignment Rule Rule Workload - Allows assigning a scaling policy based on the workload's metadata.
- namespace
Workload
Scaling Policy Assignment Rule Rule Namespace - Allows assigning a scaling policy based on the workload's namespace.
- workload
Workload
Scaling Policy Assignment Rule Rule Workload - Allows assigning a scaling policy based on the workload's metadata.
- namespace
workload
Workload Scaling Policy Assignment Rule Rule Namespace - Allows assigning a scaling policy based on the workload's namespace.
- workload
workload
Workload Scaling Policy Assignment Rule Rule Workload - Allows assigning a scaling policy based on the workload's metadata.
- namespace
workload.
Workload Scaling Policy Assignment Rule Rule Namespace - Allows assigning a scaling policy based on the workload's namespace.
- workload
workload.
Workload Scaling Policy Assignment Rule Rule Workload - Allows assigning a scaling policy based on the workload's metadata.
- namespace Property Map
- Allows assigning a scaling policy based on the workload's namespace.
- workload Property Map
- Allows assigning a scaling policy based on the workload's metadata.
WorkloadScalingPolicyAssignmentRuleRuleNamespace, WorkloadScalingPolicyAssignmentRuleRuleNamespaceArgs
- Names List<string>
- Defines matching by namespace names.
- Names []string
- Defines matching by namespace names.
- names List<String>
- Defines matching by namespace names.
- names string[]
- Defines matching by namespace names.
- names Sequence[str]
- Defines matching by namespace names.
- names List<String>
- Defines matching by namespace names.
WorkloadScalingPolicyAssignmentRuleRuleWorkload, WorkloadScalingPolicyAssignmentRuleRuleWorkloadArgs
- Gvks List<string>
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- Labels
Expressions List<Pulumi.Cast AI. Workload. Inputs. Workload Scaling Policy Assignment Rule Rule Workload Labels Expression> - Defines matching by label selector requirements.
- Gvks []string
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- Labels
Expressions WorkloadScaling Policy Assignment Rule Rule Workload Labels Expression - Defines matching by label selector requirements.
- gvks List<String>
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- labels
Expressions List<WorkloadScaling Policy Assignment Rule Rule Workload Labels Expression> - Defines matching by label selector requirements.
- gvks string[]
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- labels
Expressions workloadWorkload Scaling Policy Assignment Rule Rule Workload Labels Expression[] - Defines matching by label selector requirements.
- gvks Sequence[str]
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- labels_
expressions Sequence[workload.Workload Scaling Policy Assignment Rule Rule Workload Labels Expression] - Defines matching by label selector requirements.
- gvks List<String>
- Group, version, and kind for Kubernetes resources. Format: kind[.version][.group].
It can be either:
- only kind, e.g. "Deployment"
- group and kind: e.g."Deployment.apps"
- group, version and kind: e.g."Deployment.v1.apps"
- labels
Expressions List<Property Map> - Defines matching by label selector requirements.
WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression, WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpressionArgs
- Operator string
- The operator to use for matching the label.
- Key string
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - Values List<string>
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
- Operator string
- The operator to use for matching the label.
- Key string
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - Values []string
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
- operator String
- The operator to use for matching the label.
- key String
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - values List<String>
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
- operator string
- The operator to use for matching the label.
- key string
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - values string[]
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
- operator str
- The operator to use for matching the label.
- key str
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - values Sequence[str]
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
- operator String
- The operator to use for matching the label.
- key String
- The label key to match. Required for all operators except
RegexandContains. If not specified, it will search through all labels. - values List<String>
- A list of values to match against the label key. It is required for
In,NotIn,Regex, andContainsoperators.
WorkloadScalingPolicyConfidence, WorkloadScalingPolicyConfidenceArgs
- Threshold double
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
- Threshold float64
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
- threshold Double
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
- threshold number
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
- threshold float
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
- threshold Number
- Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
WorkloadScalingPolicyCpu, WorkloadScalingPolicyCpuArgs
- Apply
Threshold double - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- Apply
Threshold Pulumi.Strategy Cast AI. Workload. Inputs. Workload Scaling Policy Cpu Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - Args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - Function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - Limit
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Cpu Limit - Resource limit settings
- Look
Back intPeriod Seconds - The look back period in seconds for the recommendation.
- Management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - Max double
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Min double
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Overhead double
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- Apply
Threshold float64 - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- Apply
Threshold WorkloadStrategy Scaling Policy Cpu Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - Args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - Function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - Limit
Workload
Scaling Policy Cpu Limit - Resource limit settings
- Look
Back intPeriod Seconds - The look back period in seconds for the recommendation.
- Management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - Max float64
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Min float64
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Overhead float64
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold Double - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold WorkloadStrategy Scaling Policy Cpu Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args String
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function String
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
Workload
Scaling Policy Cpu Limit - Resource limit settings
- look
Back IntegerPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option String - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max Double
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min Double
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead Double
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold number - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold workloadStrategy Workload Scaling Policy Cpu Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
workload
Workload Scaling Policy Cpu Limit - Resource limit settings
- look
Back numberPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max number
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min number
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead number
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply_
threshold float - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply_
threshold_ workload.strategy Workload Scaling Policy Cpu Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args str
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function str
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
workload.
Workload Scaling Policy Cpu Limit - Resource limit settings
- look_
back_ intperiod_ seconds - The look back period in seconds for the recommendation.
- management_
option str - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max float
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min float
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead float
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold Number - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold Property MapStrategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args String
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function String
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit Property Map
- Resource limit settings
- look
Back NumberPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option String - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max Number
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min Number
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead Number
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
WorkloadScalingPolicyCpuApplyThresholdStrategy, WorkloadScalingPolicyCpuApplyThresholdStrategyArgs
- Type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- Denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Exponent double
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- Numerator double
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Percentage double
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- Type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- Denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Exponent float64
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- Numerator float64
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Percentage float64
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type String
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator String
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent Double
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator Double
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage Double
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent number
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator number
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage number
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type str
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator str
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent float
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator float
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage float
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type String
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator String
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent Number
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator Number
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage Number
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
WorkloadScalingPolicyCpuLimit, WorkloadScalingPolicyCpuLimitArgs
- Type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- Multiplier double
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- Type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- Multiplier float64
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type String
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier Double
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier number
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type str
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier float
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type String
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier Number
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
WorkloadScalingPolicyDownscaling, WorkloadScalingPolicyDownscalingArgs
- Apply
Type string - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Apply
Type string - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type String - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type string - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply_
type str - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type String - Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
WorkloadScalingPolicyMemory, WorkloadScalingPolicyMemoryArgs
- Apply
Threshold double - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- Apply
Threshold Pulumi.Strategy Cast AI. Workload. Inputs. Workload Scaling Policy Memory Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - Args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - Function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - Limit
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Memory Limit - Resource limit settings
- Look
Back intPeriod Seconds - The look back period in seconds for the recommendation.
- Management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - Max double
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Min double
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Overhead double
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- Apply
Threshold float64 - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- Apply
Threshold WorkloadStrategy Scaling Policy Memory Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - Args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - Function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - Limit
Workload
Scaling Policy Memory Limit - Resource limit settings
- Look
Back intPeriod Seconds - The look back period in seconds for the recommendation.
- Management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - Max float64
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Min float64
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- Overhead float64
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold Double - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold WorkloadStrategy Scaling Policy Memory Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args String
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function String
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
Workload
Scaling Policy Memory Limit - Resource limit settings
- look
Back IntegerPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option String - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max Double
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min Double
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead Double
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold number - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold workloadStrategy Workload Scaling Policy Memory Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args string
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function string
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
workload
Workload Scaling Policy Memory Limit - Resource limit settings
- look
Back numberPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option string - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max number
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min number
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead number
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply_
threshold float - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply_
threshold_ workload.strategy Workload Scaling Policy Memory Apply Threshold Strategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args str
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function str
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit
workload.
Workload Scaling Policy Memory Limit - Resource limit settings
- look_
back_ intperiod_ seconds - The look back period in seconds for the recommendation.
- management_
option str - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max float
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min float
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead float
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
- apply
Threshold Number - The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value
- apply
Threshold Property MapStrategy - Resource apply threshold strategy settings. The default strategy is
PERCENTAGEwith percentage value set to 0.1. - args String
- The arguments for the function - i.e. for
QUANTILEthis should be a [0, 1] float.MAXdoesn't accept any args - function String
- The function used to calculate the resource recommendation. Supported values:
QUANTILE,MAX - limit Property Map
- Resource limit settings
- look
Back NumberPeriod Seconds - The look back period in seconds for the recommendation.
- management
Option String - Disables management for a single resource when set to
READ_ONLY. The resource will use its original workload template requests and limits. Supported value:READ_ONLY. Minimum required workload-autoscaler version:v0.23.1. - max Number
- Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- min Number
- Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
- overhead Number
- Overhead for the recommendation, e.g.
0.1will result in 10% higher recommendation
WorkloadScalingPolicyMemoryApplyThresholdStrategy, WorkloadScalingPolicyMemoryApplyThresholdStrategyArgs
- Type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- Denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Exponent double
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- Numerator double
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Percentage double
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- Type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- Denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Exponent float64
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- Numerator float64
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- Percentage float64
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type String
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator String
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent Double
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator Double
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage Double
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type string
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator string
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent number
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator number
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage number
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type str
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator str
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent float
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator float
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage float
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
- type String
- Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
- denominator String
- If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
- exponent Number
- The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
- numerator Number
- The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
- percentage Number
- Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
WorkloadScalingPolicyMemoryEvent, WorkloadScalingPolicyMemoryEventArgs
- Apply
Type string - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- Apply
Type string - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type String - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type string - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply_
type str - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
- apply
Type String - Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
WorkloadScalingPolicyMemoryLimit, WorkloadScalingPolicyMemoryLimitArgs
- Type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- Multiplier double
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- Type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- Multiplier float64
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type String
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier Double
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type string
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier number
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type str
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier float
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
- type String
- Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
- multiplier Number
- Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
WorkloadScalingPolicyPredictiveScaling, WorkloadScalingPolicyPredictiveScalingArgs
- Cpu
Pulumi.
Cast AI. Workload. Inputs. Workload Scaling Policy Predictive Scaling Cpu - Defines predictive scaling resource configuration.
- Cpu
Workload
Scaling Policy Predictive Scaling Cpu - Defines predictive scaling resource configuration.
- cpu
Workload
Scaling Policy Predictive Scaling Cpu - Defines predictive scaling resource configuration.
- cpu
workload
Workload Scaling Policy Predictive Scaling Cpu - Defines predictive scaling resource configuration.
- cpu
workload.
Workload Scaling Policy Predictive Scaling Cpu - Defines predictive scaling resource configuration.
- cpu Property Map
- Defines predictive scaling resource configuration.
WorkloadScalingPolicyPredictiveScalingCpu, WorkloadScalingPolicyPredictiveScalingCpuArgs
- Enabled bool
- Defines if predictive scaling is enabled for resource.
- Enabled bool
- Defines if predictive scaling is enabled for resource.
- enabled Boolean
- Defines if predictive scaling is enabled for resource.
- enabled boolean
- Defines if predictive scaling is enabled for resource.
- enabled bool
- Defines if predictive scaling is enabled for resource.
- enabled Boolean
- Defines if predictive scaling is enabled for resource.
WorkloadScalingPolicyRolloutBehavior, WorkloadScalingPolicyRolloutBehaviorArgs
- Prefer
One boolBy One - Defines if pods should be restarted one by one to avoid service disruption.
- Type string
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
- Prefer
One boolBy One - Defines if pods should be restarted one by one to avoid service disruption.
- Type string
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
- prefer
One BooleanBy One - Defines if pods should be restarted one by one to avoid service disruption.
- type String
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
- prefer
One booleanBy One - Defines if pods should be restarted one by one to avoid service disruption.
- type string
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
- prefer_
one_ boolby_ one - Defines if pods should be restarted one by one to avoid service disruption.
- type str
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
- prefer
One BooleanBy One - Defines if pods should be restarted one by one to avoid service disruption.
- type String
- Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
WorkloadScalingPolicyStartup, WorkloadScalingPolicyStartupArgs
- Period
Seconds int - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
- Period
Seconds int - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
- period
Seconds Integer - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
- period
Seconds number - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
- period_
seconds int - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
- period
Seconds Number - Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
Package Details
- Repository
- castai castai/pulumi-castai
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
castaiTerraform Provider.
