1. Packages
  2. CAST AI
  3. API Docs
  4. WorkloadScalingPolicy
CAST AI v7.73.2 published on Wednesday, Oct 29, 2025 by CAST AI

castai.WorkloadScalingPolicy

Start a Neo task
Explain and create a castai.WorkloadScalingPolicy resource
castai logo
CAST AI v7.73.2 published on Wednesday, Oct 29, 2025 by CAST AI

    Create WorkloadScalingPolicy Resource

    Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

    Constructor syntax

    new WorkloadScalingPolicy(name: string, args: WorkloadScalingPolicyArgs, opts?: CustomResourceOptions);
    @overload
    def WorkloadScalingPolicy(resource_name: str,
                              args: WorkloadScalingPolicyArgs,
                              opts: Optional[ResourceOptions] = None)
    
    @overload
    def WorkloadScalingPolicy(resource_name: str,
                              opts: Optional[ResourceOptions] = None,
                              cpu: Optional[_workload.WorkloadScalingPolicyCpuArgs] = None,
                              apply_type: Optional[str] = None,
                              memory: Optional[_workload.WorkloadScalingPolicyMemoryArgs] = None,
                              cluster_id: Optional[str] = None,
                              management_option: Optional[str] = None,
                              downscaling: Optional[_workload.WorkloadScalingPolicyDownscalingArgs] = None,
                              anti_affinity: Optional[_workload.WorkloadScalingPolicyAntiAffinityArgs] = None,
                              confidence: Optional[_workload.WorkloadScalingPolicyConfidenceArgs] = None,
                              assignment_rules: Optional[Sequence[_workload.WorkloadScalingPolicyAssignmentRuleArgs]] = None,
                              memory_event: Optional[_workload.WorkloadScalingPolicyMemoryEventArgs] = None,
                              name: Optional[str] = None,
                              predictive_scaling: Optional[_workload.WorkloadScalingPolicyPredictiveScalingArgs] = None,
                              rollout_behavior: Optional[_workload.WorkloadScalingPolicyRolloutBehaviorArgs] = None,
                              startup: Optional[_workload.WorkloadScalingPolicyStartupArgs] = None)
    func NewWorkloadScalingPolicy(ctx *Context, name string, args WorkloadScalingPolicyArgs, opts ...ResourceOption) (*WorkloadScalingPolicy, error)
    public WorkloadScalingPolicy(string name, WorkloadScalingPolicyArgs args, CustomResourceOptions? opts = null)
    public WorkloadScalingPolicy(String name, WorkloadScalingPolicyArgs args)
    public WorkloadScalingPolicy(String name, WorkloadScalingPolicyArgs args, CustomResourceOptions options)
    
    type: castai:workload:WorkloadScalingPolicy
    properties: # The arguments to resource properties.
    options: # Bag of options to control resource's behavior.
    
    

    Parameters

    name string
    The unique name of the resource.
    args WorkloadScalingPolicyArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    resource_name str
    The unique name of the resource.
    args WorkloadScalingPolicyArgs
    The arguments to resource properties.
    opts ResourceOptions
    Bag of options to control resource's behavior.
    ctx Context
    Context object for the current deployment.
    name string
    The unique name of the resource.
    args WorkloadScalingPolicyArgs
    The arguments to resource properties.
    opts ResourceOption
    Bag of options to control resource's behavior.
    name string
    The unique name of the resource.
    args WorkloadScalingPolicyArgs
    The arguments to resource properties.
    opts CustomResourceOptions
    Bag of options to control resource's behavior.
    name String
    The unique name of the resource.
    args WorkloadScalingPolicyArgs
    The arguments to resource properties.
    options CustomResourceOptions
    Bag of options to control resource's behavior.

    WorkloadScalingPolicy Resource Properties

    To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

    Inputs

    In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

    The WorkloadScalingPolicy resource accepts the following input properties:

    ApplyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    ClusterId string
    CAST AI cluster id
    Cpu Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyCpu
    ManagementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    Memory Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemory
    AntiAffinity Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAntiAffinity
    AssignmentRules List<Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAssignmentRule>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    Confidence Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    Downscaling Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyDownscaling
    MemoryEvent Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemoryEvent
    Name string
    Scaling policy name
    PredictiveScaling Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyPredictiveScaling
    RolloutBehavior Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    Startup Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyStartup
    ApplyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    ClusterId string
    CAST AI cluster id
    Cpu WorkloadScalingPolicyCpuArgs
    ManagementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    Memory WorkloadScalingPolicyMemoryArgs
    AntiAffinity WorkloadScalingPolicyAntiAffinityArgs
    AssignmentRules WorkloadScalingPolicyAssignmentRuleArgs
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    Confidence WorkloadScalingPolicyConfidenceArgs
    Defines the confidence settings for applying recommendations.
    Downscaling WorkloadScalingPolicyDownscalingArgs
    MemoryEvent WorkloadScalingPolicyMemoryEventArgs
    Name string
    Scaling policy name
    PredictiveScaling WorkloadScalingPolicyPredictiveScalingArgs
    RolloutBehavior WorkloadScalingPolicyRolloutBehaviorArgs
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    Startup WorkloadScalingPolicyStartupArgs
    applyType String
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    clusterId String
    CAST AI cluster id
    cpu WorkloadScalingPolicyCpu
    managementOption String
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory WorkloadScalingPolicyMemory
    antiAffinity WorkloadScalingPolicyAntiAffinity
    assignmentRules List<WorkloadScalingPolicyAssignmentRule>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    confidence WorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    downscaling WorkloadScalingPolicyDownscaling
    memoryEvent WorkloadScalingPolicyMemoryEvent
    name String
    Scaling policy name
    predictiveScaling WorkloadScalingPolicyPredictiveScaling
    rolloutBehavior WorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup WorkloadScalingPolicyStartup
    applyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    clusterId string
    CAST AI cluster id
    cpu workloadWorkloadScalingPolicyCpu
    managementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory workloadWorkloadScalingPolicyMemory
    antiAffinity workloadWorkloadScalingPolicyAntiAffinity
    assignmentRules workloadWorkloadScalingPolicyAssignmentRule[]
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    confidence workloadWorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    downscaling workloadWorkloadScalingPolicyDownscaling
    memoryEvent workloadWorkloadScalingPolicyMemoryEvent
    name string
    Scaling policy name
    predictiveScaling workloadWorkloadScalingPolicyPredictiveScaling
    rolloutBehavior workloadWorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup workloadWorkloadScalingPolicyStartup
    apply_type str
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    cluster_id str
    CAST AI cluster id
    cpu workload.WorkloadScalingPolicyCpuArgs
    management_option str
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory workload.WorkloadScalingPolicyMemoryArgs
    anti_affinity workload.WorkloadScalingPolicyAntiAffinityArgs
    assignment_rules Sequence[workload.WorkloadScalingPolicyAssignmentRuleArgs]
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    confidence workload.WorkloadScalingPolicyConfidenceArgs
    Defines the confidence settings for applying recommendations.
    downscaling workload.WorkloadScalingPolicyDownscalingArgs
    memory_event workload.WorkloadScalingPolicyMemoryEventArgs
    name str
    Scaling policy name
    predictive_scaling workload.WorkloadScalingPolicyPredictiveScalingArgs
    rollout_behavior workload.WorkloadScalingPolicyRolloutBehaviorArgs
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup workload.WorkloadScalingPolicyStartupArgs
    applyType String
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    clusterId String
    CAST AI cluster id
    cpu Property Map
    managementOption String
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory Property Map
    antiAffinity Property Map
    assignmentRules List<Property Map>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    confidence Property Map
    Defines the confidence settings for applying recommendations.
    downscaling Property Map
    memoryEvent Property Map
    name String
    Scaling policy name
    predictiveScaling Property Map
    rolloutBehavior Property Map
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup Property Map

    Outputs

    All input properties are implicitly available as output properties. Additionally, the WorkloadScalingPolicy resource produces the following output properties:

    Id string
    The provider-assigned unique ID for this managed resource.
    Id string
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.
    id string
    The provider-assigned unique ID for this managed resource.
    id str
    The provider-assigned unique ID for this managed resource.
    id String
    The provider-assigned unique ID for this managed resource.

    Look up Existing WorkloadScalingPolicy Resource

    Get an existing WorkloadScalingPolicy resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

    public static get(name: string, id: Input<ID>, state?: WorkloadScalingPolicyState, opts?: CustomResourceOptions): WorkloadScalingPolicy
    @staticmethod
    def get(resource_name: str,
            id: str,
            opts: Optional[ResourceOptions] = None,
            anti_affinity: Optional[_workload.WorkloadScalingPolicyAntiAffinityArgs] = None,
            apply_type: Optional[str] = None,
            assignment_rules: Optional[Sequence[_workload.WorkloadScalingPolicyAssignmentRuleArgs]] = None,
            cluster_id: Optional[str] = None,
            confidence: Optional[_workload.WorkloadScalingPolicyConfidenceArgs] = None,
            cpu: Optional[_workload.WorkloadScalingPolicyCpuArgs] = None,
            downscaling: Optional[_workload.WorkloadScalingPolicyDownscalingArgs] = None,
            management_option: Optional[str] = None,
            memory: Optional[_workload.WorkloadScalingPolicyMemoryArgs] = None,
            memory_event: Optional[_workload.WorkloadScalingPolicyMemoryEventArgs] = None,
            name: Optional[str] = None,
            predictive_scaling: Optional[_workload.WorkloadScalingPolicyPredictiveScalingArgs] = None,
            rollout_behavior: Optional[_workload.WorkloadScalingPolicyRolloutBehaviorArgs] = None,
            startup: Optional[_workload.WorkloadScalingPolicyStartupArgs] = None) -> WorkloadScalingPolicy
    func GetWorkloadScalingPolicy(ctx *Context, name string, id IDInput, state *WorkloadScalingPolicyState, opts ...ResourceOption) (*WorkloadScalingPolicy, error)
    public static WorkloadScalingPolicy Get(string name, Input<string> id, WorkloadScalingPolicyState? state, CustomResourceOptions? opts = null)
    public static WorkloadScalingPolicy get(String name, Output<String> id, WorkloadScalingPolicyState state, CustomResourceOptions options)
    resources:  _:    type: castai:workload:WorkloadScalingPolicy    get:      id: ${id}
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    resource_name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    name
    The unique name of the resulting resource.
    id
    The unique provider ID of the resource to lookup.
    state
    Any extra arguments used during the lookup.
    opts
    A bag of options that control this resource's behavior.
    The following state arguments are supported:
    AntiAffinity Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAntiAffinity
    ApplyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    AssignmentRules List<Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAssignmentRule>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    ClusterId string
    CAST AI cluster id
    Confidence Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    Cpu Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyCpu
    Downscaling Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyDownscaling
    ManagementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    Memory Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemory
    MemoryEvent Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemoryEvent
    Name string
    Scaling policy name
    PredictiveScaling Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyPredictiveScaling
    RolloutBehavior Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    Startup Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyStartup
    AntiAffinity WorkloadScalingPolicyAntiAffinityArgs
    ApplyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    AssignmentRules WorkloadScalingPolicyAssignmentRuleArgs
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    ClusterId string
    CAST AI cluster id
    Confidence WorkloadScalingPolicyConfidenceArgs
    Defines the confidence settings for applying recommendations.
    Cpu WorkloadScalingPolicyCpuArgs
    Downscaling WorkloadScalingPolicyDownscalingArgs
    ManagementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    Memory WorkloadScalingPolicyMemoryArgs
    MemoryEvent WorkloadScalingPolicyMemoryEventArgs
    Name string
    Scaling policy name
    PredictiveScaling WorkloadScalingPolicyPredictiveScalingArgs
    RolloutBehavior WorkloadScalingPolicyRolloutBehaviorArgs
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    Startup WorkloadScalingPolicyStartupArgs
    antiAffinity WorkloadScalingPolicyAntiAffinity
    applyType String
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    assignmentRules List<WorkloadScalingPolicyAssignmentRule>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    clusterId String
    CAST AI cluster id
    confidence WorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    cpu WorkloadScalingPolicyCpu
    downscaling WorkloadScalingPolicyDownscaling
    managementOption String
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory WorkloadScalingPolicyMemory
    memoryEvent WorkloadScalingPolicyMemoryEvent
    name String
    Scaling policy name
    predictiveScaling WorkloadScalingPolicyPredictiveScaling
    rolloutBehavior WorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup WorkloadScalingPolicyStartup
    antiAffinity workloadWorkloadScalingPolicyAntiAffinity
    applyType string
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    assignmentRules workloadWorkloadScalingPolicyAssignmentRule[]
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    clusterId string
    CAST AI cluster id
    confidence workloadWorkloadScalingPolicyConfidence
    Defines the confidence settings for applying recommendations.
    cpu workloadWorkloadScalingPolicyCpu
    downscaling workloadWorkloadScalingPolicyDownscaling
    managementOption string
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory workloadWorkloadScalingPolicyMemory
    memoryEvent workloadWorkloadScalingPolicyMemoryEvent
    name string
    Scaling policy name
    predictiveScaling workloadWorkloadScalingPolicyPredictiveScaling
    rolloutBehavior workloadWorkloadScalingPolicyRolloutBehavior
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup workloadWorkloadScalingPolicyStartup
    anti_affinity workload.WorkloadScalingPolicyAntiAffinityArgs
    apply_type str
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    assignment_rules Sequence[workload.WorkloadScalingPolicyAssignmentRuleArgs]
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    cluster_id str
    CAST AI cluster id
    confidence workload.WorkloadScalingPolicyConfidenceArgs
    Defines the confidence settings for applying recommendations.
    cpu workload.WorkloadScalingPolicyCpuArgs
    downscaling workload.WorkloadScalingPolicyDownscalingArgs
    management_option str
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory workload.WorkloadScalingPolicyMemoryArgs
    memory_event workload.WorkloadScalingPolicyMemoryEventArgs
    name str
    Scaling policy name
    predictive_scaling workload.WorkloadScalingPolicyPredictiveScalingArgs
    rollout_behavior workload.WorkloadScalingPolicyRolloutBehaviorArgs
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup workload.WorkloadScalingPolicyStartupArgs
    antiAffinity Property Map
    applyType String
    Recommendation apply type. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    assignmentRules List<Property Map>
    Allows defining conditions for automatically assigning workloads to this scaling policy.
    clusterId String
    CAST AI cluster id
    confidence Property Map
    Defines the confidence settings for applying recommendations.
    cpu Property Map
    downscaling Property Map
    managementOption String
    Defines possible options for workload management. - READ_ONLY - workload watched (metrics collected), but no actions performed by CAST AI. - MANAGED - workload watched (metrics collected), CAST AI may perform actions on the workload.
    memory Property Map
    memoryEvent Property Map
    name String
    Scaling policy name
    predictiveScaling Property Map
    rolloutBehavior Property Map
    Defines the rollout behavior used when applying recommendations. Prerequisites: - Applicable to Deployment resources that support running as multi-replica. - Deployment is running with single replica (replica count = 1). - Deployment's rollout strategy allows for downtime. - Recommendation apply type is "immediate". - Cluster has workload-autoscaler component version v0.35.3 or higher.
    startup Property Map

    Supporting Types

    WorkloadScalingPolicyAntiAffinity, WorkloadScalingPolicyAntiAffinityArgs

    ConsiderAntiAffinity bool
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
    ConsiderAntiAffinity bool
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
    considerAntiAffinity Boolean
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
    considerAntiAffinity boolean
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
    consider_anti_affinity bool
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.
    considerAntiAffinity Boolean
    Defines if anti-affinity should be considered when scaling the workload. If enabled, requiring host ports, or having anti-affinity on hostname will force all recommendations to be deferred.

    WorkloadScalingPolicyAssignmentRule, WorkloadScalingPolicyAssignmentRuleArgs

    WorkloadScalingPolicyAssignmentRuleRule, WorkloadScalingPolicyAssignmentRuleRuleArgs

    Namespace Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAssignmentRuleRuleNamespace
    Allows assigning a scaling policy based on the workload's namespace.
    Workload Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAssignmentRuleRuleWorkload
    Allows assigning a scaling policy based on the workload's metadata.
    Namespace WorkloadScalingPolicyAssignmentRuleRuleNamespace
    Allows assigning a scaling policy based on the workload's namespace.
    Workload WorkloadScalingPolicyAssignmentRuleRuleWorkload
    Allows assigning a scaling policy based on the workload's metadata.
    namespace WorkloadScalingPolicyAssignmentRuleRuleNamespace
    Allows assigning a scaling policy based on the workload's namespace.
    workload WorkloadScalingPolicyAssignmentRuleRuleWorkload
    Allows assigning a scaling policy based on the workload's metadata.
    namespace workloadWorkloadScalingPolicyAssignmentRuleRuleNamespace
    Allows assigning a scaling policy based on the workload's namespace.
    workload workloadWorkloadScalingPolicyAssignmentRuleRuleWorkload
    Allows assigning a scaling policy based on the workload's metadata.
    namespace workload.WorkloadScalingPolicyAssignmentRuleRuleNamespace
    Allows assigning a scaling policy based on the workload's namespace.
    workload workload.WorkloadScalingPolicyAssignmentRuleRuleWorkload
    Allows assigning a scaling policy based on the workload's metadata.
    namespace Property Map
    Allows assigning a scaling policy based on the workload's namespace.
    workload Property Map
    Allows assigning a scaling policy based on the workload's metadata.

    WorkloadScalingPolicyAssignmentRuleRuleNamespace, WorkloadScalingPolicyAssignmentRuleRuleNamespaceArgs

    Names List<string>
    Defines matching by namespace names.
    Names []string
    Defines matching by namespace names.
    names List<String>
    Defines matching by namespace names.
    names string[]
    Defines matching by namespace names.
    names Sequence[str]
    Defines matching by namespace names.
    names List<String>
    Defines matching by namespace names.

    WorkloadScalingPolicyAssignmentRuleRuleWorkload, WorkloadScalingPolicyAssignmentRuleRuleWorkloadArgs

    Gvks List<string>
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    LabelsExpressions List<Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression>
    Defines matching by label selector requirements.
    Gvks []string
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    LabelsExpressions WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression
    Defines matching by label selector requirements.
    gvks List<String>
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    labelsExpressions List<WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression>
    Defines matching by label selector requirements.
    gvks string[]
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    labelsExpressions workloadWorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression[]
    Defines matching by label selector requirements.
    gvks Sequence[str]
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    labels_expressions Sequence[workload.WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression]
    Defines matching by label selector requirements.
    gvks List<String>
    Group, version, and kind for Kubernetes resources. Format: kind[.version][.group]. It can be either:

    • only kind, e.g. "Deployment"
    • group and kind: e.g."Deployment.apps"
    • group, version and kind: e.g."Deployment.v1.apps"
    labelsExpressions List<Property Map>
    Defines matching by label selector requirements.

    WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpression, WorkloadScalingPolicyAssignmentRuleRuleWorkloadLabelsExpressionArgs

    Operator string
    The operator to use for matching the label.
    Key string
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    Values List<string>
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.
    Operator string
    The operator to use for matching the label.
    Key string
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    Values []string
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.
    operator String
    The operator to use for matching the label.
    key String
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    values List<String>
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.
    operator string
    The operator to use for matching the label.
    key string
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    values string[]
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.
    operator str
    The operator to use for matching the label.
    key str
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    values Sequence[str]
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.
    operator String
    The operator to use for matching the label.
    key String
    The label key to match. Required for all operators except Regex and Contains. If not specified, it will search through all labels.
    values List<String>
    A list of values to match against the label key. It is required for In, NotIn, Regex, and Contains operators.

    WorkloadScalingPolicyConfidence, WorkloadScalingPolicyConfidenceArgs

    Threshold double
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
    Threshold float64
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
    threshold Double
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
    threshold number
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
    threshold float
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).
    threshold Number
    Defines the confidence threshold for applying recommendations. The smaller number indicates that we require fewer metrics data points to apply recommendations - changing this value can cause applying less precise recommendations. Do not change the default unless you want to optimize with fewer data points (e.g., short-lived workloads).

    WorkloadScalingPolicyCpu, WorkloadScalingPolicyCpuArgs

    ApplyThreshold double
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    ApplyThresholdStrategy Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyCpuApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    Args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    Function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    Limit Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyCpuLimit
    Resource limit settings
    LookBackPeriodSeconds int
    The look back period in seconds for the recommendation.
    ManagementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    Max double
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Min double
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Overhead double
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    ApplyThreshold float64
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    ApplyThresholdStrategy WorkloadScalingPolicyCpuApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    Args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    Function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    Limit WorkloadScalingPolicyCpuLimit
    Resource limit settings
    LookBackPeriodSeconds int
    The look back period in seconds for the recommendation.
    ManagementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    Max float64
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Min float64
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Overhead float64
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold Double
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy WorkloadScalingPolicyCpuApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args String
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function String
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit WorkloadScalingPolicyCpuLimit
    Resource limit settings
    lookBackPeriodSeconds Integer
    The look back period in seconds for the recommendation.
    managementOption String
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max Double
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min Double
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead Double
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold number
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy workloadWorkloadScalingPolicyCpuApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit workloadWorkloadScalingPolicyCpuLimit
    Resource limit settings
    lookBackPeriodSeconds number
    The look back period in seconds for the recommendation.
    managementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max number
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min number
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead number
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    apply_threshold float
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    apply_threshold_strategy workload.WorkloadScalingPolicyCpuApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args str
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function str
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit workload.WorkloadScalingPolicyCpuLimit
    Resource limit settings
    look_back_period_seconds int
    The look back period in seconds for the recommendation.
    management_option str
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max float
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min float
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead float
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold Number
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy Property Map
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args String
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function String
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit Property Map
    Resource limit settings
    lookBackPeriodSeconds Number
    The look back period in seconds for the recommendation.
    managementOption String
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max Number
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min Number
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead Number
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation

    WorkloadScalingPolicyCpuApplyThresholdStrategy, WorkloadScalingPolicyCpuApplyThresholdStrategyArgs

    Type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    Denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Exponent double
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    Numerator double
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Percentage double
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    Type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    Denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Exponent float64
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    Numerator float64
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Percentage float64
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type String
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator String
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent Double
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator Double
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage Double
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent number
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator number
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage number
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type str
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator str
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent float
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator float
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage float
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type String
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator String
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent Number
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator Number
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage Number
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.

    WorkloadScalingPolicyCpuLimit, WorkloadScalingPolicyCpuLimitArgs

    Type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    Multiplier double
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    Type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    Multiplier float64
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type String
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier Double
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier number
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type str
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier float
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type String
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier Number
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.

    WorkloadScalingPolicyDownscaling, WorkloadScalingPolicyDownscalingArgs

    ApplyType string
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    ApplyType string
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType String
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType string
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    apply_type str
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType String
    Defines the apply type to be used when downscaling. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)

    WorkloadScalingPolicyMemory, WorkloadScalingPolicyMemoryArgs

    ApplyThreshold double
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    ApplyThresholdStrategy Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemoryApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    Args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    Function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    Limit Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyMemoryLimit
    Resource limit settings
    LookBackPeriodSeconds int
    The look back period in seconds for the recommendation.
    ManagementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    Max double
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Min double
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Overhead double
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    ApplyThreshold float64
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    ApplyThresholdStrategy WorkloadScalingPolicyMemoryApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    Args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    Function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    Limit WorkloadScalingPolicyMemoryLimit
    Resource limit settings
    LookBackPeriodSeconds int
    The look back period in seconds for the recommendation.
    ManagementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    Max float64
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Min float64
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    Overhead float64
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold Double
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy WorkloadScalingPolicyMemoryApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args String
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function String
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit WorkloadScalingPolicyMemoryLimit
    Resource limit settings
    lookBackPeriodSeconds Integer
    The look back period in seconds for the recommendation.
    managementOption String
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max Double
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min Double
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead Double
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold number
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy workloadWorkloadScalingPolicyMemoryApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args string
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function string
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit workloadWorkloadScalingPolicyMemoryLimit
    Resource limit settings
    lookBackPeriodSeconds number
    The look back period in seconds for the recommendation.
    managementOption string
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max number
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min number
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead number
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    apply_threshold float
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    apply_threshold_strategy workload.WorkloadScalingPolicyMemoryApplyThresholdStrategy
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args str
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function str
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit workload.WorkloadScalingPolicyMemoryLimit
    Resource limit settings
    look_back_period_seconds int
    The look back period in seconds for the recommendation.
    management_option str
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max float
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min float
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead float
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation
    applyThreshold Number
    The threshold of when to apply the recommendation. Recommendation will be applied when diff of current requests and new recommendation is greater than set value

    Deprecated: Use apply_threshold_strategy instead

    applyThresholdStrategy Property Map
    Resource apply threshold strategy settings. The default strategy is PERCENTAGE with percentage value set to 0.1.
    args String
    The arguments for the function - i.e. for QUANTILE this should be a [0, 1] float. MAX doesn't accept any args
    function String
    The function used to calculate the resource recommendation. Supported values: QUANTILE, MAX
    limit Property Map
    Resource limit settings
    lookBackPeriodSeconds Number
    The look back period in seconds for the recommendation.
    managementOption String
    Disables management for a single resource when set to READ_ONLY. The resource will use its original workload template requests and limits. Supported value: READ_ONLY. Minimum required workload-autoscaler version: v0.23.1.
    max Number
    Max values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    min Number
    Min values for the recommendation, applies to every container. For memory - this is in MiB, for CPU - this is in cores.
    overhead Number
    Overhead for the recommendation, e.g. 0.1 will result in 10% higher recommendation

    WorkloadScalingPolicyMemoryApplyThresholdStrategy, WorkloadScalingPolicyMemoryApplyThresholdStrategyArgs

    Type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    Denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Exponent double
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    Numerator double
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Percentage double
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    Type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    Denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Exponent float64
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    Numerator float64
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    Percentage float64
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type String
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator String
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent Double
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator Double
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage Double
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type string
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator string
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent number
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator number
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage number
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type str
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator str
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent float
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator float
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage float
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.
    type String
    Defines apply theshold strategy type. - PERCENTAGE - recommendation will be applied when diff of current requests and new recommendation is greater than set value - DEFAULT_ADAPTIVE - will pick larger threshold percentage for small workloads and smaller percentage for large workloads. - CUSTOM_ADAPTIVE - works in same way as DEFAULT_ADAPTIVE, but it allows to tweak parameters of adaptive threshold formula: percentage = numerator/(currentRequest + denominator)^exponent. This strategy is for advance use cases, we recommend to use DEFAULT_ADAPTIVE strategy.
    denominator String
    If denominator is close or equal to 0, the threshold will be much bigger for small values.For example when numerator, exponent is 1 and denominator is 0 the threshold for 0.5 req. CPU will be 200%.It must be defined for the CUSTOM_ADAPTIVE strategy.
    exponent Number
    The exponent changes how fast the curve is going down. The smaller value will cause that we won’t pick extremely small number for big resources, for example: - if numerator is 0, denominator is 1, and exponent is 1, for 50 CPU we will pick 2% threshold - if numerator is 0, denominator is 1, and exponent is 0.8, for 50 CPU we will pick 4.3% threshold It must be defined for the CUSTOM_ADAPTIVE strategy.
    numerator Number
    The numerator affects vertical stretch of function used in adaptive threshold - smaller number will create smaller threshold.It must be defined for the CUSTOM_ADAPTIVE strategy.
    percentage Number
    Percentage of a how much difference should there be between the current pod requests and the new recommendation. It must be defined for the PERCENTAGE strategy.

    WorkloadScalingPolicyMemoryEvent, WorkloadScalingPolicyMemoryEventArgs

    ApplyType string
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    ApplyType string
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType String
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType string
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    apply_type str
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)
    applyType String
    Defines the apply type to be used when applying recommendation for memory related event. - IMMEDIATE - pods are restarted immediately when new recommendation is generated. - DEFERRED - pods are not restarted and recommendation values are applied during natural restarts only (new deployment, etc.)

    WorkloadScalingPolicyMemoryLimit, WorkloadScalingPolicyMemoryLimitArgs

    Type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    Multiplier double
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    Type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    Multiplier float64
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type String
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier Double
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type string
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier number
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type str
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier float
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.
    type String
    Defines limit strategy type. - NO_LIMIT - removes the resource limit even if it was specified in the workload spec. - KEEP_LIMITS - keep existing resource limits. While limits provide stability predictability, they may restrict workloads that need to temporarily burst beyond their allocation. - MULTIPLIER - used to calculate the resource limit. The final value is determined by multiplying the resource request by the specified factor.
    multiplier Number
    Multiplier used to calculate the resource limit. It must be defined for the MULTIPLIER strategy.

    WorkloadScalingPolicyPredictiveScaling, WorkloadScalingPolicyPredictiveScalingArgs

    Cpu Pulumi.CastAI.Workload.Inputs.WorkloadScalingPolicyPredictiveScalingCpu
    Defines predictive scaling resource configuration.
    Cpu WorkloadScalingPolicyPredictiveScalingCpu
    Defines predictive scaling resource configuration.
    cpu WorkloadScalingPolicyPredictiveScalingCpu
    Defines predictive scaling resource configuration.
    cpu workloadWorkloadScalingPolicyPredictiveScalingCpu
    Defines predictive scaling resource configuration.
    cpu workload.WorkloadScalingPolicyPredictiveScalingCpu
    Defines predictive scaling resource configuration.
    cpu Property Map
    Defines predictive scaling resource configuration.

    WorkloadScalingPolicyPredictiveScalingCpu, WorkloadScalingPolicyPredictiveScalingCpuArgs

    Enabled bool
    Defines if predictive scaling is enabled for resource.
    Enabled bool
    Defines if predictive scaling is enabled for resource.
    enabled Boolean
    Defines if predictive scaling is enabled for resource.
    enabled boolean
    Defines if predictive scaling is enabled for resource.
    enabled bool
    Defines if predictive scaling is enabled for resource.
    enabled Boolean
    Defines if predictive scaling is enabled for resource.

    WorkloadScalingPolicyRolloutBehavior, WorkloadScalingPolicyRolloutBehaviorArgs

    PreferOneByOne bool
    Defines if pods should be restarted one by one to avoid service disruption.
    Type string
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
    PreferOneByOne bool
    Defines if pods should be restarted one by one to avoid service disruption.
    Type string
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
    preferOneByOne Boolean
    Defines if pods should be restarted one by one to avoid service disruption.
    type String
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
    preferOneByOne boolean
    Defines if pods should be restarted one by one to avoid service disruption.
    type string
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
    prefer_one_by_one bool
    Defines if pods should be restarted one by one to avoid service disruption.
    type str
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.
    preferOneByOne Boolean
    Defines if pods should be restarted one by one to avoid service disruption.
    type String
    Defines the rollout type to be used when applying recommendations. - NO_DISRUPTION - pods are restarted without causing service disruption.

    WorkloadScalingPolicyStartup, WorkloadScalingPolicyStartupArgs

    PeriodSeconds int
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
    PeriodSeconds int
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
    periodSeconds Integer
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
    periodSeconds number
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
    period_seconds int
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.
    periodSeconds Number
    Defines the duration (in seconds) during which elevated resource usage is expected at startup. When set, recommendations will be adjusted to disregard resource spikes within this period. If not specified, the workload will receive standard recommendations without startup considerations.

    Package Details

    Repository
    castai castai/pulumi-castai
    License
    Apache-2.0
    Notes
    This Pulumi package is based on the castai Terraform Provider.
    castai logo
    CAST AI v7.73.2 published on Wednesday, Oct 29, 2025 by CAST AI
      Meet Neo: Your AI Platform Teammate