2022/12/08 - Auto Scaling - 3 new 4 updated api methods
Changes Adds support for metric math for target tracking scaling policies, saving you the cost and effort of publishing a custom metric to CloudWatch. Also adds support for VPC Lattice by adding the Attach/Detach/DescribeTrafficSources APIs and a new health check type to the CreateAutoScalingGroup API.
Reserved for use with Amazon VPC Lattice, which is in preview and subject to change. Do not use this API for production workloads. This API is also subject to change.
Detaches one or more traffic sources from the specified Auto Scaling group.
See also: AWS API Documentation
Request Syntax
client.detach_traffic_sources( AutoScalingGroupName='string', TrafficSources=[ { 'Identifier': 'string' }, ] )
string
[REQUIRED]
The name of the Auto Scaling group.
list
[REQUIRED]
The unique identifiers of one or more traffic sources you are detaching. You can specify up to 10 traffic sources.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group. When you detach a target group, it enters the Removing state while deregistering the instances in the group. When all instances are deregistered, then you can no longer describe the target group using the DescribeTrafficSources API call. The instances continue to run.
(dict) --
Describes the identifier of a traffic source.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group.
Identifier (string) --
The unique identifier of the traffic source.
dict
Response Syntax
{}
Response Structure
(dict) --
Reserved for use with Amazon VPC Lattice, which is in preview and subject to change. Do not use this API for production workloads. This API is also subject to change.
Attaches one or more traffic sources to the specified Auto Scaling group.
To describe the traffic sources for an Auto Scaling group, call the DescribeTrafficSources API. To detach a traffic source from the Auto Scaling group, call the DetachTrafficSources API.
This operation is additive and does not detach existing traffic sources from the Auto Scaling group.
See also: AWS API Documentation
Request Syntax
client.attach_traffic_sources( AutoScalingGroupName='string', TrafficSources=[ { 'Identifier': 'string' }, ] )
string
[REQUIRED]
The name of the Auto Scaling group.
list
[REQUIRED]
The unique identifiers of one or more traffic sources. You can specify up to 10 traffic sources.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group. Amazon EC2 Auto Scaling registers the running instances with the attached target groups. The target groups receive incoming traffic and route requests to one or more registered targets.
(dict) --
Describes the identifier of a traffic source.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group.
Identifier (string) --
The unique identifier of the traffic source.
dict
Response Syntax
{}
Response Structure
(dict) --
Reserved for use with Amazon VPC Lattice, which is in preview and subject to change. Do not use this API for production workloads. This API is also subject to change.
Gets information about the traffic sources for the specified Auto Scaling group.
See also: AWS API Documentation
Request Syntax
client.describe_traffic_sources( AutoScalingGroupName='string', TrafficSourceType='string', NextToken='string', MaxRecords=123 )
string
[REQUIRED]
The name of the Auto Scaling group.
string
[REQUIRED]
The type of traffic source you are describing. Currently, the only valid value is vpc-lattice .
string
The token for the next set of items to return. (You received this token from a previous call.)
integer
The maximum number of items to return with this call. The maximum value is 50 .
dict
Response Syntax
{ 'TrafficSources': [ { 'TrafficSource': 'string', 'State': 'string' }, ], 'NextToken': 'string' }
Response Structure
(dict) --
TrafficSources (list) --
Information about the traffic sources.
(dict) --
Describes the state of a traffic source.
TrafficSource (string) --
The unique identifier of the traffic source. Currently, this is the Amazon Resource Name (ARN) for a VPC Lattice target group.
State (string) --
The following are the possible states for a VPC Lattice target group:
Adding - The Auto Scaling instances are being registered with the target group.
Added - All Auto Scaling instances are registered with the target group.
InService - At least one Auto Scaling instance passed the VPC_LATTICE health check.
Removing - The Auto Scaling instances are being deregistered from the target group. If connection draining is enabled, VPC Lattice waits for in-flight requests to complete before deregistering the instances.
Removed - All Auto Scaling instances are deregistered from the target group.
NextToken (string) --
This string indicates that the response contains more items than can be returned in a single response. To receive additional items, specify this string for the NextToken value when requesting the next set of items. This value is null when there are no more items to return.
{'TrafficSources': [{'Identifier': 'string'}]}
We strongly recommend using a launch template when calling this operation to ensure full functionality for Amazon EC2 Auto Scaling and Amazon EC2.
Creates an Auto Scaling group with the specified name and attributes.
If you exceed your maximum limit of Auto Scaling groups, the call fails. To query this limit, call the DescribeAccountLimits API. For information about updating this limit, see Quotas for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide .
For introductory exercises for creating an Auto Scaling group, see Getting started with Amazon EC2 Auto Scaling and Tutorial: Set up a scaled and load-balanced application in the Amazon EC2 Auto Scaling User Guide . For more information, see Auto Scaling groups in the Amazon EC2 Auto Scaling User Guide .
Every Auto Scaling group has three size properties (DesiredCapacity , MaxSize , and MinSize ). Usually, you set these sizes based on a specific number of instances. However, if you configure a mixed instances policy that defines weights for the instance types, you must specify these sizes with the same units that you use for weighting instances.
See also: AWS API Documentation
Request Syntax
client.create_auto_scaling_group( AutoScalingGroupName='string', LaunchConfigurationName='string', LaunchTemplate={ 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, MixedInstancesPolicy={ 'LaunchTemplate': { 'LaunchTemplateSpecification': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'Overrides': [ { 'InstanceType': 'string', 'WeightedCapacity': 'string', 'LaunchTemplateSpecification': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'InstanceRequirements': { 'VCpuCount': { 'Min': 123, 'Max': 123 }, 'MemoryMiB': { 'Min': 123, 'Max': 123 }, 'CpuManufacturers': [ 'intel'|'amd'|'amazon-web-services', ], 'MemoryGiBPerVCpu': { 'Min': 123.0, 'Max': 123.0 }, 'ExcludedInstanceTypes': [ 'string', ], 'InstanceGenerations': [ 'current'|'previous', ], 'SpotMaxPricePercentageOverLowestPrice': 123, 'OnDemandMaxPricePercentageOverLowestPrice': 123, 'BareMetal': 'included'|'excluded'|'required', 'BurstablePerformance': 'included'|'excluded'|'required', 'RequireHibernateSupport': True|False, 'NetworkInterfaceCount': { 'Min': 123, 'Max': 123 }, 'LocalStorage': 'included'|'excluded'|'required', 'LocalStorageTypes': [ 'hdd'|'ssd', ], 'TotalLocalStorageGB': { 'Min': 123.0, 'Max': 123.0 }, 'BaselineEbsBandwidthMbps': { 'Min': 123, 'Max': 123 }, 'AcceleratorTypes': [ 'gpu'|'fpga'|'inference', ], 'AcceleratorCount': { 'Min': 123, 'Max': 123 }, 'AcceleratorManufacturers': [ 'nvidia'|'amd'|'amazon-web-services'|'xilinx', ], 'AcceleratorNames': [ 'a100'|'v100'|'k80'|'t4'|'m60'|'radeon-pro-v520'|'vu9p', ], 'AcceleratorTotalMemoryMiB': { 'Min': 123, 'Max': 123 }, 'NetworkBandwidthGbps': { 'Min': 123.0, 'Max': 123.0 }, 'AllowedInstanceTypes': [ 'string', ] } }, ] }, 'InstancesDistribution': { 'OnDemandAllocationStrategy': 'string', 'OnDemandBaseCapacity': 123, 'OnDemandPercentageAboveBaseCapacity': 123, 'SpotAllocationStrategy': 'string', 'SpotInstancePools': 123, 'SpotMaxPrice': 'string' } }, InstanceId='string', MinSize=123, MaxSize=123, DesiredCapacity=123, DefaultCooldown=123, AvailabilityZones=[ 'string', ], LoadBalancerNames=[ 'string', ], TargetGroupARNs=[ 'string', ], HealthCheckType='string', HealthCheckGracePeriod=123, PlacementGroup='string', VPCZoneIdentifier='string', TerminationPolicies=[ 'string', ], NewInstancesProtectedFromScaleIn=True|False, CapacityRebalance=True|False, LifecycleHookSpecificationList=[ { 'LifecycleHookName': 'string', 'LifecycleTransition': 'string', 'NotificationMetadata': 'string', 'HeartbeatTimeout': 123, 'DefaultResult': 'string', 'NotificationTargetARN': 'string', 'RoleARN': 'string' }, ], Tags=[ { 'ResourceId': 'string', 'ResourceType': 'string', 'Key': 'string', 'Value': 'string', 'PropagateAtLaunch': True|False }, ], ServiceLinkedRoleARN='string', MaxInstanceLifetime=123, Context='string', DesiredCapacityType='string', DefaultInstanceWarmup=123, TrafficSources=[ { 'Identifier': 'string' }, ] )
string
[REQUIRED]
The name of the Auto Scaling group. This name must be unique per Region per account.
The name can contain any ASCII character 33 to 126 including most punctuation characters, digits, and upper and lowercased letters.
Note
You cannot use a colon (:) in the name.
string
The name of the launch configuration to use to launch instances.
Conditional: You must specify either a launch template (LaunchTemplate or MixedInstancesPolicy ) or a launch configuration (LaunchConfigurationName or InstanceId ).
dict
Information used to specify the launch template and version to use to launch instances.
Conditional: You must specify either a launch template (LaunchTemplate or MixedInstancesPolicy ) or a launch configuration (LaunchConfigurationName or InstanceId ).
Note
The launch template that is specified must be configured for use with an Auto Scaling group. For more information, see Creating a launch template for an Auto Scaling group in the Amazon EC2 Auto Scaling User Guide .
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
dict
The mixed instances policy. For more information, see Auto Scaling groups with multiple instance types and purchase options in the Amazon EC2 Auto Scaling User Guide .
LaunchTemplate (dict) --
One or more launch templates and the instance types (overrides) that are used to launch EC2 instances to fulfill On-Demand and Spot capacities.
LaunchTemplateSpecification (dict) --
The launch template.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
Overrides (list) --
Any properties that you specify override the same properties in the launch template.
(dict) --
Use this structure to let Amazon EC2 Auto Scaling do the following when the Auto Scaling group has a mixed instances policy:
Override the instance type that is specified in the launch template.
Use multiple instance types.
Specify the instance types that you want, or define your instance requirements instead and let Amazon EC2 Auto Scaling provision the available instance types that meet your requirements. This can provide Amazon EC2 Auto Scaling with a larger selection of instance types to choose from when fulfilling Spot and On-Demand capacities. You can view which instance types are matched before you apply the instance requirements to your Auto Scaling group.
After you define your instance requirements, you don't have to keep updating these settings to get new EC2 instance types automatically. Amazon EC2 Auto Scaling uses the instance requirements of the Auto Scaling group to determine whether a new EC2 instance type can be used.
InstanceType (string) --
The instance type, such as m3.xlarge . You must specify an instance type that is supported in your requested Region and Availability Zones. For more information, see Instance types in the Amazon Elastic Compute Cloud User Guide .
You can specify up to 40 instance types per Auto Scaling group.
WeightedCapacity (string) --
If you provide a list of instance types to use, you can specify the number of capacity units provided by each instance type in terms of virtual CPUs, memory, storage, throughput, or other relative performance characteristic. When a Spot or On-Demand Instance is launched, the capacity units count toward the desired capacity. Amazon EC2 Auto Scaling launches instances until the desired capacity is totally fulfilled, even if this results in an overage. For example, if there are two units remaining to fulfill capacity, and Amazon EC2 Auto Scaling can only launch an instance with a WeightedCapacity of five units, the instance is launched, and the desired capacity is exceeded by three units. For more information, see Configuring instance weighting for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide . Value must be in the range of 1–999.
If you specify a value for WeightedCapacity for one instance type, you must specify a value for WeightedCapacity for all of them.
Warning
Every Auto Scaling group has three size parameters (DesiredCapacity , MaxSize , and MinSize ). Usually, you set these sizes based on a specific number of instances. However, if you configure a mixed instances policy that defines weights for the instance types, you must specify these sizes with the same units that you use for weighting instances.
LaunchTemplateSpecification (dict) --
Provides a launch template for the specified instance type or set of instance requirements. For example, some instance types might require a launch template with a different AMI. If not provided, Amazon EC2 Auto Scaling uses the launch template that's specified in the LaunchTemplate definition. For more information, see Specifying a different launch template for an instance type in the Amazon EC2 Auto Scaling User Guide .
You can specify up to 20 launch templates per Auto Scaling group. The launch templates specified in the overrides and in the LaunchTemplate definition count towards this limit.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
InstanceRequirements (dict) --
The instance requirements. Amazon EC2 Auto Scaling uses your specified requirements to identify instance types. Then, it uses your On-Demand and Spot allocation strategies to launch instances from these instance types.
You can specify up to four separate sets of instance requirements per Auto Scaling group. This is useful for provisioning instances from different Amazon Machine Images (AMIs) in the same Auto Scaling group. To do this, create the AMIs and create a new launch template for each AMI. Then, create a compatible set of instance requirements for each launch template.
Note
If you specify InstanceRequirements , you can't specify InstanceType .
VCpuCount (dict) -- [REQUIRED]
The minimum and maximum number of vCPUs for an instance type.
Min (integer) -- [REQUIRED]
The minimum number of vCPUs.
Max (integer) --
The maximum number of vCPUs.
MemoryMiB (dict) -- [REQUIRED]
The minimum and maximum instance memory size for an instance type, in MiB.
Min (integer) -- [REQUIRED]
The memory minimum in MiB.
Max (integer) --
The memory maximum in MiB.
CpuManufacturers (list) --
Lists which specific CPU manufacturers to include.
For instance types with Intel CPUs, specify intel .
For instance types with AMD CPUs, specify amd .
For instance types with Amazon Web Services CPUs, specify amazon-web-services .
Note
Don't confuse the CPU hardware manufacturer with the CPU hardware architecture. Instances will be launched with a compatible CPU architecture based on the Amazon Machine Image (AMI) that you specify in your launch template.
Default: Any manufacturer
(string) --
MemoryGiBPerVCpu (dict) --
The minimum and maximum amount of memory per vCPU for an instance type, in GiB.
Default: No minimum or maximum limits
Min (float) --
The memory minimum in GiB.
Max (float) --
The memory maximum in GiB.
ExcludedInstanceTypes (list) --
The instance types to exclude. You can use strings with one or more wild cards, represented by an asterisk (* ), to exclude an instance family, type, size, or generation. The following are examples: m5.8xlarge , c5*.* , m5a.* , r* , *3* .
For example, if you specify c5* , you are excluding the entire C5 instance family, which includes all C5a and C5n instance types. If you specify m5a.* , Amazon EC2 Auto Scaling will exclude all the M5a instance types, but not the M5n instance types.
Note
If you specify ExcludedInstanceTypes , you can't specify AllowedInstanceTypes .
Default: No excluded instance types
(string) --
InstanceGenerations (list) --
Indicates whether current or previous generation instance types are included.
For current generation instance types, specify current . The current generation includes EC2 instance types currently recommended for use. This typically includes the latest two to three generations in each instance family. For more information, see Instance types in the Amazon EC2 User Guide for Linux Instances .
For previous generation instance types, specify previous .
Default: Any current or previous generation
(string) --
SpotMaxPricePercentageOverLowestPrice (integer) --
The price protection threshold for Spot Instances. This is the maximum you’ll pay for a Spot Instance, expressed as a percentage higher than the least expensive current generation M, C, or R instance type with your specified attributes. When Amazon EC2 Auto Scaling selects instance types with your attributes, we will exclude instance types whose price is higher than your threshold. The parameter accepts an integer, which Amazon EC2 Auto Scaling interprets as a percentage. To turn off price protection, specify a high value, such as 999999 .
If you set DesiredCapacityType to vcpu or memory-mib , the price protection threshold is applied based on the per vCPU or per memory price instead of the per instance price.
Default: 100
OnDemandMaxPricePercentageOverLowestPrice (integer) --
The price protection threshold for On-Demand Instances. This is the maximum you’ll pay for an On-Demand Instance, expressed as a percentage higher than the least expensive current generation M, C, or R instance type with your specified attributes. When Amazon EC2 Auto Scaling selects instance types with your attributes, we will exclude instance types whose price is higher than your threshold. The parameter accepts an integer, which Amazon EC2 Auto Scaling interprets as a percentage. To turn off price protection, specify a high value, such as 999999 .
If you set DesiredCapacityType to vcpu or memory-mib , the price protection threshold is applied based on the per vCPU or per memory price instead of the per instance price.
Default: 20
BareMetal (string) --
Indicates whether bare metal instance types are included, excluded, or required.
Default: excluded
BurstablePerformance (string) --
Indicates whether burstable performance instance types are included, excluded, or required. For more information, see Burstable performance instances in the Amazon EC2 User Guide for Linux Instances .
Default: excluded
RequireHibernateSupport (boolean) --
Indicates whether instance types must provide On-Demand Instance hibernation support.
Default: false
NetworkInterfaceCount (dict) --
The minimum and maximum number of network interfaces for an instance type.
Default: No minimum or maximum limits
Min (integer) --
The minimum number of network interfaces.
Max (integer) --
The maximum number of network interfaces.
LocalStorage (string) --
Indicates whether instance types with instance store volumes are included, excluded, or required. For more information, see Amazon EC2 instance store in the Amazon EC2 User Guide for Linux Instances .
Default: included
LocalStorageTypes (list) --
Indicates the type of local storage that is required.
For instance types with hard disk drive (HDD) storage, specify hdd .
For instance types with solid state drive (SSD) storage, specify ssd .
Default: Any local storage type
(string) --
TotalLocalStorageGB (dict) --
The minimum and maximum total local storage size for an instance type, in GB.
Default: No minimum or maximum limits
Min (float) --
The storage minimum in GB.
Max (float) --
The storage maximum in GB.
BaselineEbsBandwidthMbps (dict) --
The minimum and maximum baseline bandwidth performance for an instance type, in Mbps. For more information, see Amazon EBS–optimized instances in the Amazon EC2 User Guide for Linux Instances .
Default: No minimum or maximum limits
Min (integer) --
The minimum value in Mbps.
Max (integer) --
The maximum value in Mbps.
AcceleratorTypes (list) --
Lists the accelerator types that must be on an instance type.
For instance types with GPU accelerators, specify gpu .
For instance types with FPGA accelerators, specify fpga .
For instance types with inference accelerators, specify inference .
Default: Any accelerator type
(string) --
AcceleratorCount (dict) --
The minimum and maximum number of accelerators (GPUs, FPGAs, or Amazon Web Services Inferentia chips) for an instance type.
To exclude accelerator-enabled instance types, set Max to 0 .
Default: No minimum or maximum limits
Min (integer) --
The minimum value.
Max (integer) --
The maximum value.
AcceleratorManufacturers (list) --
Indicates whether instance types must have accelerators by specific manufacturers.
For instance types with NVIDIA devices, specify nvidia .
For instance types with AMD devices, specify amd .
For instance types with Amazon Web Services devices, specify amazon-web-services .
For instance types with Xilinx devices, specify xilinx .
Default: Any manufacturer
(string) --
AcceleratorNames (list) --
Lists the accelerators that must be on an instance type.
For instance types with NVIDIA A100 GPUs, specify a100 .
For instance types with NVIDIA V100 GPUs, specify v100 .
For instance types with NVIDIA K80 GPUs, specify k80 .
For instance types with NVIDIA T4 GPUs, specify t4 .
For instance types with NVIDIA M60 GPUs, specify m60 .
For instance types with AMD Radeon Pro V520 GPUs, specify radeon-pro-v520 .
For instance types with Xilinx VU9P FPGAs, specify vu9p .
Default: Any accelerator
(string) --
AcceleratorTotalMemoryMiB (dict) --
The minimum and maximum total memory size for the accelerators on an instance type, in MiB.
Default: No minimum or maximum limits
Min (integer) --
The memory minimum in MiB.
Max (integer) --
The memory maximum in MiB.
NetworkBandwidthGbps (dict) --
The minimum and maximum amount of network bandwidth, in gigabits per second (Gbps).
Default: No minimum or maximum limits
Min (float) --
The minimum amount of network bandwidth, in gigabits per second (Gbps).
Max (float) --
The maximum amount of network bandwidth, in gigabits per second (Gbps).
AllowedInstanceTypes (list) --
The instance types to apply your specified attributes against. All other instance types are ignored, even if they match your specified attributes.
You can use strings with one or more wild cards, represented by an asterisk (* ), to allow an instance type, size, or generation. The following are examples: m5.8xlarge , c5*.* , m5a.* , r* , *3* .
For example, if you specify c5* , Amazon EC2 Auto Scaling will allow the entire C5 instance family, which includes all C5a and C5n instance types. If you specify m5a.* , Amazon EC2 Auto Scaling will allow all the M5a instance types, but not the M5n instance types.
Note
If you specify AllowedInstanceTypes , you can't specify ExcludedInstanceTypes .
Default: All instance types
(string) --
InstancesDistribution (dict) --
The instances distribution.
OnDemandAllocationStrategy (string) --
The allocation strategy to apply to your On-Demand Instances when they are launched. Possible instance types are determined by the launch template overrides that you specify.
The following lists the valid values:
lowest-price
Uses price to determine which instance types are the highest priority, launching the lowest priced instance types within an Availability Zone first. This is the default value for Auto Scaling groups that specify InstanceRequirements .
prioritized
You set the order of instance types for the launch template overrides from highest to lowest priority (from first to last in the list). Amazon EC2 Auto Scaling launches your highest priority instance types first. If all your On-Demand capacity cannot be fulfilled using your highest priority instance type, then Amazon EC2 Auto Scaling launches the remaining capacity using the second priority instance type, and so on. This is the default value for Auto Scaling groups that don't specify InstanceRequirements and cannot be used for groups that do.
OnDemandBaseCapacity (integer) --
The minimum amount of the Auto Scaling group's capacity that must be fulfilled by On-Demand Instances. This base portion is launched first as your group scales.
This number has the same unit of measurement as the group's desired capacity. If you change the default unit of measurement (number of instances) by specifying weighted capacity values in your launch template overrides list, or by changing the default desired capacity type setting of the group, you must specify this number using the same unit of measurement.
Default: 0
OnDemandPercentageAboveBaseCapacity (integer) --
Controls the percentages of On-Demand Instances and Spot Instances for your additional capacity beyond OnDemandBaseCapacity . Expressed as a number (for example, 20 specifies 20% On-Demand Instances, 80% Spot Instances). If set to 100, only On-Demand Instances are used.
Default: 100
SpotAllocationStrategy (string) --
The allocation strategy to apply to your Spot Instances when they are launched. Possible instance types are determined by the launch template overrides that you specify.
The following lists the valid values:
capacity-optimized
Requests Spot Instances using pools that are optimally chosen based on the available Spot capacity. This strategy has the lowest risk of interruption. To give certain instance types a higher chance of launching first, use capacity-optimized-prioritized .
capacity-optimized-prioritized
You set the order of instance types for the launch template overrides from highest to lowest priority (from first to last in the list). Amazon EC2 Auto Scaling honors the instance type priorities on a best effort basis but optimizes for capacity first. Note that if the On-Demand allocation strategy is set to prioritized , the same priority is applied when fulfilling On-Demand capacity. This is not a valid value for Auto Scaling groups that specify InstanceRequirements .
lowest-price
Requests Spot Instances using the lowest priced pools within an Availability Zone, across the number of Spot pools that you specify for the SpotInstancePools property. To ensure that your desired capacity is met, you might receive Spot Instances from several pools. This is the default value, but it might lead to high interruption rates because this strategy only considers instance price and not available capacity.
price-capacity-optimized (recommended)
The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price.
SpotInstancePools (integer) --
The number of Spot Instance pools across which to allocate your Spot Instances. The Spot pools are determined from the different instance types in the overrides. Valid only when the SpotAllocationStrategy is lowest-price . Value must be in the range of 1–20.
Default: 2
SpotMaxPrice (string) --
The maximum price per unit hour that you are willing to pay for a Spot Instance. If your maximum price is lower than the Spot price for the instance types that you selected, your Spot Instances are not launched. We do not recommend specifying a maximum price because it can lead to increased interruptions. When Spot Instances launch, you pay the current Spot price. To remove a maximum price that you previously set, include the property but specify an empty string ("") for the value.
Warning
If you specify a maximum price, your instances will be interrupted more frequently than if you do not specify one.
Valid Range: Minimum value of 0.001
string
The ID of the instance used to base the launch configuration on. If specified, Amazon EC2 Auto Scaling uses the configuration values from the specified instance to create a new launch configuration. To get the instance ID, use the Amazon EC2 DescribeInstances API operation. For more information, see Creating an Auto Scaling group using an EC2 instance in the Amazon EC2 Auto Scaling User Guide .
integer
[REQUIRED]
The minimum size of the group.
integer
[REQUIRED]
The maximum size of the group.
Note
With a mixed instances policy that uses instance weighting, Amazon EC2 Auto Scaling may need to go above MaxSize to meet your capacity requirements. In this event, Amazon EC2 Auto Scaling will never go above MaxSize by more than your largest instance weight (weights that define how many units each instance contributes to the desired capacity of the group).
integer
The desired capacity is the initial capacity of the Auto Scaling group at the time of its creation and the capacity it attempts to maintain. It can scale beyond this capacity if you configure auto scaling. This number must be greater than or equal to the minimum size of the group and less than or equal to the maximum size of the group. If you do not specify a desired capacity, the default is the minimum size of the group.
integer
Only needed if you use simple scaling policies.
The amount of time, in seconds, between one scaling activity ending and another one starting due to simple scaling policies. For more information, see Scaling cooldowns for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide .
Default: 300 seconds
list
A list of Availability Zones where instances in the Auto Scaling group can be created. Used for launching into the default VPC subnet in each Availability Zone when not using the VPCZoneIdentifier property, or for attaching a network interface when an existing network interface ID is specified in a launch template.
(string) --
list
A list of Classic Load Balancers associated with this Auto Scaling group. For Application Load Balancers, Network Load Balancers, and Gateway Load Balancer, specify the TargetGroupARNs property instead.
(string) --
list
The Amazon Resource Names (ARN) of the Elastic Load Balancing target groups to associate with the Auto Scaling group. Instances are registered as targets with the target groups. The target groups receive incoming traffic and route requests to one or more registered targets. For more information, see Use Elastic Load Balancing to distribute traffic across the instances in your Auto Scaling group in the Amazon EC2 Auto Scaling User Guide .
(string) --
string
Determines whether any additional health checks are performed on the instances in this group. Amazon EC2 health checks are always on. For more information, see Health checks for Auto Scaling instances in the Amazon EC2 Auto Scaling User Guide .
The valid values are EC2 (default), ELB , and VPC_LATTICE . The VPC_LATTICE health check type is reserved for use with VPC Lattice, which is in preview release and is subject to change.
integer
The amount of time, in seconds, that Amazon EC2 Auto Scaling waits before checking the health status of an EC2 instance that has come into service and marking it unhealthy due to a failed health check. This is useful if your instances do not immediately pass their health checks after they enter the InService state. For more information, see Set the health check grace period for an Auto Scaling group in the Amazon EC2 Auto Scaling User Guide .
Default: 0 seconds
string
The name of the placement group into which to launch your instances. For more information, see Placement groups in the Amazon EC2 User Guide for Linux Instances .
Note
A cluster placement group is a logical grouping of instances within a single Availability Zone. You cannot specify multiple Availability Zones and a cluster placement group.
string
A comma-separated list of subnet IDs for a virtual private cloud (VPC) where instances in the Auto Scaling group can be created. If you specify VPCZoneIdentifier with AvailabilityZones , the subnets that you specify must reside in those Availability Zones.
list
A policy or a list of policies that are used to select the instance to terminate. These policies are executed in the order that you list them. For more information, see Work with Amazon EC2 Auto Scaling termination policies in the Amazon EC2 Auto Scaling User Guide .
Valid values: Default | AllocationStrategy | ClosestToNextInstanceHour | NewestInstance | OldestInstance | OldestLaunchConfiguration | OldestLaunchTemplate | arn:aws:lambda:region:account-id:function:my-function:my-alias
(string) --
boolean
Indicates whether newly launched instances are protected from termination by Amazon EC2 Auto Scaling when scaling in. For more information about preventing instances from terminating on scale in, see Using instance scale-in protection in the Amazon EC2 Auto Scaling User Guide .
boolean
Indicates whether Capacity Rebalancing is enabled. Otherwise, Capacity Rebalancing is disabled. When you turn on Capacity Rebalancing, Amazon EC2 Auto Scaling attempts to launch a Spot Instance whenever Amazon EC2 notifies that a Spot Instance is at an elevated risk of interruption. After launching a new instance, it then terminates an old instance. For more information, see Use Capacity Rebalancing to handle Amazon EC2 Spot Interruptions in the in the Amazon EC2 Auto Scaling User Guide .
list
One or more lifecycle hooks to add to the Auto Scaling group before instances are launched.
(dict) --
Describes information used to specify a lifecycle hook for an Auto Scaling group.
For more information, see Amazon EC2 Auto Scaling lifecycle hooks in the Amazon EC2 Auto Scaling User Guide .
LifecycleHookName (string) -- [REQUIRED]
The name of the lifecycle hook.
LifecycleTransition (string) -- [REQUIRED]
The lifecycle transition. For Auto Scaling groups, there are two major lifecycle transitions.
To create a lifecycle hook for scale-out events, specify autoscaling:EC2_INSTANCE_LAUNCHING .
To create a lifecycle hook for scale-in events, specify autoscaling:EC2_INSTANCE_TERMINATING .
NotificationMetadata (string) --
Additional information that you want to include any time Amazon EC2 Auto Scaling sends a message to the notification target.
HeartbeatTimeout (integer) --
The maximum time, in seconds, that can elapse before the lifecycle hook times out. The range is from 30 to 7200 seconds. The default value is 3600 seconds (1 hour).
DefaultResult (string) --
The action the Auto Scaling group takes when the lifecycle hook timeout elapses or if an unexpected failure occurs. The default value is ABANDON .
Valid values: CONTINUE | ABANDON
NotificationTargetARN (string) --
The Amazon Resource Name (ARN) of the notification target that Amazon EC2 Auto Scaling sends notifications to when an instance is in a wait state for the lifecycle hook. You can specify an Amazon SNS topic or an Amazon SQS queue.
RoleARN (string) --
The ARN of the IAM role that allows the Auto Scaling group to publish to the specified notification target. For information about creating this role, see Configure a notification target for a lifecycle hook in the Amazon EC2 Auto Scaling User Guide .
Valid only if the notification target is an Amazon SNS topic or an Amazon SQS queue.
list
One or more tags. You can tag your Auto Scaling group and propagate the tags to the Amazon EC2 instances it launches. Tags are not propagated to Amazon EBS volumes. To add tags to Amazon EBS volumes, specify the tags in a launch template but use caution. If the launch template specifies an instance tag with a key that is also specified for the Auto Scaling group, Amazon EC2 Auto Scaling overrides the value of that instance tag with the value specified by the Auto Scaling group. For more information, see Tag Auto Scaling groups and instances in the Amazon EC2 Auto Scaling User Guide .
(dict) --
Describes a tag for an Auto Scaling group.
ResourceId (string) --
The name of the Auto Scaling group.
ResourceType (string) --
The type of resource. The only supported value is auto-scaling-group .
Key (string) -- [REQUIRED]
The tag key.
Value (string) --
The tag value.
PropagateAtLaunch (boolean) --
Determines whether the tag is added to new instances as they are launched in the group.
string
The Amazon Resource Name (ARN) of the service-linked role that the Auto Scaling group uses to call other Amazon Web Services service on your behalf. By default, Amazon EC2 Auto Scaling uses a service-linked role named AWSServiceRoleForAutoScaling , which it creates if it does not exist. For more information, see Service-linked roles in the Amazon EC2 Auto Scaling User Guide .
integer
The maximum amount of time, in seconds, that an instance can be in service. The default is null. If specified, the value must be either 0 or a number equal to or greater than 86,400 seconds (1 day). For more information, see Replacing Auto Scaling instances based on maximum instance lifetime in the Amazon EC2 Auto Scaling User Guide .
string
Reserved.
string
The unit of measurement for the value specified for desired capacity. Amazon EC2 Auto Scaling supports DesiredCapacityType for attribute-based instance type selection only. For more information, see Creating an Auto Scaling group using attribute-based instance type selection in the Amazon EC2 Auto Scaling User Guide .
By default, Amazon EC2 Auto Scaling specifies units , which translates into number of instances.
Valid values: units | vcpu | memory-mib
integer
The amount of time, in seconds, until a newly launched instance can contribute to the Amazon CloudWatch metrics. This delay lets an instance finish initializing before Amazon EC2 Auto Scaling aggregates instance metrics, resulting in more reliable usage data. Set this value equal to the amount of time that it takes for resource consumption to become stable after an instance reaches the InService state. For more information, see Set the default instance warmup for an Auto Scaling group in the Amazon EC2 Auto Scaling User Guide .
Warning
To manage your warm-up settings at the group level, we recommend that you set the default instance warmup, even if its value is set to 0 seconds . This also optimizes the performance of scaling policies that scale continuously, such as target tracking and step scaling policies.
If you need to remove a value that you previously set, include the property but specify -1 for the value. However, we strongly recommend keeping the default instance warmup enabled by specifying a minimum value of 0 .
Default: None
list
Reserved for use with Amazon VPC Lattice, which is in preview release and is subject to change. Do not use this parameter for production workloads. It is also subject to change.
The unique identifiers of one or more traffic sources.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group. Amazon EC2 Auto Scaling registers the running instances with the attached target groups. The target groups receive incoming traffic and route requests to one or more registered targets.
(dict) --
Describes the identifier of a traffic source.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group.
Identifier (string) --
The unique identifier of the traffic source.
None
{'AutoScalingGroups': {'TrafficSources': [{'Identifier': 'string'}]}}
Gets information about the Auto Scaling groups in the account and Region.
If you specify Auto Scaling group names, the output includes information for only the specified Auto Scaling groups. If you specify filters, the output includes information for only those Auto Scaling groups that meet the filter criteria. If you do not specify group names or filters, the output includes information for all Auto Scaling groups.
This operation also returns information about instances in Auto Scaling groups. To retrieve information about the instances in a warm pool, you must call the DescribeWarmPool API.
See also: AWS API Documentation
Request Syntax
client.describe_auto_scaling_groups( AutoScalingGroupNames=[ 'string', ], NextToken='string', MaxRecords=123, Filters=[ { 'Name': 'string', 'Values': [ 'string', ] }, ] )
list
The names of the Auto Scaling groups. By default, you can only specify up to 50 names. You can optionally increase this limit using the MaxRecords property.
If you omit this property, all Auto Scaling groups are described.
(string) --
string
The token for the next set of items to return. (You received this token from a previous call.)
integer
The maximum number of items to return with this call. The default value is 50 and the maximum value is 100 .
list
One or more filters to limit the results based on specific tags.
(dict) --
Describes a filter that is used to return a more specific list of results from a describe operation.
If you specify multiple filters, the filters are automatically logically joined with an AND , and the request returns only the results that match all of the specified filters.
For more information, see Tag Auto Scaling groups and instances in the Amazon EC2 Auto Scaling User Guide .
Name (string) --
The name of the filter.
The valid values for Name depend on which API operation you're using with the filter ( DescribeAutoScalingGroups or DescribeTags ).
DescribeAutoScalingGroups
Valid values for Name include the following:
tag-key - Accepts tag keys. The results only include information about the Auto Scaling groups associated with these tag keys.
tag-value - Accepts tag values. The results only include information about the Auto Scaling groups associated with these tag values.
tag:<key> - Accepts the key/value combination of the tag. Use the tag key in the filter name and the tag value as the filter value. The results only include information about the Auto Scaling groups associated with the specified key/value combination.
DescribeTags
Valid values for Name include the following:
auto-scaling-group - Accepts the names of Auto Scaling groups. The results only include information about the tags associated with these Auto Scaling groups.
key - Accepts tag keys. The results only include information about the tags associated with these tag keys.
value - Accepts tag values. The results only include information about the tags associated with these tag values.
propagate-at-launch - Accepts a Boolean value, which specifies whether tags propagate to instances at launch. The results only include information about the tags associated with the specified Boolean value.
Values (list) --
One or more filter values. Filter values are case-sensitive.
If you specify multiple values for a filter, the values are automatically logically joined with an OR , and the request returns all results that match any of the specified values. For example, specify "tag:environment" for the filter name and "production,development" for the filter values to find Auto Scaling groups with the tag "environment=production" or "environment=development".
(string) --
dict
Response Syntax
{ 'AutoScalingGroups': [ { 'AutoScalingGroupName': 'string', 'AutoScalingGroupARN': 'string', 'LaunchConfigurationName': 'string', 'LaunchTemplate': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'MixedInstancesPolicy': { 'LaunchTemplate': { 'LaunchTemplateSpecification': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'Overrides': [ { 'InstanceType': 'string', 'WeightedCapacity': 'string', 'LaunchTemplateSpecification': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'InstanceRequirements': { 'VCpuCount': { 'Min': 123, 'Max': 123 }, 'MemoryMiB': { 'Min': 123, 'Max': 123 }, 'CpuManufacturers': [ 'intel'|'amd'|'amazon-web-services', ], 'MemoryGiBPerVCpu': { 'Min': 123.0, 'Max': 123.0 }, 'ExcludedInstanceTypes': [ 'string', ], 'InstanceGenerations': [ 'current'|'previous', ], 'SpotMaxPricePercentageOverLowestPrice': 123, 'OnDemandMaxPricePercentageOverLowestPrice': 123, 'BareMetal': 'included'|'excluded'|'required', 'BurstablePerformance': 'included'|'excluded'|'required', 'RequireHibernateSupport': True|False, 'NetworkInterfaceCount': { 'Min': 123, 'Max': 123 }, 'LocalStorage': 'included'|'excluded'|'required', 'LocalStorageTypes': [ 'hdd'|'ssd', ], 'TotalLocalStorageGB': { 'Min': 123.0, 'Max': 123.0 }, 'BaselineEbsBandwidthMbps': { 'Min': 123, 'Max': 123 }, 'AcceleratorTypes': [ 'gpu'|'fpga'|'inference', ], 'AcceleratorCount': { 'Min': 123, 'Max': 123 }, 'AcceleratorManufacturers': [ 'nvidia'|'amd'|'amazon-web-services'|'xilinx', ], 'AcceleratorNames': [ 'a100'|'v100'|'k80'|'t4'|'m60'|'radeon-pro-v520'|'vu9p', ], 'AcceleratorTotalMemoryMiB': { 'Min': 123, 'Max': 123 }, 'NetworkBandwidthGbps': { 'Min': 123.0, 'Max': 123.0 }, 'AllowedInstanceTypes': [ 'string', ] } }, ] }, 'InstancesDistribution': { 'OnDemandAllocationStrategy': 'string', 'OnDemandBaseCapacity': 123, 'OnDemandPercentageAboveBaseCapacity': 123, 'SpotAllocationStrategy': 'string', 'SpotInstancePools': 123, 'SpotMaxPrice': 'string' } }, 'MinSize': 123, 'MaxSize': 123, 'DesiredCapacity': 123, 'PredictedCapacity': 123, 'DefaultCooldown': 123, 'AvailabilityZones': [ 'string', ], 'LoadBalancerNames': [ 'string', ], 'TargetGroupARNs': [ 'string', ], 'HealthCheckType': 'string', 'HealthCheckGracePeriod': 123, 'Instances': [ { 'InstanceId': 'string', 'InstanceType': 'string', 'AvailabilityZone': 'string', 'LifecycleState': 'Pending'|'Pending:Wait'|'Pending:Proceed'|'Quarantined'|'InService'|'Terminating'|'Terminating:Wait'|'Terminating:Proceed'|'Terminated'|'Detaching'|'Detached'|'EnteringStandby'|'Standby'|'Warmed:Pending'|'Warmed:Pending:Wait'|'Warmed:Pending:Proceed'|'Warmed:Terminating'|'Warmed:Terminating:Wait'|'Warmed:Terminating:Proceed'|'Warmed:Terminated'|'Warmed:Stopped'|'Warmed:Running'|'Warmed:Hibernated', 'HealthStatus': 'string', 'LaunchConfigurationName': 'string', 'LaunchTemplate': { 'LaunchTemplateId': 'string', 'LaunchTemplateName': 'string', 'Version': 'string' }, 'ProtectedFromScaleIn': True|False, 'WeightedCapacity': 'string' }, ], 'CreatedTime': datetime(2015, 1, 1), 'SuspendedProcesses': [ { 'ProcessName': 'string', 'SuspensionReason': 'string' }, ], 'PlacementGroup': 'string', 'VPCZoneIdentifier': 'string', 'EnabledMetrics': [ { 'Metric': 'string', 'Granularity': 'string' }, ], 'Status': 'string', 'Tags': [ { 'ResourceId': 'string', 'ResourceType': 'string', 'Key': 'string', 'Value': 'string', 'PropagateAtLaunch': True|False }, ], 'TerminationPolicies': [ 'string', ], 'NewInstancesProtectedFromScaleIn': True|False, 'ServiceLinkedRoleARN': 'string', 'MaxInstanceLifetime': 123, 'CapacityRebalance': True|False, 'WarmPoolConfiguration': { 'MaxGroupPreparedCapacity': 123, 'MinSize': 123, 'PoolState': 'Stopped'|'Running'|'Hibernated', 'Status': 'PendingDelete', 'InstanceReusePolicy': { 'ReuseOnScaleIn': True|False } }, 'WarmPoolSize': 123, 'Context': 'string', 'DesiredCapacityType': 'string', 'DefaultInstanceWarmup': 123, 'TrafficSources': [ { 'Identifier': 'string' }, ] }, ], 'NextToken': 'string' }
Response Structure
(dict) --
AutoScalingGroups (list) --
The groups.
(dict) --
Describes an Auto Scaling group.
AutoScalingGroupName (string) --
The name of the Auto Scaling group.
AutoScalingGroupARN (string) --
The Amazon Resource Name (ARN) of the Auto Scaling group.
LaunchConfigurationName (string) --
The name of the associated launch configuration.
LaunchTemplate (dict) --
The launch template for the group.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
MixedInstancesPolicy (dict) --
The mixed instances policy for the group.
LaunchTemplate (dict) --
One or more launch templates and the instance types (overrides) that are used to launch EC2 instances to fulfill On-Demand and Spot capacities.
LaunchTemplateSpecification (dict) --
The launch template.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
Overrides (list) --
Any properties that you specify override the same properties in the launch template.
(dict) --
Use this structure to let Amazon EC2 Auto Scaling do the following when the Auto Scaling group has a mixed instances policy:
Override the instance type that is specified in the launch template.
Use multiple instance types.
Specify the instance types that you want, or define your instance requirements instead and let Amazon EC2 Auto Scaling provision the available instance types that meet your requirements. This can provide Amazon EC2 Auto Scaling with a larger selection of instance types to choose from when fulfilling Spot and On-Demand capacities. You can view which instance types are matched before you apply the instance requirements to your Auto Scaling group.
After you define your instance requirements, you don't have to keep updating these settings to get new EC2 instance types automatically. Amazon EC2 Auto Scaling uses the instance requirements of the Auto Scaling group to determine whether a new EC2 instance type can be used.
InstanceType (string) --
The instance type, such as m3.xlarge . You must specify an instance type that is supported in your requested Region and Availability Zones. For more information, see Instance types in the Amazon Elastic Compute Cloud User Guide .
You can specify up to 40 instance types per Auto Scaling group.
WeightedCapacity (string) --
If you provide a list of instance types to use, you can specify the number of capacity units provided by each instance type in terms of virtual CPUs, memory, storage, throughput, or other relative performance characteristic. When a Spot or On-Demand Instance is launched, the capacity units count toward the desired capacity. Amazon EC2 Auto Scaling launches instances until the desired capacity is totally fulfilled, even if this results in an overage. For example, if there are two units remaining to fulfill capacity, and Amazon EC2 Auto Scaling can only launch an instance with a WeightedCapacity of five units, the instance is launched, and the desired capacity is exceeded by three units. For more information, see Configuring instance weighting for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide . Value must be in the range of 1–999.
If you specify a value for WeightedCapacity for one instance type, you must specify a value for WeightedCapacity for all of them.
Warning
Every Auto Scaling group has three size parameters (DesiredCapacity , MaxSize , and MinSize ). Usually, you set these sizes based on a specific number of instances. However, if you configure a mixed instances policy that defines weights for the instance types, you must specify these sizes with the same units that you use for weighting instances.
LaunchTemplateSpecification (dict) --
Provides a launch template for the specified instance type or set of instance requirements. For example, some instance types might require a launch template with a different AMI. If not provided, Amazon EC2 Auto Scaling uses the launch template that's specified in the LaunchTemplate definition. For more information, see Specifying a different launch template for an instance type in the Amazon EC2 Auto Scaling User Guide .
You can specify up to 20 launch templates per Auto Scaling group. The launch templates specified in the overrides and in the LaunchTemplate definition count towards this limit.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
InstanceRequirements (dict) --
The instance requirements. Amazon EC2 Auto Scaling uses your specified requirements to identify instance types. Then, it uses your On-Demand and Spot allocation strategies to launch instances from these instance types.
You can specify up to four separate sets of instance requirements per Auto Scaling group. This is useful for provisioning instances from different Amazon Machine Images (AMIs) in the same Auto Scaling group. To do this, create the AMIs and create a new launch template for each AMI. Then, create a compatible set of instance requirements for each launch template.
Note
If you specify InstanceRequirements , you can't specify InstanceType .
VCpuCount (dict) --
The minimum and maximum number of vCPUs for an instance type.
Min (integer) --
The minimum number of vCPUs.
Max (integer) --
The maximum number of vCPUs.
MemoryMiB (dict) --
The minimum and maximum instance memory size for an instance type, in MiB.
Min (integer) --
The memory minimum in MiB.
Max (integer) --
The memory maximum in MiB.
CpuManufacturers (list) --
Lists which specific CPU manufacturers to include.
For instance types with Intel CPUs, specify intel .
For instance types with AMD CPUs, specify amd .
For instance types with Amazon Web Services CPUs, specify amazon-web-services .
Note
Don't confuse the CPU hardware manufacturer with the CPU hardware architecture. Instances will be launched with a compatible CPU architecture based on the Amazon Machine Image (AMI) that you specify in your launch template.
Default: Any manufacturer
(string) --
MemoryGiBPerVCpu (dict) --
The minimum and maximum amount of memory per vCPU for an instance type, in GiB.
Default: No minimum or maximum limits
Min (float) --
The memory minimum in GiB.
Max (float) --
The memory maximum in GiB.
ExcludedInstanceTypes (list) --
The instance types to exclude. You can use strings with one or more wild cards, represented by an asterisk (* ), to exclude an instance family, type, size, or generation. The following are examples: m5.8xlarge , c5*.* , m5a.* , r* , *3* .
For example, if you specify c5* , you are excluding the entire C5 instance family, which includes all C5a and C5n instance types. If you specify m5a.* , Amazon EC2 Auto Scaling will exclude all the M5a instance types, but not the M5n instance types.
Note
If you specify ExcludedInstanceTypes , you can't specify AllowedInstanceTypes .
Default: No excluded instance types
(string) --
InstanceGenerations (list) --
Indicates whether current or previous generation instance types are included.
For current generation instance types, specify current . The current generation includes EC2 instance types currently recommended for use. This typically includes the latest two to three generations in each instance family. For more information, see Instance types in the Amazon EC2 User Guide for Linux Instances .
For previous generation instance types, specify previous .
Default: Any current or previous generation
(string) --
SpotMaxPricePercentageOverLowestPrice (integer) --
The price protection threshold for Spot Instances. This is the maximum you’ll pay for a Spot Instance, expressed as a percentage higher than the least expensive current generation M, C, or R instance type with your specified attributes. When Amazon EC2 Auto Scaling selects instance types with your attributes, we will exclude instance types whose price is higher than your threshold. The parameter accepts an integer, which Amazon EC2 Auto Scaling interprets as a percentage. To turn off price protection, specify a high value, such as 999999 .
If you set DesiredCapacityType to vcpu or memory-mib , the price protection threshold is applied based on the per vCPU or per memory price instead of the per instance price.
Default: 100
OnDemandMaxPricePercentageOverLowestPrice (integer) --
The price protection threshold for On-Demand Instances. This is the maximum you’ll pay for an On-Demand Instance, expressed as a percentage higher than the least expensive current generation M, C, or R instance type with your specified attributes. When Amazon EC2 Auto Scaling selects instance types with your attributes, we will exclude instance types whose price is higher than your threshold. The parameter accepts an integer, which Amazon EC2 Auto Scaling interprets as a percentage. To turn off price protection, specify a high value, such as 999999 .
If you set DesiredCapacityType to vcpu or memory-mib , the price protection threshold is applied based on the per vCPU or per memory price instead of the per instance price.
Default: 20
BareMetal (string) --
Indicates whether bare metal instance types are included, excluded, or required.
Default: excluded
BurstablePerformance (string) --
Indicates whether burstable performance instance types are included, excluded, or required. For more information, see Burstable performance instances in the Amazon EC2 User Guide for Linux Instances .
Default: excluded
RequireHibernateSupport (boolean) --
Indicates whether instance types must provide On-Demand Instance hibernation support.
Default: false
NetworkInterfaceCount (dict) --
The minimum and maximum number of network interfaces for an instance type.
Default: No minimum or maximum limits
Min (integer) --
The minimum number of network interfaces.
Max (integer) --
The maximum number of network interfaces.
LocalStorage (string) --
Indicates whether instance types with instance store volumes are included, excluded, or required. For more information, see Amazon EC2 instance store in the Amazon EC2 User Guide for Linux Instances .
Default: included
LocalStorageTypes (list) --
Indicates the type of local storage that is required.
For instance types with hard disk drive (HDD) storage, specify hdd .
For instance types with solid state drive (SSD) storage, specify ssd .
Default: Any local storage type
(string) --
TotalLocalStorageGB (dict) --
The minimum and maximum total local storage size for an instance type, in GB.
Default: No minimum or maximum limits
Min (float) --
The storage minimum in GB.
Max (float) --
The storage maximum in GB.
BaselineEbsBandwidthMbps (dict) --
The minimum and maximum baseline bandwidth performance for an instance type, in Mbps. For more information, see Amazon EBS–optimized instances in the Amazon EC2 User Guide for Linux Instances .
Default: No minimum or maximum limits
Min (integer) --
The minimum value in Mbps.
Max (integer) --
The maximum value in Mbps.
AcceleratorTypes (list) --
Lists the accelerator types that must be on an instance type.
For instance types with GPU accelerators, specify gpu .
For instance types with FPGA accelerators, specify fpga .
For instance types with inference accelerators, specify inference .
Default: Any accelerator type
(string) --
AcceleratorCount (dict) --
The minimum and maximum number of accelerators (GPUs, FPGAs, or Amazon Web Services Inferentia chips) for an instance type.
To exclude accelerator-enabled instance types, set Max to 0 .
Default: No minimum or maximum limits
Min (integer) --
The minimum value.
Max (integer) --
The maximum value.
AcceleratorManufacturers (list) --
Indicates whether instance types must have accelerators by specific manufacturers.
For instance types with NVIDIA devices, specify nvidia .
For instance types with AMD devices, specify amd .
For instance types with Amazon Web Services devices, specify amazon-web-services .
For instance types with Xilinx devices, specify xilinx .
Default: Any manufacturer
(string) --
AcceleratorNames (list) --
Lists the accelerators that must be on an instance type.
For instance types with NVIDIA A100 GPUs, specify a100 .
For instance types with NVIDIA V100 GPUs, specify v100 .
For instance types with NVIDIA K80 GPUs, specify k80 .
For instance types with NVIDIA T4 GPUs, specify t4 .
For instance types with NVIDIA M60 GPUs, specify m60 .
For instance types with AMD Radeon Pro V520 GPUs, specify radeon-pro-v520 .
For instance types with Xilinx VU9P FPGAs, specify vu9p .
Default: Any accelerator
(string) --
AcceleratorTotalMemoryMiB (dict) --
The minimum and maximum total memory size for the accelerators on an instance type, in MiB.
Default: No minimum or maximum limits
Min (integer) --
The memory minimum in MiB.
Max (integer) --
The memory maximum in MiB.
NetworkBandwidthGbps (dict) --
The minimum and maximum amount of network bandwidth, in gigabits per second (Gbps).
Default: No minimum or maximum limits
Min (float) --
The minimum amount of network bandwidth, in gigabits per second (Gbps).
Max (float) --
The maximum amount of network bandwidth, in gigabits per second (Gbps).
AllowedInstanceTypes (list) --
The instance types to apply your specified attributes against. All other instance types are ignored, even if they match your specified attributes.
You can use strings with one or more wild cards, represented by an asterisk (* ), to allow an instance type, size, or generation. The following are examples: m5.8xlarge , c5*.* , m5a.* , r* , *3* .
For example, if you specify c5* , Amazon EC2 Auto Scaling will allow the entire C5 instance family, which includes all C5a and C5n instance types. If you specify m5a.* , Amazon EC2 Auto Scaling will allow all the M5a instance types, but not the M5n instance types.
Note
If you specify AllowedInstanceTypes , you can't specify ExcludedInstanceTypes .
Default: All instance types
(string) --
InstancesDistribution (dict) --
The instances distribution.
OnDemandAllocationStrategy (string) --
The allocation strategy to apply to your On-Demand Instances when they are launched. Possible instance types are determined by the launch template overrides that you specify.
The following lists the valid values:
lowest-price
Uses price to determine which instance types are the highest priority, launching the lowest priced instance types within an Availability Zone first. This is the default value for Auto Scaling groups that specify InstanceRequirements .
prioritized
You set the order of instance types for the launch template overrides from highest to lowest priority (from first to last in the list). Amazon EC2 Auto Scaling launches your highest priority instance types first. If all your On-Demand capacity cannot be fulfilled using your highest priority instance type, then Amazon EC2 Auto Scaling launches the remaining capacity using the second priority instance type, and so on. This is the default value for Auto Scaling groups that don't specify InstanceRequirements and cannot be used for groups that do.
OnDemandBaseCapacity (integer) --
The minimum amount of the Auto Scaling group's capacity that must be fulfilled by On-Demand Instances. This base portion is launched first as your group scales.
This number has the same unit of measurement as the group's desired capacity. If you change the default unit of measurement (number of instances) by specifying weighted capacity values in your launch template overrides list, or by changing the default desired capacity type setting of the group, you must specify this number using the same unit of measurement.
Default: 0
OnDemandPercentageAboveBaseCapacity (integer) --
Controls the percentages of On-Demand Instances and Spot Instances for your additional capacity beyond OnDemandBaseCapacity . Expressed as a number (for example, 20 specifies 20% On-Demand Instances, 80% Spot Instances). If set to 100, only On-Demand Instances are used.
Default: 100
SpotAllocationStrategy (string) --
The allocation strategy to apply to your Spot Instances when they are launched. Possible instance types are determined by the launch template overrides that you specify.
The following lists the valid values:
capacity-optimized
Requests Spot Instances using pools that are optimally chosen based on the available Spot capacity. This strategy has the lowest risk of interruption. To give certain instance types a higher chance of launching first, use capacity-optimized-prioritized .
capacity-optimized-prioritized
You set the order of instance types for the launch template overrides from highest to lowest priority (from first to last in the list). Amazon EC2 Auto Scaling honors the instance type priorities on a best effort basis but optimizes for capacity first. Note that if the On-Demand allocation strategy is set to prioritized , the same priority is applied when fulfilling On-Demand capacity. This is not a valid value for Auto Scaling groups that specify InstanceRequirements .
lowest-price
Requests Spot Instances using the lowest priced pools within an Availability Zone, across the number of Spot pools that you specify for the SpotInstancePools property. To ensure that your desired capacity is met, you might receive Spot Instances from several pools. This is the default value, but it might lead to high interruption rates because this strategy only considers instance price and not available capacity.
price-capacity-optimized (recommended)
The price and capacity optimized allocation strategy looks at both price and capacity to select the Spot Instance pools that are the least likely to be interrupted and have the lowest possible price.
SpotInstancePools (integer) --
The number of Spot Instance pools across which to allocate your Spot Instances. The Spot pools are determined from the different instance types in the overrides. Valid only when the SpotAllocationStrategy is lowest-price . Value must be in the range of 1–20.
Default: 2
SpotMaxPrice (string) --
The maximum price per unit hour that you are willing to pay for a Spot Instance. If your maximum price is lower than the Spot price for the instance types that you selected, your Spot Instances are not launched. We do not recommend specifying a maximum price because it can lead to increased interruptions. When Spot Instances launch, you pay the current Spot price. To remove a maximum price that you previously set, include the property but specify an empty string ("") for the value.
Warning
If you specify a maximum price, your instances will be interrupted more frequently than if you do not specify one.
Valid Range: Minimum value of 0.001
MinSize (integer) --
The minimum size of the group.
MaxSize (integer) --
The maximum size of the group.
DesiredCapacity (integer) --
The desired size of the group.
PredictedCapacity (integer) --
The predicted capacity of the group when it has a predictive scaling policy.
DefaultCooldown (integer) --
The duration of the default cooldown period, in seconds.
AvailabilityZones (list) --
One or more Availability Zones for the group.
(string) --
LoadBalancerNames (list) --
One or more load balancers associated with the group.
(string) --
TargetGroupARNs (list) --
The Amazon Resource Names (ARN) of the target groups for your load balancer.
(string) --
HealthCheckType (string) --
Determines whether any additional health checks are performed on the instances in this group. Amazon EC2 health checks are always on.
The valid values are EC2 (default), ELB , and VPC_LATTICE . The VPC_LATTICE health check type is reserved for use with VPC Lattice, which is in preview release and is subject to change.
HealthCheckGracePeriod (integer) --
The duration of the health check grace period, in seconds.
Instances (list) --
The EC2 instances associated with the group.
(dict) --
Describes an EC2 instance.
InstanceId (string) --
The ID of the instance.
InstanceType (string) --
The instance type of the EC2 instance.
AvailabilityZone (string) --
The Availability Zone in which the instance is running.
LifecycleState (string) --
A description of the current lifecycle state. The Quarantined state is not used. For information about lifecycle states, see Instance lifecycle in the Amazon EC2 Auto Scaling User Guide .
HealthStatus (string) --
The last reported health status of the instance. "Healthy" means that the instance is healthy and should remain in service. "Unhealthy" means that the instance is unhealthy and that Amazon EC2 Auto Scaling should terminate and replace it.
LaunchConfigurationName (string) --
The launch configuration associated with the instance.
LaunchTemplate (dict) --
The launch template for the instance.
LaunchTemplateId (string) --
The ID of the launch template. To get the template ID, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
LaunchTemplateName (string) --
The name of the launch template. To get the template name, use the Amazon EC2 DescribeLaunchTemplates API operation. New launch templates can be created using the Amazon EC2 CreateLaunchTemplate API.
Conditional: You must specify either a LaunchTemplateId or a LaunchTemplateName .
Version (string) --
The version number, $Latest , or $Default . To get the version number, use the Amazon EC2 DescribeLaunchTemplateVersions API operation. New launch template versions can be created using the Amazon EC2 CreateLaunchTemplateVersion API. If the value is $Latest , Amazon EC2 Auto Scaling selects the latest version of the launch template when launching instances. If the value is $Default , Amazon EC2 Auto Scaling selects the default version of the launch template when launching instances. The default value is $Default .
ProtectedFromScaleIn (boolean) --
Indicates whether the instance is protected from termination by Amazon EC2 Auto Scaling when scaling in.
WeightedCapacity (string) --
The number of capacity units contributed by the instance based on its instance type.
Valid Range: Minimum value of 1. Maximum value of 999.
CreatedTime (datetime) --
The date and time the group was created.
SuspendedProcesses (list) --
The suspended processes associated with the group.
(dict) --
Describes an auto scaling process that has been suspended.
For more information, see Scaling processes in the Amazon EC2 Auto Scaling User Guide .
ProcessName (string) --
The name of the suspended process.
SuspensionReason (string) --
The reason that the process was suspended.
PlacementGroup (string) --
The name of the placement group into which to launch your instances, if any.
VPCZoneIdentifier (string) --
One or more subnet IDs, if applicable, separated by commas.
EnabledMetrics (list) --
The metrics enabled for the group.
(dict) --
Describes an enabled Auto Scaling group metric.
Metric (string) --
One of the following metrics:
GroupMinSize
GroupMaxSize
GroupDesiredCapacity
GroupInServiceInstances
GroupPendingInstances
GroupStandbyInstances
GroupTerminatingInstances
GroupTotalInstances
GroupInServiceCapacity
GroupPendingCapacity
GroupStandbyCapacity
GroupTerminatingCapacity
GroupTotalCapacity
WarmPoolDesiredCapacity
WarmPoolWarmedCapacity
WarmPoolPendingCapacity
WarmPoolTerminatingCapacity
WarmPoolTotalCapacity
GroupAndWarmPoolDesiredCapacity
GroupAndWarmPoolTotalCapacity
For more information, see Auto Scaling group metrics in the Amazon EC2 Auto Scaling User Guide .
Granularity (string) --
The granularity of the metric. The only valid value is 1Minute .
Status (string) --
The current state of the group when the DeleteAutoScalingGroup operation is in progress.
Tags (list) --
The tags for the group.
(dict) --
Describes a tag for an Auto Scaling group.
ResourceId (string) --
The name of the group.
ResourceType (string) --
The type of resource. The only supported value is auto-scaling-group .
Key (string) --
The tag key.
Value (string) --
The tag value.
PropagateAtLaunch (boolean) --
Determines whether the tag is added to new instances as they are launched in the group.
TerminationPolicies (list) --
The termination policies for the group.
(string) --
NewInstancesProtectedFromScaleIn (boolean) --
Indicates whether newly launched instances are protected from termination by Amazon EC2 Auto Scaling when scaling in.
ServiceLinkedRoleARN (string) --
The Amazon Resource Name (ARN) of the service-linked role that the Auto Scaling group uses to call other Amazon Web Services on your behalf.
MaxInstanceLifetime (integer) --
The maximum amount of time, in seconds, that an instance can be in service.
Valid Range: Minimum value of 0.
CapacityRebalance (boolean) --
Indicates whether Capacity Rebalancing is enabled.
WarmPoolConfiguration (dict) --
The warm pool for the group.
MaxGroupPreparedCapacity (integer) --
The maximum number of instances that are allowed to be in the warm pool or in any state except Terminated for the Auto Scaling group.
MinSize (integer) --
The minimum number of instances to maintain in the warm pool.
PoolState (string) --
The instance state to transition to after the lifecycle actions are complete.
Status (string) --
The status of a warm pool that is marked for deletion.
InstanceReusePolicy (dict) --
The instance reuse policy.
ReuseOnScaleIn (boolean) --
Specifies whether instances in the Auto Scaling group can be returned to the warm pool on scale in.
WarmPoolSize (integer) --
The current size of the warm pool.
Context (string) --
Reserved.
DesiredCapacityType (string) --
The unit of measurement for the value specified for desired capacity. Amazon EC2 Auto Scaling supports DesiredCapacityType for attribute-based instance type selection only.
DefaultInstanceWarmup (integer) --
The duration of the default instance warmup, in seconds.
TrafficSources (list) --
The unique identifiers of the traffic sources.
(dict) --
Describes the identifier of a traffic source.
Currently, you must specify an Amazon Resource Name (ARN) for an existing VPC Lattice target group.
Identifier (string) --
The unique identifier of the traffic source.
NextToken (string) --
A string that indicates that the response contains more items than can be returned in a single response. To receive additional items, specify this string for the NextToken value when requesting the next set of items. This value is null when there are no more items to return.
{'ScalingPolicies': {'TargetTrackingConfiguration': {'CustomizedMetricSpecification': {'Metrics': [{'Expression': 'string', 'Id': 'string', 'Label': 'string', 'MetricStat': {'Metric': {'Dimensions': [{'Name': 'string', 'Value': 'string'}], 'MetricName': 'string', 'Namespace': 'string'}, 'Stat': 'string', 'Unit': 'string'}, 'ReturnData': 'boolean'}]}}}}
Gets information about the scaling policies in the account and Region.
See also: AWS API Documentation
Request Syntax
client.describe_policies( AutoScalingGroupName='string', PolicyNames=[ 'string', ], PolicyTypes=[ 'string', ], NextToken='string', MaxRecords=123 )
string
The name of the Auto Scaling group.
list
The names of one or more policies. If you omit this property, all policies are described. If a group name is provided, the results are limited to that group. If you specify an unknown policy name, it is ignored with no error.
Array Members: Maximum number of 50 items.
(string) --
list
One or more policy types. The valid values are SimpleScaling , StepScaling , TargetTrackingScaling , and PredictiveScaling .
(string) --
string
The token for the next set of items to return. (You received this token from a previous call.)
integer
The maximum number of items to be returned with each call. The default value is 50 and the maximum value is 100 .
dict
Response Syntax
{ 'ScalingPolicies': [ { 'AutoScalingGroupName': 'string', 'PolicyName': 'string', 'PolicyARN': 'string', 'PolicyType': 'string', 'AdjustmentType': 'string', 'MinAdjustmentStep': 123, 'MinAdjustmentMagnitude': 123, 'ScalingAdjustment': 123, 'Cooldown': 123, 'StepAdjustments': [ { 'MetricIntervalLowerBound': 123.0, 'MetricIntervalUpperBound': 123.0, 'ScalingAdjustment': 123 }, ], 'MetricAggregationType': 'string', 'EstimatedInstanceWarmup': 123, 'Alarms': [ { 'AlarmName': 'string', 'AlarmARN': 'string' }, ], 'TargetTrackingConfiguration': { 'PredefinedMetricSpecification': { 'PredefinedMetricType': 'ASGAverageCPUUtilization'|'ASGAverageNetworkIn'|'ASGAverageNetworkOut'|'ALBRequestCountPerTarget', 'ResourceLabel': 'string' }, 'CustomizedMetricSpecification': { 'MetricName': 'string', 'Namespace': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ], 'Statistic': 'Average'|'Minimum'|'Maximum'|'SampleCount'|'Sum', 'Unit': 'string', 'Metrics': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'TargetValue': 123.0, 'DisableScaleIn': True|False }, 'Enabled': True|False, 'PredictiveScalingConfiguration': { 'MetricSpecifications': [ { 'TargetValue': 123.0, 'PredefinedMetricPairSpecification': { 'PredefinedMetricType': 'ASGCPUUtilization'|'ASGNetworkIn'|'ASGNetworkOut'|'ALBRequestCount', 'ResourceLabel': 'string' }, 'PredefinedScalingMetricSpecification': { 'PredefinedMetricType': 'ASGAverageCPUUtilization'|'ASGAverageNetworkIn'|'ASGAverageNetworkOut'|'ALBRequestCountPerTarget', 'ResourceLabel': 'string' }, 'PredefinedLoadMetricSpecification': { 'PredefinedMetricType': 'ASGTotalCPUUtilization'|'ASGTotalNetworkIn'|'ASGTotalNetworkOut'|'ALBTargetGroupRequestCount', 'ResourceLabel': 'string' }, 'CustomizedScalingMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'CustomizedLoadMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'CustomizedCapacityMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] } }, ], 'Mode': 'ForecastAndScale'|'ForecastOnly', 'SchedulingBufferTime': 123, 'MaxCapacityBreachBehavior': 'HonorMaxCapacity'|'IncreaseMaxCapacity', 'MaxCapacityBuffer': 123 } }, ], 'NextToken': 'string' }
Response Structure
(dict) --
ScalingPolicies (list) --
The scaling policies.
(dict) --
Describes a scaling policy.
AutoScalingGroupName (string) --
The name of the Auto Scaling group.
PolicyName (string) --
The name of the scaling policy.
PolicyARN (string) --
The Amazon Resource Name (ARN) of the policy.
PolicyType (string) --
One of the following policy types:
TargetTrackingScaling
StepScaling
SimpleScaling (default)
PredictiveScaling
For more information, see Target tracking scaling policies and Step and simple scaling policies in the Amazon EC2 Auto Scaling User Guide .
AdjustmentType (string) --
Specifies how the scaling adjustment is interpreted (for example, an absolute number or a percentage). The valid values are ChangeInCapacity , ExactCapacity , and PercentChangeInCapacity .
MinAdjustmentStep (integer) --
Available for backward compatibility. Use MinAdjustmentMagnitude instead.
MinAdjustmentMagnitude (integer) --
The minimum value to scale by when the adjustment type is PercentChangeInCapacity .
ScalingAdjustment (integer) --
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity.
Cooldown (integer) --
The duration of the policy's cooldown period, in seconds.
StepAdjustments (list) --
A set of adjustments that enable you to scale based on the size of the alarm breach.
(dict) --
Describes information used to create a step adjustment for a step scaling policy.
For the following examples, suppose that you have an alarm with a breach threshold of 50:
To trigger the adjustment when the metric is greater than or equal to 50 and less than 60, specify a lower bound of 0 and an upper bound of 10.
To trigger the adjustment when the metric is greater than 40 and less than or equal to 50, specify a lower bound of -10 and an upper bound of 0.
There are a few rules for the step adjustments for your step policy:
The ranges of your step adjustments can't overlap or have a gap.
At most, one step adjustment can have a null lower bound. If one step adjustment has a negative lower bound, then there must be a step adjustment with a null lower bound.
At most, one step adjustment can have a null upper bound. If one step adjustment has a positive upper bound, then there must be a step adjustment with a null upper bound.
The upper and lower bound can't be null in the same step adjustment.
For more information, see Step adjustments in the Amazon EC2 Auto Scaling User Guide .
MetricIntervalLowerBound (float) --
The lower bound for the difference between the alarm threshold and the CloudWatch metric. If the metric value is above the breach threshold, the lower bound is inclusive (the metric must be greater than or equal to the threshold plus the lower bound). Otherwise, it is exclusive (the metric must be greater than the threshold plus the lower bound). A null value indicates negative infinity.
MetricIntervalUpperBound (float) --
The upper bound for the difference between the alarm threshold and the CloudWatch metric. If the metric value is above the breach threshold, the upper bound is exclusive (the metric must be less than the threshold plus the upper bound). Otherwise, it is inclusive (the metric must be less than or equal to the threshold plus the upper bound). A null value indicates positive infinity.
The upper bound must be greater than the lower bound.
ScalingAdjustment (integer) --
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity.
The amount by which to scale. The adjustment is based on the value that you specified in the AdjustmentType property (either an absolute number or a percentage). A positive value adds to the current capacity and a negative number subtracts from the current capacity.
MetricAggregationType (string) --
The aggregation type for the CloudWatch metrics. The valid values are Minimum , Maximum , and Average .
EstimatedInstanceWarmup (integer) --
The estimated time, in seconds, until a newly launched instance can contribute to the CloudWatch metrics.
Alarms (list) --
The CloudWatch alarms related to the policy.
(dict) --
Describes an alarm.
AlarmName (string) --
The name of the alarm.
AlarmARN (string) --
The Amazon Resource Name (ARN) of the alarm.
TargetTrackingConfiguration (dict) --
A target tracking scaling policy.
PredefinedMetricSpecification (dict) --
A predefined metric. You must specify either a predefined metric or a customized metric.
PredefinedMetricType (string) --
The metric type. The following predefined metrics are available:
ASGAverageCPUUtilization - Average CPU utilization of the Auto Scaling group.
ASGAverageNetworkIn - Average number of bytes received on all network interfaces by the Auto Scaling group.
ASGAverageNetworkOut - Average number of bytes sent out on all network interfaces by the Auto Scaling group.
ALBRequestCountPerTarget - Average Application Load Balancer request count per target for your Auto Scaling group.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
CustomizedMetricSpecification (dict) --
A customized metric. You must specify either a predefined metric or a customized metric.
MetricName (string) --
The name of the metric. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) --
The namespace of the metric.
Dimensions (list) --
The dimensions of the metric.
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) --
The name of the dimension.
Value (string) --
The value of the dimension.
Statistic (string) --
The statistic of the metric.
Unit (string) --
The unit of the metric. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Metrics (list) --
The metrics to include in the target tracking scaling policy, as a metric data query. This can include both raw metric and metric math expressions.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
Id (string) --
A short name that identifies the object's results in the response. This name must be unique among all TargetTrackingMetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each TargetTrackingMetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each TargetTrackingMetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) --
Represents a specific metric.
Namespace (string) --
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) --
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) --
The name of the dimension.
Value (string) --
The value of the dimension.
Stat (string) --
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for scaling is Average
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
TargetValue (float) --
The target value for the metric.
Note
Some metrics are based on a count instead of a percentage, such as the request count for an Application Load Balancer or the number of messages in an SQS queue. If the scaling policy specifies one of these metrics, specify the target utilization as the optimal average request or message count per instance during any one-minute interval.
DisableScaleIn (boolean) --
Indicates whether scaling in by the target tracking scaling policy is disabled. If scaling in is disabled, the target tracking scaling policy doesn't remove instances from the Auto Scaling group. Otherwise, the target tracking scaling policy can remove instances from the Auto Scaling group. The default is false .
Enabled (boolean) --
Indicates whether the policy is enabled (true ) or disabled (false ).
PredictiveScalingConfiguration (dict) --
A predictive scaling policy.
MetricSpecifications (list) --
This structure includes the metrics and target utilization to use for predictive scaling.
This is an array, but we currently only support a single metric specification. That is, you can specify a target value and a single metric pair, or a target value and one scaling metric and one load metric.
(dict) --
This structure specifies the metrics and target utilization settings for a predictive scaling policy.
You must specify either a metric pair, or a load metric and a scaling metric individually. Specifying a metric pair instead of individual metrics provides a simpler way to configure metrics for a scaling policy. You choose the metric pair, and the policy automatically knows the correct sum and average statistics to use for the load metric and the scaling metric.
Example
You create a predictive scaling policy and specify ALBRequestCount as the value for the metric pair and 1000.0 as the target value. For this type of metric, you must provide the metric dimension for the corresponding target group, so you also provide a resource label for the Application Load Balancer target group that is attached to your Auto Scaling group.
The number of requests the target group receives per minute provides the load metric, and the request count averaged between the members of the target group provides the scaling metric. In CloudWatch, this refers to the RequestCount and RequestCountPerTarget metrics, respectively.
For optimal use of predictive scaling, you adhere to the best practice of using a dynamic scaling policy to automatically scale between the minimum capacity and maximum capacity in response to real-time changes in resource utilization.
Amazon EC2 Auto Scaling consumes data points for the load metric over the last 14 days and creates an hourly load forecast for predictive scaling. (A minimum of 24 hours of data is required.)
After creating the load forecast, Amazon EC2 Auto Scaling determines when to reduce or increase the capacity of your Auto Scaling group in each hour of the forecast period so that the average number of requests received by each instance is as close to 1000 requests per minute as possible at all times.
For information about using custom metrics with predictive scaling, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
TargetValue (float) --
Specifies the target utilization.
Note
Some metrics are based on a count instead of a percentage, such as the request count for an Application Load Balancer or the number of messages in an SQS queue. If the scaling policy specifies one of these metrics, specify the target utilization as the optimal average request or message count per instance during any one-minute interval.
PredefinedMetricPairSpecification (dict) --
The predefined metric pair specification from which Amazon EC2 Auto Scaling determines the appropriate scaling metric and load metric to use.
PredefinedMetricType (string) --
Indicates which metrics to use. There are two different types of metrics for each metric type: one is a load metric and one is a scaling metric. For example, if the metric type is ASGCPUUtilization , the Auto Scaling group's total CPU metric is used as the load metric, and the average CPU metric is used for the scaling metric.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the total and average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
PredefinedScalingMetricSpecification (dict) --
The predefined scaling metric specification.
PredefinedMetricType (string) --
The metric type.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
PredefinedLoadMetricSpecification (dict) --
The predefined load metric specification.
PredefinedMetricType (string) --
The metric type.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
CustomizedScalingMetricSpecification (dict) --
The customized scaling metric specification.
MetricDataQueries (list) --
One or more metric data queries to provide the data points for a scaling metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) --
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) --
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) --
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) --
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) --
The name of the dimension.
Value (string) --
The value of the dimension.
Stat (string) --
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
CustomizedLoadMetricSpecification (dict) --
The customized load metric specification.
MetricDataQueries (list) --
One or more metric data queries to provide the data points for a load metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) --
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) --
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) --
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) --
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) --
The name of the dimension.
Value (string) --
The value of the dimension.
Stat (string) --
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
CustomizedCapacityMetricSpecification (dict) --
The customized capacity metric specification.
MetricDataQueries (list) --
One or more metric data queries to provide the data points for a capacity metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) --
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) --
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) --
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) --
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) --
The name of the dimension.
Value (string) --
The value of the dimension.
Stat (string) --
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
Mode (string) --
The predictive scaling mode. Defaults to ForecastOnly if not specified.
SchedulingBufferTime (integer) --
The amount of time, in seconds, by which the instance launch time can be advanced. For example, the forecast says to add capacity at 10:00 AM, and you choose to pre-launch instances by 5 minutes. In that case, the instances will be launched at 9:55 AM. The intention is to give resources time to be provisioned. It can take a few minutes to launch an EC2 instance. The actual amount of time required depends on several factors, such as the size of the instance and whether there are startup scripts to complete.
The value must be less than the forecast interval duration of 3600 seconds (60 minutes). Defaults to 300 seconds if not specified.
MaxCapacityBreachBehavior (string) --
Defines the behavior that should be applied if the forecast capacity approaches or exceeds the maximum capacity of the Auto Scaling group. Defaults to HonorMaxCapacity if not specified.
The following are possible values:
HonorMaxCapacity - Amazon EC2 Auto Scaling cannot scale out capacity higher than the maximum capacity. The maximum capacity is enforced as a hard limit.
IncreaseMaxCapacity - Amazon EC2 Auto Scaling can scale out capacity higher than the maximum capacity when the forecast capacity is close to or exceeds the maximum capacity. The upper limit is determined by the forecasted capacity and the value for MaxCapacityBuffer .
MaxCapacityBuffer (integer) --
The size of the capacity buffer to use when the forecast capacity is close to or exceeds the maximum capacity. The value is specified as a percentage relative to the forecast capacity. For example, if the buffer is 10, this means a 10 percent buffer, such that if the forecast capacity is 50, and the maximum capacity is 40, then the effective maximum capacity is 55.
If set to 0, Amazon EC2 Auto Scaling may scale capacity higher than the maximum capacity to equal but not exceed forecast capacity.
Required if the MaxCapacityBreachBehavior property is set to IncreaseMaxCapacity , and cannot be used otherwise.
NextToken (string) --
A string that indicates that the response contains more items than can be returned in a single response. To receive additional items, specify this string for the NextToken value when requesting the next set of items. This value is null when there are no more items to return.
{'TargetTrackingConfiguration': {'CustomizedMetricSpecification': {'Metrics': [{'Expression': 'string', 'Id': 'string', 'Label': 'string', 'MetricStat': {'Metric': {'Dimensions': [{'Name': 'string', 'Value': 'string'}], 'MetricName': 'string', 'Namespace': 'string'}, 'Stat': 'string', 'Unit': 'string'}, 'ReturnData': 'boolean'}]}}}
Creates or updates a scaling policy for an Auto Scaling group. Scaling policies are used to scale an Auto Scaling group based on configurable metrics. If no policies are defined, the dynamic scaling and predictive scaling features are not used.
For more information about using dynamic scaling, see Target tracking scaling policies and Step and simple scaling policies in the Amazon EC2 Auto Scaling User Guide .
For more information about using predictive scaling, see Predictive scaling for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide .
You can view the scaling policies for an Auto Scaling group using the DescribePolicies API call. If you are no longer using a scaling policy, you can delete it by calling the DeletePolicy API.
See also: AWS API Documentation
Request Syntax
client.put_scaling_policy( AutoScalingGroupName='string', PolicyName='string', PolicyType='string', AdjustmentType='string', MinAdjustmentStep=123, MinAdjustmentMagnitude=123, ScalingAdjustment=123, Cooldown=123, MetricAggregationType='string', StepAdjustments=[ { 'MetricIntervalLowerBound': 123.0, 'MetricIntervalUpperBound': 123.0, 'ScalingAdjustment': 123 }, ], EstimatedInstanceWarmup=123, TargetTrackingConfiguration={ 'PredefinedMetricSpecification': { 'PredefinedMetricType': 'ASGAverageCPUUtilization'|'ASGAverageNetworkIn'|'ASGAverageNetworkOut'|'ALBRequestCountPerTarget', 'ResourceLabel': 'string' }, 'CustomizedMetricSpecification': { 'MetricName': 'string', 'Namespace': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ], 'Statistic': 'Average'|'Minimum'|'Maximum'|'SampleCount'|'Sum', 'Unit': 'string', 'Metrics': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'TargetValue': 123.0, 'DisableScaleIn': True|False }, Enabled=True|False, PredictiveScalingConfiguration={ 'MetricSpecifications': [ { 'TargetValue': 123.0, 'PredefinedMetricPairSpecification': { 'PredefinedMetricType': 'ASGCPUUtilization'|'ASGNetworkIn'|'ASGNetworkOut'|'ALBRequestCount', 'ResourceLabel': 'string' }, 'PredefinedScalingMetricSpecification': { 'PredefinedMetricType': 'ASGAverageCPUUtilization'|'ASGAverageNetworkIn'|'ASGAverageNetworkOut'|'ALBRequestCountPerTarget', 'ResourceLabel': 'string' }, 'PredefinedLoadMetricSpecification': { 'PredefinedMetricType': 'ASGTotalCPUUtilization'|'ASGTotalNetworkIn'|'ASGTotalNetworkOut'|'ALBTargetGroupRequestCount', 'ResourceLabel': 'string' }, 'CustomizedScalingMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'CustomizedLoadMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] }, 'CustomizedCapacityMetricSpecification': { 'MetricDataQueries': [ { 'Id': 'string', 'Expression': 'string', 'MetricStat': { 'Metric': { 'Namespace': 'string', 'MetricName': 'string', 'Dimensions': [ { 'Name': 'string', 'Value': 'string' }, ] }, 'Stat': 'string', 'Unit': 'string' }, 'Label': 'string', 'ReturnData': True|False }, ] } }, ], 'Mode': 'ForecastAndScale'|'ForecastOnly', 'SchedulingBufferTime': 123, 'MaxCapacityBreachBehavior': 'HonorMaxCapacity'|'IncreaseMaxCapacity', 'MaxCapacityBuffer': 123 } )
string
[REQUIRED]
The name of the Auto Scaling group.
string
[REQUIRED]
The name of the policy.
string
One of the following policy types:
TargetTrackingScaling
StepScaling
SimpleScaling (default)
PredictiveScaling
string
Specifies how the scaling adjustment is interpreted (for example, an absolute number or a percentage). The valid values are ChangeInCapacity , ExactCapacity , and PercentChangeInCapacity .
Required if the policy type is StepScaling or SimpleScaling . For more information, see Scaling adjustment types in the Amazon EC2 Auto Scaling User Guide .
integer
Available for backward compatibility. Use MinAdjustmentMagnitude instead.
integer
The minimum value to scale by when the adjustment type is PercentChangeInCapacity . For example, suppose that you create a step scaling policy to scale out an Auto Scaling group by 25 percent and you specify a MinAdjustmentMagnitude of 2. If the group has 4 instances and the scaling policy is performed, 25 percent of 4 is 1. However, because you specified a MinAdjustmentMagnitude of 2, Amazon EC2 Auto Scaling scales out the group by 2 instances.
Valid only if the policy type is StepScaling or SimpleScaling . For more information, see Scaling adjustment types in the Amazon EC2 Auto Scaling User Guide .
Note
Some Auto Scaling groups use instance weights. In this case, set the MinAdjustmentMagnitude to a value that is at least as large as your largest instance weight.
integer
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity. For exact capacity, you must specify a positive value.
Required if the policy type is SimpleScaling . (Not used with any other policy type.)
integer
A cooldown period, in seconds, that applies to a specific simple scaling policy. When a cooldown period is specified here, it overrides the default cooldown.
Valid only if the policy type is SimpleScaling . For more information, see Scaling cooldowns for Amazon EC2 Auto Scaling in the Amazon EC2 Auto Scaling User Guide .
Default: None
string
The aggregation type for the CloudWatch metrics. The valid values are Minimum , Maximum , and Average . If the aggregation type is null, the value is treated as Average .
Valid only if the policy type is StepScaling .
list
A set of adjustments that enable you to scale based on the size of the alarm breach.
Required if the policy type is StepScaling . (Not used with any other policy type.)
(dict) --
Describes information used to create a step adjustment for a step scaling policy.
For the following examples, suppose that you have an alarm with a breach threshold of 50:
To trigger the adjustment when the metric is greater than or equal to 50 and less than 60, specify a lower bound of 0 and an upper bound of 10.
To trigger the adjustment when the metric is greater than 40 and less than or equal to 50, specify a lower bound of -10 and an upper bound of 0.
There are a few rules for the step adjustments for your step policy:
The ranges of your step adjustments can't overlap or have a gap.
At most, one step adjustment can have a null lower bound. If one step adjustment has a negative lower bound, then there must be a step adjustment with a null lower bound.
At most, one step adjustment can have a null upper bound. If one step adjustment has a positive upper bound, then there must be a step adjustment with a null upper bound.
The upper and lower bound can't be null in the same step adjustment.
For more information, see Step adjustments in the Amazon EC2 Auto Scaling User Guide .
MetricIntervalLowerBound (float) --
The lower bound for the difference between the alarm threshold and the CloudWatch metric. If the metric value is above the breach threshold, the lower bound is inclusive (the metric must be greater than or equal to the threshold plus the lower bound). Otherwise, it is exclusive (the metric must be greater than the threshold plus the lower bound). A null value indicates negative infinity.
MetricIntervalUpperBound (float) --
The upper bound for the difference between the alarm threshold and the CloudWatch metric. If the metric value is above the breach threshold, the upper bound is exclusive (the metric must be less than the threshold plus the upper bound). Otherwise, it is inclusive (the metric must be less than or equal to the threshold plus the upper bound). A null value indicates positive infinity.
The upper bound must be greater than the lower bound.
ScalingAdjustment (integer) -- [REQUIRED]
The amount by which to scale, based on the specified adjustment type. A positive value adds to the current capacity while a negative number removes from the current capacity.
The amount by which to scale. The adjustment is based on the value that you specified in the AdjustmentType property (either an absolute number or a percentage). A positive value adds to the current capacity and a negative number subtracts from the current capacity.
integer
Not needed if the default instance warmup is defined for the group.
The estimated time, in seconds, until a newly launched instance can contribute to the CloudWatch metrics. This warm-up period applies to instances launched due to a specific target tracking or step scaling policy. When a warm-up period is specified here, it overrides the default instance warmup.
Valid only if the policy type is TargetTrackingScaling or StepScaling .
Note
The default is to use the value for the default instance warmup defined for the group. If default instance warmup is null, then EstimatedInstanceWarmup falls back to the value of default cooldown.
dict
A target tracking scaling policy. Provides support for predefined or custom metrics.
The following predefined metrics are available:
ASGAverageCPUUtilization
ASGAverageNetworkIn
ASGAverageNetworkOut
ALBRequestCountPerTarget
If you specify ALBRequestCountPerTarget for the metric, you must specify the ResourceLabel property with the PredefinedMetricSpecification .
For more information, see TargetTrackingConfiguration in the Amazon EC2 Auto Scaling API Reference .
Required if the policy type is TargetTrackingScaling .
PredefinedMetricSpecification (dict) --
A predefined metric. You must specify either a predefined metric or a customized metric.
PredefinedMetricType (string) -- [REQUIRED]
The metric type. The following predefined metrics are available:
ASGAverageCPUUtilization - Average CPU utilization of the Auto Scaling group.
ASGAverageNetworkIn - Average number of bytes received on all network interfaces by the Auto Scaling group.
ASGAverageNetworkOut - Average number of bytes sent out on all network interfaces by the Auto Scaling group.
ALBRequestCountPerTarget - Average Application Load Balancer request count per target for your Auto Scaling group.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
CustomizedMetricSpecification (dict) --
A customized metric. You must specify either a predefined metric or a customized metric.
MetricName (string) --
The name of the metric. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) --
The namespace of the metric.
Dimensions (list) --
The dimensions of the metric.
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) -- [REQUIRED]
The name of the dimension.
Value (string) -- [REQUIRED]
The value of the dimension.
Statistic (string) --
The statistic of the metric.
Unit (string) --
The unit of the metric. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Metrics (list) --
The metrics to include in the target tracking scaling policy, as a metric data query. This can include both raw metric and metric math expressions.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
Id (string) -- [REQUIRED]
A short name that identifies the object's results in the response. This name must be unique among all TargetTrackingMetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each TargetTrackingMetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each TargetTrackingMetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) -- [REQUIRED]
Represents a specific metric.
Namespace (string) -- [REQUIRED]
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) -- [REQUIRED]
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) -- [REQUIRED]
The name of the dimension.
Value (string) -- [REQUIRED]
The value of the dimension.
Stat (string) -- [REQUIRED]
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for scaling is Average
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
TargetValue (float) -- [REQUIRED]
The target value for the metric.
Note
Some metrics are based on a count instead of a percentage, such as the request count for an Application Load Balancer or the number of messages in an SQS queue. If the scaling policy specifies one of these metrics, specify the target utilization as the optimal average request or message count per instance during any one-minute interval.
DisableScaleIn (boolean) --
Indicates whether scaling in by the target tracking scaling policy is disabled. If scaling in is disabled, the target tracking scaling policy doesn't remove instances from the Auto Scaling group. Otherwise, the target tracking scaling policy can remove instances from the Auto Scaling group. The default is false .
boolean
Indicates whether the scaling policy is enabled or disabled. The default is enabled. For more information, see Disabling a scaling policy for an Auto Scaling group in the Amazon EC2 Auto Scaling User Guide .
dict
A predictive scaling policy. Provides support for predefined and custom metrics.
Predefined metrics include CPU utilization, network in/out, and the Application Load Balancer request count.
For more information, see PredictiveScalingConfiguration in the Amazon EC2 Auto Scaling API Reference .
Required if the policy type is PredictiveScaling .
MetricSpecifications (list) -- [REQUIRED]
This structure includes the metrics and target utilization to use for predictive scaling.
This is an array, but we currently only support a single metric specification. That is, you can specify a target value and a single metric pair, or a target value and one scaling metric and one load metric.
(dict) --
This structure specifies the metrics and target utilization settings for a predictive scaling policy.
You must specify either a metric pair, or a load metric and a scaling metric individually. Specifying a metric pair instead of individual metrics provides a simpler way to configure metrics for a scaling policy. You choose the metric pair, and the policy automatically knows the correct sum and average statistics to use for the load metric and the scaling metric.
Example
You create a predictive scaling policy and specify ALBRequestCount as the value for the metric pair and 1000.0 as the target value. For this type of metric, you must provide the metric dimension for the corresponding target group, so you also provide a resource label for the Application Load Balancer target group that is attached to your Auto Scaling group.
The number of requests the target group receives per minute provides the load metric, and the request count averaged between the members of the target group provides the scaling metric. In CloudWatch, this refers to the RequestCount and RequestCountPerTarget metrics, respectively.
For optimal use of predictive scaling, you adhere to the best practice of using a dynamic scaling policy to automatically scale between the minimum capacity and maximum capacity in response to real-time changes in resource utilization.
Amazon EC2 Auto Scaling consumes data points for the load metric over the last 14 days and creates an hourly load forecast for predictive scaling. (A minimum of 24 hours of data is required.)
After creating the load forecast, Amazon EC2 Auto Scaling determines when to reduce or increase the capacity of your Auto Scaling group in each hour of the forecast period so that the average number of requests received by each instance is as close to 1000 requests per minute as possible at all times.
For information about using custom metrics with predictive scaling, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
TargetValue (float) -- [REQUIRED]
Specifies the target utilization.
Note
Some metrics are based on a count instead of a percentage, such as the request count for an Application Load Balancer or the number of messages in an SQS queue. If the scaling policy specifies one of these metrics, specify the target utilization as the optimal average request or message count per instance during any one-minute interval.
PredefinedMetricPairSpecification (dict) --
The predefined metric pair specification from which Amazon EC2 Auto Scaling determines the appropriate scaling metric and load metric to use.
PredefinedMetricType (string) -- [REQUIRED]
Indicates which metrics to use. There are two different types of metrics for each metric type: one is a load metric and one is a scaling metric. For example, if the metric type is ASGCPUUtilization , the Auto Scaling group's total CPU metric is used as the load metric, and the average CPU metric is used for the scaling metric.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the total and average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
PredefinedScalingMetricSpecification (dict) --
The predefined scaling metric specification.
PredefinedMetricType (string) -- [REQUIRED]
The metric type.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the average request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
PredefinedLoadMetricSpecification (dict) --
The predefined load metric specification.
PredefinedMetricType (string) -- [REQUIRED]
The metric type.
ResourceLabel (string) --
A label that uniquely identifies a specific Application Load Balancer target group from which to determine the request count served by your Auto Scaling group. You can't specify a resource label unless the target group is attached to the Auto Scaling group.
You create the resource label by appending the final portion of the load balancer ARN and the final portion of the target group ARN into a single value, separated by a forward slash (/). The format of the resource label is:
app/my-alb/778d41231b141a0f/targetgroup/my-alb-target-group/943f017f100becff .
Where:
app/<load-balancer-name>/<load-balancer-id> is the final portion of the load balancer ARN
targetgroup/<target-group-name>/<target-group-id> is the final portion of the target group ARN.
To find the ARN for an Application Load Balancer, use the DescribeLoadBalancers API operation. To find the ARN for the target group, use the DescribeTargetGroups API operation.
CustomizedScalingMetricSpecification (dict) --
The customized scaling metric specification.
MetricDataQueries (list) -- [REQUIRED]
One or more metric data queries to provide the data points for a scaling metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) -- [REQUIRED]
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) -- [REQUIRED]
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) -- [REQUIRED]
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) -- [REQUIRED]
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) -- [REQUIRED]
The name of the dimension.
Value (string) -- [REQUIRED]
The value of the dimension.
Stat (string) -- [REQUIRED]
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
CustomizedLoadMetricSpecification (dict) --
The customized load metric specification.
MetricDataQueries (list) -- [REQUIRED]
One or more metric data queries to provide the data points for a load metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) -- [REQUIRED]
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) -- [REQUIRED]
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) -- [REQUIRED]
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) -- [REQUIRED]
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) -- [REQUIRED]
The name of the dimension.
Value (string) -- [REQUIRED]
The value of the dimension.
Stat (string) -- [REQUIRED]
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
CustomizedCapacityMetricSpecification (dict) --
The customized capacity metric specification.
MetricDataQueries (list) -- [REQUIRED]
One or more metric data queries to provide the data points for a capacity metric. Use multiple metric data queries only if you are performing a math expression on returned data.
(dict) --
The metric data to return. Also defines whether this call is returning data for one metric only, or whether it is performing a math expression on the values of returned metric statistics to create a new time series. A time series is a series of data points, each of which is associated with a timestamp.
For more information and examples, see Advanced predictive scaling policy configurations using custom metrics in the Amazon EC2 Auto Scaling User Guide .
Id (string) -- [REQUIRED]
A short name that identifies the object's results in the response. This name must be unique among all MetricDataQuery objects specified for a single scaling policy. If you are performing math expressions on this set of data, this name represents that data and can serve as a variable in the mathematical expression. The valid characters are letters, numbers, and underscores. The first character must be a lowercase letter.
Expression (string) --
The math expression to perform on the returned data, if this object is performing a math expression. This expression can use the Id of the other metrics to refer to those metrics, and can also use the Id of other expressions to use the result of those expressions.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
MetricStat (dict) --
Information about the metric data to return.
Conditional: Within each MetricDataQuery object, you must specify either Expression or MetricStat , but not both.
Metric (dict) -- [REQUIRED]
The CloudWatch metric to return, including the metric name, namespace, and dimensions. To get the exact metric name, namespace, and dimensions, inspect the Metric object that is returned by a call to ListMetrics .
Namespace (string) -- [REQUIRED]
The namespace of the metric. For more information, see the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
MetricName (string) -- [REQUIRED]
The name of the metric.
Dimensions (list) --
The dimensions for the metric. For the list of available dimensions, see the Amazon Web Services documentation available from the table in Amazon Web Services services that publish CloudWatch metrics in the Amazon CloudWatch User Guide .
Conditional: If you published your metric with dimensions, you must specify the same dimensions in your scaling policy.
(dict) --
Describes the dimension of a metric.
Name (string) -- [REQUIRED]
The name of the dimension.
Value (string) -- [REQUIRED]
The value of the dimension.
Stat (string) -- [REQUIRED]
The statistic to return. It can include any CloudWatch statistic or extended statistic. For a list of valid values, see the table in Statistics in the Amazon CloudWatch User Guide .
The most commonly used metrics for predictive scaling are Average and Sum .
Unit (string) --
The unit to use for the returned data points. For a complete list of the units that CloudWatch supports, see the MetricDatum data type in the Amazon CloudWatch API Reference .
Label (string) --
A human-readable label for this metric or expression. This is especially useful if this is a math expression, so that you know what the value represents.
ReturnData (boolean) --
Indicates whether to return the timestamps and raw data values of this metric.
If you use any math expressions, specify true for this value for only the final math expression that the metric specification is based on. You must specify false for ReturnData for all the other metrics and expressions used in the metric specification.
If you are only retrieving metrics and not performing any math expressions, do not specify anything for ReturnData . This sets it to its default (true ).
Mode (string) --
The predictive scaling mode. Defaults to ForecastOnly if not specified.
SchedulingBufferTime (integer) --
The amount of time, in seconds, by which the instance launch time can be advanced. For example, the forecast says to add capacity at 10:00 AM, and you choose to pre-launch instances by 5 minutes. In that case, the instances will be launched at 9:55 AM. The intention is to give resources time to be provisioned. It can take a few minutes to launch an EC2 instance. The actual amount of time required depends on several factors, such as the size of the instance and whether there are startup scripts to complete.
The value must be less than the forecast interval duration of 3600 seconds (60 minutes). Defaults to 300 seconds if not specified.
MaxCapacityBreachBehavior (string) --
Defines the behavior that should be applied if the forecast capacity approaches or exceeds the maximum capacity of the Auto Scaling group. Defaults to HonorMaxCapacity if not specified.
The following are possible values:
HonorMaxCapacity - Amazon EC2 Auto Scaling cannot scale out capacity higher than the maximum capacity. The maximum capacity is enforced as a hard limit.
IncreaseMaxCapacity - Amazon EC2 Auto Scaling can scale out capacity higher than the maximum capacity when the forecast capacity is close to or exceeds the maximum capacity. The upper limit is determined by the forecasted capacity and the value for MaxCapacityBuffer .
MaxCapacityBuffer (integer) --
The size of the capacity buffer to use when the forecast capacity is close to or exceeds the maximum capacity. The value is specified as a percentage relative to the forecast capacity. For example, if the buffer is 10, this means a 10 percent buffer, such that if the forecast capacity is 50, and the maximum capacity is 40, then the effective maximum capacity is 55.
If set to 0, Amazon EC2 Auto Scaling may scale capacity higher than the maximum capacity to equal but not exceed forecast capacity.
Required if the MaxCapacityBreachBehavior property is set to IncreaseMaxCapacity , and cannot be used otherwise.
dict
Response Syntax
{ 'PolicyARN': 'string', 'Alarms': [ { 'AlarmName': 'string', 'AlarmARN': 'string' }, ] }
Response Structure
(dict) --
Contains the output of PutScalingPolicy.
PolicyARN (string) --
The Amazon Resource Name (ARN) of the policy.
Alarms (list) --
The CloudWatch alarms created for the target tracking scaling policy.
(dict) --
Describes an alarm.
AlarmName (string) --
The name of the alarm.
AlarmARN (string) --
The Amazon Resource Name (ARN) of the alarm.