Welcome
This is the Application Auto Scaling API Reference. With Application Auto Scaling, you can configure automatic scaling for the following resources:
-
Amazon AppStream 2.0 fleets
-
Amazon Aurora Replicas
-
Amazon Comprehend document classification and entity recognizer endpoints
-
Amazon DynamoDB tables and global secondary indexes throughput capacity
-
Amazon ECS services
-
Amazon ElastiCache for Redis clusters (replication groups)
-
Amazon EMR clusters
-
Amazon Keyspaces (for Apache Cassandra) tables
-
AWS Lambda function provisioned concurrency
-
Amazon Managed Streaming for Apache Kafka broker storage
-
Amazon Neptune clusters
-
Amazon SageMaker endpoint variants
-
Amazon SageMaker inference components
-
Amazon SageMaker serverless endpoint provisioned concurrency
-
Spot Fleets (Amazon EC2)
-
Pool of WorkSpaces
-
Custom resources provided by your own applications or services
To learn more about Application Auto Scaling, see the Application Auto Scaling User Guide.
API Summary
The Application Auto Scaling service API includes three key sets of actions:
-
Register and manage scalable targets - Register AWS or custom resources as scalable targets (a resource that Application Auto Scaling can scale), set minimum and maximum capacity limits, and retrieve information on existing scalable targets.
-
Configure and manage automatic scaling - Define scaling policies to dynamically scale your resources in response to CloudWatch alarms, schedule one-time or recurring scaling actions, and retrieve your recent scaling activity history.
-
Suspend and resume scaling - Temporarily suspend and later resume automatic scaling by calling the RegisterScalableTarget API action for any Application Auto Scaling scalable target. You can suspend and resume (individually or in combination) scale-out activities that are triggered by a scaling policy, scale-in activities that are triggered by a scaling policy, and scheduled scaling.
The documentation for each action shows the request syntax, the
request parameters, and the response elements and provides links to language-specific SDK
reference topics. For more information, see AWS
SDKs
API request rate
Application Auto Scaling uses the token bucket algorithm to implement API throttling. With this
algorithm, your account has a bucket that holds a specific number of tokens. The number of
tokens in the bucket represents your throttling limit at any given second. Application Auto Scaling throttles
API requests based on a shared API bucket. For example, calls to the DescribeScalableTargets and DescribeScheduledActions API
operations use tokens from the same bucket. Throttling means that Application Auto Scaling rejects a request
because the request exceeds the service's limit for the number of requests per second. When a
request is throttled, Application Auto Scaling returns a RateExceeded
error. For more
information, see My Auto Scaling API
calls are getting throttled. What can I do to avoid this?
This document was last published on October 8, 2024.