ManagedKafkaEventSourceProps
- class aws_cdk.aws_lambda_event_sources.ManagedKafkaEventSourceProps(*, starting_position, batch_size=None, enabled=None, max_batching_window=None, provisioned_poller_config=None, topic, consumer_group_id=None, filter_encryption=None, filters=None, on_failure=None, secret=None, cluster_arn)
Bases:
KafkaEventSourceProps
Properties for a MSK event source.
- Parameters:
starting_position (
StartingPosition
) – Where to begin consuming the stream.batch_size (
Union
[int
,float
,None
]) – The largest number of records that AWS Lambda will retrieve from your event source at the time of invoking your function. Your function receives an event with all the retrieved records. Valid Range: - Minimum value of 1 - Maximum value of: - 1000 forDynamoEventSource
- 10000 forKinesisEventSource
,ManagedKafkaEventSource
andSelfManagedKafkaEventSource
Default: 100enabled (
Optional
[bool
]) – If the stream event source mapping should be enabled. Default: truemax_batching_window (
Optional
[Duration
]) – The maximum amount of time to gather records before invoking the function. Maximum of Duration.minutes(5). Default: - Duration.seconds(0) for Kinesis, DynamoDB, and SQS event sources, Duration.millis(500) for MSK, self-managed Kafka, and Amazon MQ.provisioned_poller_config (
Union
[ProvisionedPollerConfig
,Dict
[str
,Any
],None
]) – Configuration for provisioned pollers that read from the event source. When specified, allows control over the minimum and maximum number of pollers that can be provisioned to process events from the source. Default: - no provisioned pollerstopic (
str
) – The Kafka topic to subscribe to.consumer_group_id (
Optional
[str
]) – The identifier for the Kafka consumer group to join. The consumer group ID must be unique among all your Kafka event sources. After creating a Kafka event source mapping with the consumer group ID specified, you cannot update this value. The value must have a lenght between 1 and 200 and full the pattern ‘[a-zA-Z0-9-/:_+=.@-]’. Default: - nonefilter_encryption (
Optional
[IKey
]) – Add Customer managed KMS key to encrypt Filter Criteria. Default: - nonefilters (
Optional
[Sequence
[Mapping
[str
,Any
]]]) – Add filter criteria to Event Source. Default: - noneon_failure (
Optional
[IEventSourceDlq
]) – Add an on Failure Destination for this Kafka event. SNS/SQS/S3 are supported Default: - discarded records are ignoredsecret (
Optional
[ISecret
]) – The secret with the Kafka credentials, see https://docs.aws.amazon.com/msk/latest/developerguide/msk-password.html for details This field is required if your Kafka brokers are accessed over the Internet. Default: nonecluster_arn (
str
) – An MSK cluster construct.
- ExampleMetadata:
infused
Example:
from aws_cdk.aws_secretsmanager import Secret from aws_cdk.aws_lambda_event_sources import ManagedKafkaEventSource # my_function: lambda.Function # Your MSK cluster arn cluster_arn = "arn:aws:kafka:us-east-1:0123456789019:cluster/SalesCluster/abcd1234-abcd-cafe-abab-9876543210ab-4" # The Kafka topic you want to subscribe to topic = "some-cool-topic" # The secret that allows access to your MSK cluster # You still have to make sure that it is associated with your cluster as described in the documentation secret = Secret(self, "Secret", secret_name="AmazonMSK_KafkaSecret") my_function.add_event_source(ManagedKafkaEventSource( cluster_arn=cluster_arn, topic=topic, secret=secret, batch_size=100, # default starting_position=lambda_.StartingPosition.TRIM_HORIZON ))
Attributes
- batch_size
The largest number of records that AWS Lambda will retrieve from your event source at the time of invoking your function.
Your function receives an event with all the retrieved records.
Valid Range:
Minimum value of 1
Maximum value of:
1000 for
DynamoEventSource
10000 for
KinesisEventSource
,ManagedKafkaEventSource
andSelfManagedKafkaEventSource
- Default:
100
- cluster_arn
An MSK cluster construct.
- consumer_group_id
The identifier for the Kafka consumer group to join.
The consumer group ID must be unique among all your Kafka event sources. After creating a Kafka event source mapping with the consumer group ID specified, you cannot update this value. The value must have a lenght between 1 and 200 and full the pattern ‘[a-zA-Z0-9-/:_+=.@-]’.
- enabled
If the stream event source mapping should be enabled.
- Default:
true
- filter_encryption
Add Customer managed KMS key to encrypt Filter Criteria.
- filters
Add filter criteria to Event Source.
- Default:
none
- See:
https://docs.aws.amazon.com/lambda/latest/dg/invocation-eventfiltering.html
- max_batching_window
The maximum amount of time to gather records before invoking the function.
Maximum of Duration.minutes(5).
- Default:
Duration.seconds(0) for Kinesis, DynamoDB, and SQS event sources, Duration.millis(500) for MSK, self-managed Kafka, and Amazon MQ.
- See:
- on_failure
Add an on Failure Destination for this Kafka event.
SNS/SQS/S3 are supported
- Default:
discarded records are ignored
- provisioned_poller_config
Configuration for provisioned pollers that read from the event source.
When specified, allows control over the minimum and maximum number of pollers that can be provisioned to process events from the source.
- Default:
no provisioned pollers
- secret
//docs.aws.amazon.com/msk/latest/developerguide/msk-password.html for details This field is required if your Kafka brokers are accessed over the Internet.
- Default:
none
- Type:
The secret with the Kafka credentials, see https
- starting_position
Where to begin consuming the stream.
- topic
The Kafka topic to subscribe to.