CfnConnector
- class aws_cdk.aws_kafkaconnect.CfnConnector(scope, id, *, capacity, connector_configuration, connector_name, kafka_cluster, kafka_cluster_client_authentication, kafka_cluster_encryption_in_transit, kafka_connect_version, plugins, service_execution_role_arn, connector_description=None, log_delivery=None, worker_configuration=None)
Bases:
CfnResource
A CloudFormation
AWS::KafkaConnect::Connector
.Creates a connector using the specified properties.
- CloudformationResource:
AWS::KafkaConnect::Connector
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect cfn_connector = kafkaconnect.CfnConnector(self, "MyCfnConnector", capacity=kafkaconnect.CfnConnector.CapacityProperty( auto_scaling=kafkaconnect.CfnConnector.AutoScalingProperty( max_worker_count=123, mcu_count=123, min_worker_count=123, scale_in_policy=kafkaconnect.CfnConnector.ScaleInPolicyProperty( cpu_utilization_percentage=123 ), scale_out_policy=kafkaconnect.CfnConnector.ScaleOutPolicyProperty( cpu_utilization_percentage=123 ) ), provisioned_capacity=kafkaconnect.CfnConnector.ProvisionedCapacityProperty( worker_count=123, # the properties below are optional mcu_count=123 ) ), connector_configuration={ "connector_configuration_key": "connectorConfiguration" }, connector_name="connectorName", kafka_cluster=kafkaconnect.CfnConnector.KafkaClusterProperty( apache_kafka_cluster=kafkaconnect.CfnConnector.ApacheKafkaClusterProperty( bootstrap_servers="bootstrapServers", vpc=kafkaconnect.CfnConnector.VpcProperty( security_groups=["securityGroups"], subnets=["subnets"] ) ) ), kafka_cluster_client_authentication=kafkaconnect.CfnConnector.KafkaClusterClientAuthenticationProperty( authentication_type="authenticationType" ), kafka_cluster_encryption_in_transit=kafkaconnect.CfnConnector.KafkaClusterEncryptionInTransitProperty( encryption_type="encryptionType" ), kafka_connect_version="kafkaConnectVersion", plugins=[kafkaconnect.CfnConnector.PluginProperty( custom_plugin=kafkaconnect.CfnConnector.CustomPluginProperty( custom_plugin_arn="customPluginArn", revision=123 ) )], service_execution_role_arn="serviceExecutionRoleArn", # the properties below are optional connector_description="connectorDescription", log_delivery=kafkaconnect.CfnConnector.LogDeliveryProperty( worker_log_delivery=kafkaconnect.CfnConnector.WorkerLogDeliveryProperty( cloud_watch_logs=kafkaconnect.CfnConnector.CloudWatchLogsLogDeliveryProperty( enabled=False, # the properties below are optional log_group="logGroup" ), firehose=kafkaconnect.CfnConnector.FirehoseLogDeliveryProperty( enabled=False, # the properties below are optional delivery_stream="deliveryStream" ), s3=kafkaconnect.CfnConnector.S3LogDeliveryProperty( enabled=False, # the properties below are optional bucket="bucket", prefix="prefix" ) ) ), worker_configuration=kafkaconnect.CfnConnector.WorkerConfigurationProperty( revision=123, worker_configuration_arn="workerConfigurationArn" ) )
Create a new
AWS::KafkaConnect::Connector
.- Parameters:
scope (
Construct
) –scope in which this resource is defined.
id (
str
) –scoped id of the resource.
capacity (
Union
[CapacityProperty
,Dict
[str
,Any
],IResolvable
]) – The connector’s compute capacity settings.connector_configuration (
Union
[IResolvable
,Mapping
[str
,str
]]) – The configuration of the connector.connector_name (
str
) – The name of the connector.kafka_cluster (
Union
[IResolvable
,KafkaClusterProperty
,Dict
[str
,Any
]]) – The details of the Apache Kafka cluster to which the connector is connected.kafka_cluster_client_authentication (
Union
[IResolvable
,KafkaClusterClientAuthenticationProperty
,Dict
[str
,Any
]]) – The type of client authentication used to connect to the Apache Kafka cluster. The value is NONE when no client authentication is used.kafka_cluster_encryption_in_transit (
Union
[IResolvable
,KafkaClusterEncryptionInTransitProperty
,Dict
[str
,Any
]]) – Details of encryption in transit to the Apache Kafka cluster.kafka_connect_version (
str
) – The version of Kafka Connect. It has to be compatible with both the Apache Kafka cluster’s version and the plugins.plugins (
Union
[IResolvable
,Sequence
[Union
[IResolvable
,PluginProperty
,Dict
[str
,Any
]]]]) – Specifies which plugin to use for the connector. You must specify a single-element list. Amazon MSK Connect does not currently support specifying multiple plugins.service_execution_role_arn (
str
) – The Amazon Resource Name (ARN) of the IAM role used by the connector to access Amazon Web Services resources.connector_description (
Optional
[str
]) – The description of the connector.log_delivery (
Union
[IResolvable
,LogDeliveryProperty
,Dict
[str
,Any
],None
]) – The settings for delivering connector logs to Amazon CloudWatch Logs.worker_configuration (
Union
[IResolvable
,WorkerConfigurationProperty
,Dict
[str
,Any
],None
]) – The worker configurations that are in use with the connector.
Methods
- add_deletion_override(path)
Syntactic sugar for
addOverride(path, undefined)
.- Parameters:
path (
str
) – The path of the value to delete.- Return type:
None
- add_depends_on(target)
Indicates that this resource depends on another resource and cannot be provisioned unless the other resource has been successfully provisioned.
This can be used for resources across stacks (or nested stack) boundaries and the dependency will automatically be transferred to the relevant scope.
- Parameters:
target (
CfnResource
) –- Return type:
None
- add_metadata(key, value)
Add a value to the CloudFormation Resource Metadata.
- Parameters:
key (
str
) –value (
Any
) –
- See:
- Return type:
None
https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/metadata-section-structure.html
Note that this is a different set of metadata from CDK node metadata; this metadata ends up in the stack template under the resource, whereas CDK node metadata ends up in the Cloud Assembly.
- add_override(path, value)
Adds an override to the synthesized CloudFormation resource.
To add a property override, either use
addPropertyOverride
or prefixpath
with “Properties.” (i.e.Properties.TopicName
).If the override is nested, separate each nested level using a dot (.) in the path parameter. If there is an array as part of the nesting, specify the index in the path.
To include a literal
.
in the property name, prefix with a\
. In most programming languages you will need to write this as"\\."
because the\
itself will need to be escaped.For example:
cfn_resource.add_override("Properties.GlobalSecondaryIndexes.0.Projection.NonKeyAttributes", ["myattribute"]) cfn_resource.add_override("Properties.GlobalSecondaryIndexes.1.ProjectionType", "INCLUDE")
would add the overrides Example:
"Properties": { "GlobalSecondaryIndexes": [ { "Projection": { "NonKeyAttributes": [ "myattribute" ] ... } ... }, { "ProjectionType": "INCLUDE" ... }, ] ... }
The
value
argument toaddOverride
will not be processed or translated in any way. Pass raw JSON values in here with the correct capitalization for CloudFormation. If you pass CDK classes or structs, they will be rendered with lowercased key names, and CloudFormation will reject the template.- Parameters:
path (
str
) –The path of the property, you can use dot notation to override values in complex types. Any intermdediate keys will be created as needed.
value (
Any
) –The value. Could be primitive or complex.
- Return type:
None
- add_property_deletion_override(property_path)
Adds an override that deletes the value of a property from the resource definition.
- Parameters:
property_path (
str
) – The path to the property.- Return type:
None
- add_property_override(property_path, value)
Adds an override to a resource property.
Syntactic sugar for
addOverride("Properties.<...>", value)
.- Parameters:
property_path (
str
) – The path of the property.value (
Any
) – The value.
- Return type:
None
- apply_removal_policy(policy=None, *, apply_to_update_replace_policy=None, default=None)
Sets the deletion policy of the resource based on the removal policy specified.
The Removal Policy controls what happens to this resource when it stops being managed by CloudFormation, either because you’ve removed it from the CDK application or because you’ve made a change that requires the resource to be replaced.
The resource can be deleted (
RemovalPolicy.DESTROY
), or left in your AWS account for data recovery and cleanup later (RemovalPolicy.RETAIN
).- Parameters:
policy (
Optional
[RemovalPolicy
]) –apply_to_update_replace_policy (
Optional
[bool
]) – Apply the same deletion policy to the resource’s “UpdateReplacePolicy”. Default: truedefault (
Optional
[RemovalPolicy
]) – The default policy to apply in case the removal policy is not defined. Default: - Default value is resource specific. To determine the default value for a resoure, please consult that specific resource’s documentation.
- Return type:
None
- get_att(attribute_name)
Returns a token for an runtime attribute of this resource.
Ideally, use generated attribute accessors (e.g.
resource.arn
), but this can be used for future compatibility in case there is no generated attribute.- Parameters:
attribute_name (
str
) – The name of the attribute.- Return type:
- get_metadata(key)
Retrieve a value value from the CloudFormation Resource Metadata.
- Parameters:
key (
str
) –- See:
- Return type:
Any
https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/metadata-section-structure.html
Note that this is a different set of metadata from CDK node metadata; this metadata ends up in the stack template under the resource, whereas CDK node metadata ends up in the Cloud Assembly.
- inspect(inspector)
Examines the CloudFormation resource and discloses attributes.
- Parameters:
inspector (
TreeInspector
) –tree inspector to collect and process attributes.
- Return type:
None
- override_logical_id(new_logical_id)
Overrides the auto-generated logical ID with a specific ID.
- Parameters:
new_logical_id (
str
) – The new logical ID to use for this stack element.- Return type:
None
- to_string()
Returns a string representation of this construct.
- Return type:
str
- Returns:
a string representation of this resource
Attributes
- CFN_RESOURCE_TYPE_NAME = 'AWS::KafkaConnect::Connector'
- attr_connector_arn
The Amazon Resource Name (ARN) of the newly created connector.
- CloudformationAttribute:
ConnectorArn
- capacity
The connector’s compute capacity settings.
- cfn_options
Options for this resource, such as condition, update policy etc.
- cfn_resource_type
AWS resource type.
- connector_configuration
The configuration of the connector.
- connector_description
The description of the connector.
- connector_name
The name of the connector.
- creation_stack
return:
the stack trace of the point where this Resource was created from, sourced from the +metadata+ entry typed +aws:cdk:logicalId+, and with the bottom-most node +internal+ entries filtered.
- kafka_cluster
The details of the Apache Kafka cluster to which the connector is connected.
- kafka_cluster_client_authentication
The type of client authentication used to connect to the Apache Kafka cluster.
The value is NONE when no client authentication is used.
- kafka_cluster_encryption_in_transit
Details of encryption in transit to the Apache Kafka cluster.
- kafka_connect_version
The version of Kafka Connect.
It has to be compatible with both the Apache Kafka cluster’s version and the plugins.
- log_delivery
The settings for delivering connector logs to Amazon CloudWatch Logs.
- logical_id
The logical ID for this CloudFormation stack element.
The logical ID of the element is calculated from the path of the resource node in the construct tree.
To override this value, use
overrideLogicalId(newLogicalId)
.- Returns:
the logical ID as a stringified token. This value will only get resolved during synthesis.
- node
The construct tree node associated with this construct.
- plugins
Specifies which plugin to use for the connector.
You must specify a single-element list. Amazon MSK Connect does not currently support specifying multiple plugins.
- ref
Return a string that will be resolved to a CloudFormation
{ Ref }
for this element.If, by any chance, the intrinsic reference of a resource is not a string, you could coerce it to an IResolvable through
Lazy.any({ produce: resource.ref })
.
- service_execution_role_arn
The Amazon Resource Name (ARN) of the IAM role used by the connector to access Amazon Web Services resources.
- stack
The stack in which this element is defined.
CfnElements must be defined within a stack scope (directly or indirectly).
- worker_configuration
The worker configurations that are in use with the connector.
Static Methods
- classmethod is_cfn_element(x)
Returns
true
if a construct is a stack element (i.e. part of the synthesized cloudformation template).Uses duck-typing instead of
instanceof
to allow stack elements from different versions of this library to be included in the same stack.- Parameters:
x (
Any
) –- Return type:
bool
- Returns:
The construct as a stack element or undefined if it is not a stack element.
- classmethod is_cfn_resource(construct)
Check whether the given construct is a CfnResource.
- Parameters:
construct (
IConstruct
) –- Return type:
bool
- classmethod is_construct(x)
Return whether the given object is a Construct.
- Parameters:
x (
Any
) –- Return type:
bool
ApacheKafkaClusterProperty
- class CfnConnector.ApacheKafkaClusterProperty(*, bootstrap_servers, vpc)
Bases:
object
The details of the Apache Kafka cluster to which the connector is connected.
- Parameters:
bootstrap_servers (
str
) – The bootstrap servers of the cluster.vpc (
Union
[IResolvable
,VpcProperty
,Dict
[str
,Any
]]) – Details of an Amazon VPC which has network connectivity to the Apache Kafka cluster.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect apache_kafka_cluster_property = kafkaconnect.CfnConnector.ApacheKafkaClusterProperty( bootstrap_servers="bootstrapServers", vpc=kafkaconnect.CfnConnector.VpcProperty( security_groups=["securityGroups"], subnets=["subnets"] ) )
Attributes
- bootstrap_servers
The bootstrap servers of the cluster.
- vpc
Details of an Amazon VPC which has network connectivity to the Apache Kafka cluster.
AutoScalingProperty
- class CfnConnector.AutoScalingProperty(*, max_worker_count, mcu_count, min_worker_count, scale_in_policy, scale_out_policy)
Bases:
object
Specifies how the connector scales.
- Parameters:
max_worker_count (
Union
[int
,float
]) – The maximum number of workers allocated to the connector.mcu_count (
Union
[int
,float
]) – The number of microcontroller units (MCUs) allocated to each connector worker. The valid values are 1,2,4,8.min_worker_count (
Union
[int
,float
]) – The minimum number of workers allocated to the connector.scale_in_policy (
Union
[IResolvable
,ScaleInPolicyProperty
,Dict
[str
,Any
]]) – The sacle-in policy for the connector.scale_out_policy (
Union
[IResolvable
,ScaleOutPolicyProperty
,Dict
[str
,Any
]]) – The sacle-out policy for the connector.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect auto_scaling_property = kafkaconnect.CfnConnector.AutoScalingProperty( max_worker_count=123, mcu_count=123, min_worker_count=123, scale_in_policy=kafkaconnect.CfnConnector.ScaleInPolicyProperty( cpu_utilization_percentage=123 ), scale_out_policy=kafkaconnect.CfnConnector.ScaleOutPolicyProperty( cpu_utilization_percentage=123 ) )
Attributes
- max_worker_count
The maximum number of workers allocated to the connector.
- mcu_count
The number of microcontroller units (MCUs) allocated to each connector worker.
The valid values are 1,2,4,8.
- min_worker_count
The minimum number of workers allocated to the connector.
- scale_in_policy
The sacle-in policy for the connector.
- scale_out_policy
The sacle-out policy for the connector.
CapacityProperty
- class CfnConnector.CapacityProperty(*, auto_scaling=None, provisioned_capacity=None)
Bases:
object
Information about the capacity of the connector, whether it is auto scaled or provisioned.
- Parameters:
auto_scaling (
Union
[IResolvable
,AutoScalingProperty
,Dict
[str
,Any
],None
]) – Information about the auto scaling parameters for the connector.provisioned_capacity (
Union
[IResolvable
,ProvisionedCapacityProperty
,Dict
[str
,Any
],None
]) – Details about a fixed capacity allocated to a connector.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect capacity_property = kafkaconnect.CfnConnector.CapacityProperty( auto_scaling=kafkaconnect.CfnConnector.AutoScalingProperty( max_worker_count=123, mcu_count=123, min_worker_count=123, scale_in_policy=kafkaconnect.CfnConnector.ScaleInPolicyProperty( cpu_utilization_percentage=123 ), scale_out_policy=kafkaconnect.CfnConnector.ScaleOutPolicyProperty( cpu_utilization_percentage=123 ) ), provisioned_capacity=kafkaconnect.CfnConnector.ProvisionedCapacityProperty( worker_count=123, # the properties below are optional mcu_count=123 ) )
Attributes
- auto_scaling
Information about the auto scaling parameters for the connector.
- provisioned_capacity
Details about a fixed capacity allocated to a connector.
CloudWatchLogsLogDeliveryProperty
- class CfnConnector.CloudWatchLogsLogDeliveryProperty(*, enabled, log_group=None)
Bases:
object
The settings for delivering connector logs to Amazon CloudWatch Logs.
- Parameters:
enabled (
Union
[bool
,IResolvable
]) – Whether log delivery to Amazon CloudWatch Logs is enabled.log_group (
Optional
[str
]) – The name of the CloudWatch log group that is the destination for log delivery.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect cloud_watch_logs_log_delivery_property = kafkaconnect.CfnConnector.CloudWatchLogsLogDeliveryProperty( enabled=False, # the properties below are optional log_group="logGroup" )
Attributes
- enabled
Whether log delivery to Amazon CloudWatch Logs is enabled.
- log_group
The name of the CloudWatch log group that is the destination for log delivery.
CustomPluginProperty
- class CfnConnector.CustomPluginProperty(*, custom_plugin_arn, revision)
Bases:
object
A plugin is an AWS resource that contains the code that defines a connector’s logic.
- Parameters:
custom_plugin_arn (
str
) – The Amazon Resource Name (ARN) of the custom plugin.revision (
Union
[int
,float
]) – The revision of the custom plugin.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect custom_plugin_property = kafkaconnect.CfnConnector.CustomPluginProperty( custom_plugin_arn="customPluginArn", revision=123 )
Attributes
- custom_plugin_arn
The Amazon Resource Name (ARN) of the custom plugin.
- revision
The revision of the custom plugin.
FirehoseLogDeliveryProperty
- class CfnConnector.FirehoseLogDeliveryProperty(*, enabled, delivery_stream=None)
Bases:
object
The settings for delivering logs to Amazon Kinesis Data Firehose.
- Parameters:
enabled (
Union
[bool
,IResolvable
]) – Specifies whether connector logs get delivered to Amazon Kinesis Data Firehose.delivery_stream (
Optional
[str
]) – The name of the Kinesis Data Firehose delivery stream that is the destination for log delivery.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect firehose_log_delivery_property = kafkaconnect.CfnConnector.FirehoseLogDeliveryProperty( enabled=False, # the properties below are optional delivery_stream="deliveryStream" )
Attributes
- delivery_stream
The name of the Kinesis Data Firehose delivery stream that is the destination for log delivery.
- enabled
Specifies whether connector logs get delivered to Amazon Kinesis Data Firehose.
KafkaClusterClientAuthenticationProperty
- class CfnConnector.KafkaClusterClientAuthenticationProperty(*, authentication_type)
Bases:
object
The client authentication information used in order to authenticate with the Apache Kafka cluster.
- Parameters:
authentication_type (
str
) – The type of client authentication used to connect to the Apache Kafka cluster. Value NONE means that no client authentication is used.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect kafka_cluster_client_authentication_property = kafkaconnect.CfnConnector.KafkaClusterClientAuthenticationProperty( authentication_type="authenticationType" )
Attributes
- authentication_type
The type of client authentication used to connect to the Apache Kafka cluster.
Value NONE means that no client authentication is used.
KafkaClusterEncryptionInTransitProperty
- class CfnConnector.KafkaClusterEncryptionInTransitProperty(*, encryption_type)
Bases:
object
Details of encryption in transit to the Apache Kafka cluster.
- Parameters:
encryption_type (
str
) – The type of encryption in transit to the Apache Kafka cluster.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect kafka_cluster_encryption_in_transit_property = kafkaconnect.CfnConnector.KafkaClusterEncryptionInTransitProperty( encryption_type="encryptionType" )
Attributes
- encryption_type
The type of encryption in transit to the Apache Kafka cluster.
KafkaClusterProperty
- class CfnConnector.KafkaClusterProperty(*, apache_kafka_cluster)
Bases:
object
The details of the Apache Kafka cluster to which the connector is connected.
- Parameters:
apache_kafka_cluster (
Union
[IResolvable
,ApacheKafkaClusterProperty
,Dict
[str
,Any
]]) – The Apache Kafka cluster to which the connector is connected.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect kafka_cluster_property = kafkaconnect.CfnConnector.KafkaClusterProperty( apache_kafka_cluster=kafkaconnect.CfnConnector.ApacheKafkaClusterProperty( bootstrap_servers="bootstrapServers", vpc=kafkaconnect.CfnConnector.VpcProperty( security_groups=["securityGroups"], subnets=["subnets"] ) ) )
Attributes
- apache_kafka_cluster
The Apache Kafka cluster to which the connector is connected.
LogDeliveryProperty
- class CfnConnector.LogDeliveryProperty(*, worker_log_delivery)
Bases:
object
Details about log delivery.
- Parameters:
worker_log_delivery (
Union
[IResolvable
,WorkerLogDeliveryProperty
,Dict
[str
,Any
]]) – The workers can send worker logs to different destination types. This configuration specifies the details of these destinations.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect log_delivery_property = kafkaconnect.CfnConnector.LogDeliveryProperty( worker_log_delivery=kafkaconnect.CfnConnector.WorkerLogDeliveryProperty( cloud_watch_logs=kafkaconnect.CfnConnector.CloudWatchLogsLogDeliveryProperty( enabled=False, # the properties below are optional log_group="logGroup" ), firehose=kafkaconnect.CfnConnector.FirehoseLogDeliveryProperty( enabled=False, # the properties below are optional delivery_stream="deliveryStream" ), s3=kafkaconnect.CfnConnector.S3LogDeliveryProperty( enabled=False, # the properties below are optional bucket="bucket", prefix="prefix" ) ) )
Attributes
- worker_log_delivery
The workers can send worker logs to different destination types.
This configuration specifies the details of these destinations.
PluginProperty
- class CfnConnector.PluginProperty(*, custom_plugin)
Bases:
object
A plugin is an AWS resource that contains the code that defines your connector logic.
- Parameters:
custom_plugin (
Union
[IResolvable
,CustomPluginProperty
,Dict
[str
,Any
]]) – Details about a custom plugin.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect plugin_property = kafkaconnect.CfnConnector.PluginProperty( custom_plugin=kafkaconnect.CfnConnector.CustomPluginProperty( custom_plugin_arn="customPluginArn", revision=123 ) )
Attributes
- custom_plugin
Details about a custom plugin.
ProvisionedCapacityProperty
- class CfnConnector.ProvisionedCapacityProperty(*, worker_count, mcu_count=None)
Bases:
object
Details about a connector’s provisioned capacity.
- Parameters:
worker_count (
Union
[int
,float
]) – The number of workers that are allocated to the connector.mcu_count (
Union
[int
,float
,None
]) – The number of microcontroller units (MCUs) allocated to each connector worker. The valid values are 1,2,4,8.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect provisioned_capacity_property = kafkaconnect.CfnConnector.ProvisionedCapacityProperty( worker_count=123, # the properties below are optional mcu_count=123 )
Attributes
- mcu_count
The number of microcontroller units (MCUs) allocated to each connector worker.
The valid values are 1,2,4,8.
- worker_count
The number of workers that are allocated to the connector.
S3LogDeliveryProperty
- class CfnConnector.S3LogDeliveryProperty(*, enabled, bucket=None, prefix=None)
Bases:
object
Details about delivering logs to Amazon S3.
- Parameters:
enabled (
Union
[bool
,IResolvable
]) – Specifies whether connector logs get sent to the specified Amazon S3 destination.bucket (
Optional
[str
]) – The name of the S3 bucket that is the destination for log delivery.prefix (
Optional
[str
]) – The S3 prefix that is the destination for log delivery.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect s3_log_delivery_property = kafkaconnect.CfnConnector.S3LogDeliveryProperty( enabled=False, # the properties below are optional bucket="bucket", prefix="prefix" )
Attributes
- bucket
The name of the S3 bucket that is the destination for log delivery.
- enabled
Specifies whether connector logs get sent to the specified Amazon S3 destination.
- prefix
The S3 prefix that is the destination for log delivery.
ScaleInPolicyProperty
- class CfnConnector.ScaleInPolicyProperty(*, cpu_utilization_percentage)
Bases:
object
The scale-in policy for the connector.
- Parameters:
cpu_utilization_percentage (
Union
[int
,float
]) – Specifies the CPU utilization percentage threshold at which you want connector scale in to be triggered.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect scale_in_policy_property = kafkaconnect.CfnConnector.ScaleInPolicyProperty( cpu_utilization_percentage=123 )
Attributes
- cpu_utilization_percentage
Specifies the CPU utilization percentage threshold at which you want connector scale in to be triggered.
ScaleOutPolicyProperty
- class CfnConnector.ScaleOutPolicyProperty(*, cpu_utilization_percentage)
Bases:
object
The scale-out policy for the connector.
- Parameters:
cpu_utilization_percentage (
Union
[int
,float
]) – The CPU utilization percentage threshold at which you want connector scale out to be triggered.- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect scale_out_policy_property = kafkaconnect.CfnConnector.ScaleOutPolicyProperty( cpu_utilization_percentage=123 )
Attributes
- cpu_utilization_percentage
The CPU utilization percentage threshold at which you want connector scale out to be triggered.
VpcProperty
- class CfnConnector.VpcProperty(*, security_groups, subnets)
Bases:
object
Information about the VPC in which the connector resides.
- Parameters:
security_groups (
Sequence
[str
]) – The security groups for the connector.subnets (
Sequence
[str
]) – The subnets for the connector.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect vpc_property = kafkaconnect.CfnConnector.VpcProperty( security_groups=["securityGroups"], subnets=["subnets"] )
Attributes
- security_groups
The security groups for the connector.
- subnets
The subnets for the connector.
WorkerConfigurationProperty
- class CfnConnector.WorkerConfigurationProperty(*, revision, worker_configuration_arn)
Bases:
object
The configuration of the workers, which are the processes that run the connector logic.
- Parameters:
revision (
Union
[int
,float
]) – The revision of the worker configuration.worker_configuration_arn (
str
) – The Amazon Resource Name (ARN) of the worker configuration.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect worker_configuration_property = kafkaconnect.CfnConnector.WorkerConfigurationProperty( revision=123, worker_configuration_arn="workerConfigurationArn" )
Attributes
- revision
The revision of the worker configuration.
- worker_configuration_arn
The Amazon Resource Name (ARN) of the worker configuration.
WorkerLogDeliveryProperty
- class CfnConnector.WorkerLogDeliveryProperty(*, cloud_watch_logs=None, firehose=None, s3=None)
Bases:
object
Workers can send worker logs to different destination types.
This configuration specifies the details of these destinations.
- Parameters:
cloud_watch_logs (
Union
[IResolvable
,CloudWatchLogsLogDeliveryProperty
,Dict
[str
,Any
],None
]) – Details about delivering logs to Amazon CloudWatch Logs.firehose (
Union
[IResolvable
,FirehoseLogDeliveryProperty
,Dict
[str
,Any
],None
]) – Details about delivering logs to Amazon Kinesis Data Firehose.s3 (
Union
[IResolvable
,S3LogDeliveryProperty
,Dict
[str
,Any
],None
]) – Details about delivering logs to Amazon S3.
- Link:
- ExampleMetadata:
fixture=_generated
Example:
# The code below shows an example of how to instantiate this type. # The values are placeholders you should change. import aws_cdk.aws_kafkaconnect as kafkaconnect worker_log_delivery_property = kafkaconnect.CfnConnector.WorkerLogDeliveryProperty( cloud_watch_logs=kafkaconnect.CfnConnector.CloudWatchLogsLogDeliveryProperty( enabled=False, # the properties below are optional log_group="logGroup" ), firehose=kafkaconnect.CfnConnector.FirehoseLogDeliveryProperty( enabled=False, # the properties below are optional delivery_stream="deliveryStream" ), s3=kafkaconnect.CfnConnector.S3LogDeliveryProperty( enabled=False, # the properties below are optional bucket="bucket", prefix="prefix" ) )
Attributes
- cloud_watch_logs
Details about delivering logs to Amazon CloudWatch Logs.
- firehose
Details about delivering logs to Amazon Kinesis Data Firehose.
- s3
Details about delivering logs to Amazon S3.