@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class ContinuousExportDescription extends Object implements Serializable, Cloneable, StructuredPojo
A list of continuous export descriptions.
Constructor and Description |
---|
ContinuousExportDescription() |
Modifier and Type | Method and Description |
---|---|
ContinuousExportDescription |
addSchemaStorageConfigEntry(String key,
String value)
Add a single SchemaStorageConfig entry
|
ContinuousExportDescription |
clearSchemaStorageConfigEntries()
Removes all the entries added into SchemaStorageConfig.
|
ContinuousExportDescription |
clone() |
boolean |
equals(Object obj) |
String |
getDataSource()
The type of data collector used to gather this data (currently only offered for AGENT).
|
String |
getExportId()
The unique ID assigned to this export.
|
String |
getS3Bucket()
The name of the s3 bucket where the export data parquet files are stored.
|
Map<String,String> |
getSchemaStorageConfig()
An object which describes how the data is stored.
|
Date |
getStartTime()
The timestamp representing when the continuous export was started.
|
String |
getStatus()
Describes the status of the export.
|
String |
getStatusDetail()
Contains information about any errors that have occurred.
|
Date |
getStopTime()
The timestamp that represents when this continuous export was stopped.
|
int |
hashCode() |
void |
marshall(ProtocolMarshaller protocolMarshaller)
Marshalls this structured data using the given
ProtocolMarshaller . |
void |
setDataSource(String dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
|
void |
setExportId(String exportId)
The unique ID assigned to this export.
|
void |
setS3Bucket(String s3Bucket)
The name of the s3 bucket where the export data parquet files are stored.
|
void |
setSchemaStorageConfig(Map<String,String> schemaStorageConfig)
An object which describes how the data is stored.
|
void |
setStartTime(Date startTime)
The timestamp representing when the continuous export was started.
|
void |
setStatus(String status)
Describes the status of the export.
|
void |
setStatusDetail(String statusDetail)
Contains information about any errors that have occurred.
|
void |
setStopTime(Date stopTime)
The timestamp that represents when this continuous export was stopped.
|
String |
toString()
Returns a string representation of this object.
|
ContinuousExportDescription |
withDataSource(DataSource dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
|
ContinuousExportDescription |
withDataSource(String dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
|
ContinuousExportDescription |
withExportId(String exportId)
The unique ID assigned to this export.
|
ContinuousExportDescription |
withS3Bucket(String s3Bucket)
The name of the s3 bucket where the export data parquet files are stored.
|
ContinuousExportDescription |
withSchemaStorageConfig(Map<String,String> schemaStorageConfig)
An object which describes how the data is stored.
|
ContinuousExportDescription |
withStartTime(Date startTime)
The timestamp representing when the continuous export was started.
|
ContinuousExportDescription |
withStatus(ContinuousExportStatus status)
Describes the status of the export.
|
ContinuousExportDescription |
withStatus(String status)
Describes the status of the export.
|
ContinuousExportDescription |
withStatusDetail(String statusDetail)
Contains information about any errors that have occurred.
|
ContinuousExportDescription |
withStopTime(Date stopTime)
The timestamp that represents when this continuous export was stopped.
|
public void setExportId(String exportId)
The unique ID assigned to this export.
exportId
- The unique ID assigned to this export.public String getExportId()
The unique ID assigned to this export.
public ContinuousExportDescription withExportId(String exportId)
The unique ID assigned to this export.
exportId
- The unique ID assigned to this export.public void setStatus(String status)
Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
status
- Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
ContinuousExportStatus
public String getStatus()
Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
ContinuousExportStatus
public ContinuousExportDescription withStatus(String status)
Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
status
- Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
ContinuousExportStatus
public ContinuousExportDescription withStatus(ContinuousExportStatus status)
Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
status
- Describes the status of the export. Can be one of the following values:
START_IN_PROGRESS - setting up resources to start continuous export.
START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.
ACTIVE - data is being exported to the customer bucket.
ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.
STOP_IN_PROGRESS - stopping the export.
STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.
INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.
ContinuousExportStatus
public void setStatusDetail(String statusDetail)
Contains information about any errors that have occurred. This data type can have the following values:
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
statusDetail
- Contains information about any errors that have occurred. This data type can have the following
values:
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
public String getStatusDetail()
Contains information about any errors that have occurred. This data type can have the following values:
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
public ContinuousExportDescription withStatusDetail(String statusDetail)
Contains information about any errors that have occurred. This data type can have the following values:
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
statusDetail
- Contains information about any errors that have occurred. This data type can have the following
values:
ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.
DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.
FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.
FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.
INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.
LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.
You can use one of the following two ways to resolve this issue.
If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.
You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.
AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.
AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.
S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.
S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.
public void setS3Bucket(String s3Bucket)
The name of the s3 bucket where the export data parquet files are stored.
s3Bucket
- The name of the s3 bucket where the export data parquet files are stored.public String getS3Bucket()
The name of the s3 bucket where the export data parquet files are stored.
public ContinuousExportDescription withS3Bucket(String s3Bucket)
The name of the s3 bucket where the export data parquet files are stored.
s3Bucket
- The name of the s3 bucket where the export data parquet files are stored.public void setStartTime(Date startTime)
The timestamp representing when the continuous export was started.
startTime
- The timestamp representing when the continuous export was started.public Date getStartTime()
The timestamp representing when the continuous export was started.
public ContinuousExportDescription withStartTime(Date startTime)
The timestamp representing when the continuous export was started.
startTime
- The timestamp representing when the continuous export was started.public void setStopTime(Date stopTime)
The timestamp that represents when this continuous export was stopped.
stopTime
- The timestamp that represents when this continuous export was stopped.public Date getStopTime()
The timestamp that represents when this continuous export was stopped.
public ContinuousExportDescription withStopTime(Date stopTime)
The timestamp that represents when this continuous export was stopped.
stopTime
- The timestamp that represents when this continuous export was stopped.public void setDataSource(String dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
dataSource
- The type of data collector used to gather this data (currently only offered for AGENT).DataSource
public String getDataSource()
The type of data collector used to gather this data (currently only offered for AGENT).
DataSource
public ContinuousExportDescription withDataSource(String dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
dataSource
- The type of data collector used to gather this data (currently only offered for AGENT).DataSource
public ContinuousExportDescription withDataSource(DataSource dataSource)
The type of data collector used to gather this data (currently only offered for AGENT).
dataSource
- The type of data collector used to gather this data (currently only offered for AGENT).DataSource
public Map<String,String> getSchemaStorageConfig()
An object which describes how the data is stored.
databaseName
- the name of the Glue database used to store the schema.
databaseName
- the name of the Glue database used to store the schema.
public void setSchemaStorageConfig(Map<String,String> schemaStorageConfig)
An object which describes how the data is stored.
databaseName
- the name of the Glue database used to store the schema.
schemaStorageConfig
- An object which describes how the data is stored.
databaseName
- the name of the Glue database used to store the schema.
public ContinuousExportDescription withSchemaStorageConfig(Map<String,String> schemaStorageConfig)
An object which describes how the data is stored.
databaseName
- the name of the Glue database used to store the schema.
schemaStorageConfig
- An object which describes how the data is stored.
databaseName
- the name of the Glue database used to store the schema.
public ContinuousExportDescription addSchemaStorageConfigEntry(String key, String value)
public ContinuousExportDescription clearSchemaStorageConfigEntries()
public String toString()
toString
in class Object
Object.toString()
public ContinuousExportDescription clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojo
ProtocolMarshaller
.marshall
in interface StructuredPojo
protocolMarshaller
- Implementation of ProtocolMarshaller
used to marshall this object's data.