@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class TransformInput extends Object implements Serializable, Cloneable, StructuredPojo
Describes the input source of a transform job and the way the transform job consumes it.
| Constructor and Description |
|---|
TransformInput() |
| Modifier and Type | Method and Description |
|---|---|
TransformInput |
clone() |
boolean |
equals(Object obj) |
String |
getCompressionType()
If your transform data is compressed, specify the compression type.
|
String |
getContentType()
The multipurpose internet mail extension (MIME) type of the data.
|
TransformDataSource |
getDataSource()
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
String |
getSplitType()
The method to use to split the transform job's data files into smaller batches.
|
int |
hashCode() |
void |
marshall(ProtocolMarshaller protocolMarshaller)
Marshalls this structured data using the given
ProtocolMarshaller. |
void |
setCompressionType(String compressionType)
If your transform data is compressed, specify the compression type.
|
void |
setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
void |
setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
void |
setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
String |
toString()
Returns a string representation of this object.
|
TransformInput |
withCompressionType(CompressionType compressionType)
If your transform data is compressed, specify the compression type.
|
TransformInput |
withCompressionType(String compressionType)
If your transform data is compressed, specify the compression type.
|
TransformInput |
withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
TransformInput |
withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
TransformInput |
withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches.
|
TransformInput |
withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
public void setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.public TransformDataSource getDataSource()
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
public TransformInput withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource - Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.public void setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public String getContentType()
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
public TransformInput withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType - The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public void setCompressionType(String compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None.
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic String getCompressionType()
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None.
None.CompressionTypepublic TransformInput withCompressionType(String compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None.
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic TransformInput withCompressionType(CompressionType compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None.
compressionType - If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None.CompressionTypepublic void setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
SplitTypepublic String getSplitType()
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports
a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in
each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
SplitTypepublic TransformInput withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
SplitTypepublic TransformInput withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType is
None, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line to split records on a newline
character boundary. SplitType also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy and
MaxPayloadInMB parameters. When the value of BatchStrategy is MultiRecord,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy is SingleRecord, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy is set to
SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
splitType - The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType is None, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line to split records on a newline character boundary. SplitType also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy and MaxPayloadInMB parameters. When the value of
BatchStrategy is MultiRecord, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB limit. If the value of
BatchStrategy is SingleRecord, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy is set
to SingleRecord. Padding is not removed if the value of BatchStrategy is set to
MultiRecord.
For more information about RecordIO, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord, see Consuming TFRecord data in the
TensorFlow documentation.
SplitTypepublic String toString()
toString in class ObjectObject.toString()public TransformInput clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojoProtocolMarshaller.marshall in interface StructuredPojoprotocolMarshaller - Implementation of ProtocolMarshaller used to marshall this object's data.