@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public class TransformInput extends Object implements Serializable, Cloneable, StructuredPojo
Describes the input source of a transform job and the way the transform job consumes it.
Constructor and Description |
---|
TransformInput() |
Modifier and Type | Method and Description |
---|---|
TransformInput |
clone() |
boolean |
equals(Object obj) |
String |
getCompressionType()
If your transform data is compressed, specify the compression type.
|
String |
getContentType()
The multipurpose internet mail extension (MIME) type of the data.
|
TransformDataSource |
getDataSource()
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
String |
getSplitType()
The method to use to split the transform job's data files into smaller batches.
|
int |
hashCode() |
void |
marshall(ProtocolMarshaller protocolMarshaller)
Marshalls this structured data using the given
ProtocolMarshaller . |
void |
setCompressionType(String compressionType)
If your transform data is compressed, specify the compression type.
|
void |
setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
void |
setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
void |
setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
String |
toString()
Returns a string representation of this object.
|
TransformInput |
withCompressionType(CompressionType compressionType)
If your transform data is compressed, specify the compression type.
|
TransformInput |
withCompressionType(String compressionType)
If your transform data is compressed, specify the compression type.
|
TransformInput |
withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data.
|
TransformInput |
withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.
|
TransformInput |
withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches.
|
TransformInput |
withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches.
|
public void setDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource
- Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.public TransformDataSource getDataSource()
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
public TransformInput withDataSource(TransformDataSource dataSource)
Describes the location of the channel data, which is, the S3 location of the input data that the model can consume.
dataSource
- Describes the location of the channel data, which is, the S3 location of the input data that the model can
consume.public void setContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType
- The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public String getContentType()
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
public TransformInput withContentType(String contentType)
The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with each http call to transfer data to the transform job.
contentType
- The multipurpose internet mail extension (MIME) type of the data. Amazon SageMaker uses the MIME type with
each http call to transfer data to the transform job.public void setCompressionType(String compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None
.
compressionType
- If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None
.CompressionType
public String getCompressionType()
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None
.
None
.CompressionType
public TransformInput withCompressionType(String compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None
.
compressionType
- If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None
.CompressionType
public TransformInput withCompressionType(CompressionType compressionType)
If your transform data is compressed, specify the compression type. Amazon SageMaker automatically decompresses
the data for the transform job accordingly. The default value is None
.
compressionType
- If your transform data is compressed, specify the compression type. Amazon SageMaker automatically
decompresses the data for the transform job accordingly. The default value is None
.CompressionType
public void setSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType
is
None
, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line
to split records on a newline
character boundary. SplitType
also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy
and
MaxPayloadInMB
parameters. When the value of BatchStrategy
is MultiRecord
,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy
is set to
SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
splitType
- The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType
is None
, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line
to split records on a newline character boundary. SplitType
also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy
and MaxPayloadInMB
parameters. When the value of
BatchStrategy
is MultiRecord
, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB
limit. If the value of
BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy
is set
to SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType
public String getSplitType()
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType
is
None
, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line
to split records on a newline
character boundary. SplitType
also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy
and
MaxPayloadInMB
parameters. When the value of BatchStrategy
is MultiRecord
,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy
is set to
SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType
is None
, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line
to split records on a newline character boundary. SplitType
also supports
a number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy
and MaxPayloadInMB
parameters. When the value of
BatchStrategy
is MultiRecord
, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB
limit. If the value of
BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual records in
each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy
is set
to SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType
public TransformInput withSplitType(String splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType
is
None
, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line
to split records on a newline
character boundary. SplitType
also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy
and
MaxPayloadInMB
parameters. When the value of BatchStrategy
is MultiRecord
,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy
is set to
SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
splitType
- The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType
is None
, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line
to split records on a newline character boundary. SplitType
also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy
and MaxPayloadInMB
parameters. When the value of
BatchStrategy
is MultiRecord
, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB
limit. If the value of
BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy
is set
to SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType
public TransformInput withSplitType(SplitType splitType)
The method to use to split the transform job's data files into smaller batches. Splitting is necessary when the
total size of each object is too large to fit in a single request. You can also use data splitting to improve
performance by processing multiple concurrent mini-batches. The default value for SplitType
is
None
, which indicates that input data files are not split, and request payloads contain the entire
contents of an input object. Set the value of this parameter to Line
to split records on a newline
character boundary. SplitType
also supports a number of record-oriented binary data formats.
Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the BatchStrategy
and
MaxPayloadInMB
parameters. When the value of BatchStrategy
is MultiRecord
,
Amazon SageMaker sends the maximum number of records in each request, up to the MaxPayloadInMB
limit. If the value of BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual
records in each request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting is
applied to a binary data format, padding is removed if the value of BatchStrategy
is set to
SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create
a Dataset Using RecordIO in the MXNet documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
splitType
- The method to use to split the transform job's data files into smaller batches. Splitting is necessary
when the total size of each object is too large to fit in a single request. You can also use data
splitting to improve performance by processing multiple concurrent mini-batches. The default value for
SplitType
is None
, which indicates that input data files are not split, and
request payloads contain the entire contents of an input object. Set the value of this parameter to
Line
to split records on a newline character boundary. SplitType
also supports a
number of record-oriented binary data formats. Currently, the supported record formats are:
RecordIO
TFRecord
When splitting is enabled, the size of a mini-batch depends on the values of the
BatchStrategy
and MaxPayloadInMB
parameters. When the value of
BatchStrategy
is MultiRecord
, Amazon SageMaker sends the maximum number of
records in each request, up to the MaxPayloadInMB
limit. If the value of
BatchStrategy
is SingleRecord
, Amazon SageMaker sends individual records in each
request.
Some data formats represent a record as a binary payload wrapped with extra padding bytes. When splitting
is applied to a binary data format, padding is removed if the value of BatchStrategy
is set
to SingleRecord
. Padding is not removed if the value of BatchStrategy
is set to
MultiRecord
.
For more information about RecordIO
, see Create a Dataset Using RecordIO in the MXNet
documentation. For more information about TFRecord
, see Consuming TFRecord data in the
TensorFlow documentation.
SplitType
public String toString()
toString
in class Object
Object.toString()
public TransformInput clone()
public void marshall(ProtocolMarshaller protocolMarshaller)
StructuredPojo
ProtocolMarshaller
.marshall
in interface StructuredPojo
protocolMarshaller
- Implementation of ProtocolMarshaller
used to marshall this object's data.