public static interface CfnDeliveryStream.ParquetSerDeProperty
For more information, see Apache Parquet .
Example:
// The code below shows an example of how to instantiate this type. // The values are placeholders you should change. import software.amazon.awscdk.services.kinesisfirehose.*; ParquetSerDeProperty parquetSerDeProperty = ParquetSerDeProperty.builder() .blockSizeBytes(123) .compression("compression") .enableDictionaryCompression(false) .maxPaddingBytes(123) .pageSizeBytes(123) .writerVersion("writerVersion") .build();
Modifier and Type | Interface and Description |
---|---|
static class |
CfnDeliveryStream.ParquetSerDeProperty.Builder
A builder for
CfnDeliveryStream.ParquetSerDeProperty |
static class |
CfnDeliveryStream.ParquetSerDeProperty.Jsii$Proxy
An implementation for
CfnDeliveryStream.ParquetSerDeProperty |
Modifier and Type | Method and Description |
---|---|
static CfnDeliveryStream.ParquetSerDeProperty.Builder |
builder() |
default java.lang.Number |
getBlockSizeBytes()
The Hadoop Distributed File System (HDFS) block size.
|
default java.lang.String |
getCompression()
The compression code to use over data blocks.
|
default java.lang.Object |
getEnableDictionaryCompression()
Indicates whether to enable dictionary compression.
|
default java.lang.Number |
getMaxPaddingBytes()
The maximum amount of padding to apply.
|
default java.lang.Number |
getPageSizeBytes()
The Parquet page size.
|
default java.lang.String |
getWriterVersion()
Indicates the version of row format to output.
|
default java.lang.Number getBlockSizeBytes()
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
default java.lang.String getCompression()
The possible values are UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
. Use SNAPPY
for higher decompression speed. Use GZIP
if the compression ratio is more important than speed.
default java.lang.Object getEnableDictionaryCompression()
default java.lang.Number getMaxPaddingBytes()
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
default java.lang.Number getPageSizeBytes()
Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
default java.lang.String getWriterVersion()
The possible values are V1
and V2
. The default is V1
.
static CfnDeliveryStream.ParquetSerDeProperty.Builder builder()