AWS services or capabilities described in AWS Documentation may vary by region/location. Click Getting Started with Amazon AWS to see specific differences applicable to the China (Beijing) Region.
A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet.
Namespace: Amazon.KinesisFirehose.Model
Assembly: AWSSDK.KinesisFirehose.dll
Version: 3.x.y.z
public class ParquetSerDe
The ParquetSerDe type exposes the following members
Name | Description | |
---|---|---|
ParquetSerDe() |
Name | Type | Description | |
---|---|---|---|
BlockSizeBytes | System.Int32 |
Gets and sets the property BlockSizeBytes. The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations. |
|
Compression | Amazon.KinesisFirehose.ParquetCompression |
Gets and sets the property Compression.
The compression code to use over data blocks. The possible values are |
|
EnableDictionaryCompression | System.Boolean |
Gets and sets the property EnableDictionaryCompression. Indicates whether to enable dictionary compression. |
|
MaxPaddingBytes | System.Int32 |
Gets and sets the property MaxPaddingBytes. The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0. |
|
PageSizeBytes | System.Int32 |
Gets and sets the property PageSizeBytes. The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB. |
|
WriterVersion | Amazon.KinesisFirehose.ParquetWriterVersion |
Gets and sets the property WriterVersion.
Indicates the version of row format to output. The possible values are |
.NET Core App:
Supported in: 3.1
.NET Standard:
Supported in: 2.0
.NET Framework:
Supported in: 4.5, 4.0, 3.5