interface ParquetOutputFormatProps
| Language | Type name |
|---|---|
.NET | Amazon.CDK.AWS.KinesisFirehose.ParquetOutputFormatProps |
Go | github.com/aws/aws-cdk-go/awscdk/v2/awskinesisfirehose#ParquetOutputFormatProps |
Java | software.amazon.awscdk.services.kinesisfirehose.ParquetOutputFormatProps |
Python | aws_cdk.aws_kinesisfirehose.ParquetOutputFormatProps |
TypeScript (source) | aws-cdk-lib » aws_kinesisfirehose » ParquetOutputFormatProps |
Props for Parquet output format for data record format conversion.
Example
const outputFormat = new firehose.ParquetOutputFormat({
blockSize: Size.mebibytes(512),
compression: firehose.ParquetCompression.UNCOMPRESSED,
enableDictionaryCompression: true,
maxPadding: Size.bytes(10),
pageSize: Size.mebibytes(2),
writerVersion: firehose.ParquetWriterVersion.V2,
})
Properties
| Name | Type | Description |
|---|---|---|
| block | Size | The Hadoop Distributed File System (HDFS) block size. |
| compression? | Parquet | The compression code to use over data blocks. |
| enable | boolean | Indicates whether to enable dictionary compression. |
| max | Size | The maximum amount of padding to apply. |
| page | Size | The Parquet page size. |
| writer | Parquet | Indicates the version of Parquet to output. |
blockSize?
Type:
Size
(optional, default: Size.mebibytes(256))
The Hadoop Distributed File System (HDFS) block size.
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. Firehose uses this value for padding calculations.
compression?
Type:
Parquet
(optional, default: SNAPPY)
The compression code to use over data blocks.
The possible values are UNCOMPRESSED , SNAPPY , and GZIP.
Use SNAPPY for higher decompression speed.
Use GZIP if the compression ratio is more important than speed.
enableDictionaryCompression?
Type:
boolean
(optional, default: false)
Indicates whether to enable dictionary compression.
maxPadding?
Type:
Size
(optional, default: no padding is applied)
The maximum amount of padding to apply.
This is useful if you intend to copy the data from Amazon S3 to HDFS before querying.
pageSize?
Type:
Size
(optional, default: Size.mebibytes(1))
The Parquet page size.
Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
writerVersion?
Type:
Parquet
(optional, default: V1)
Indicates the version of Parquet to output.
The possible values are V1 and V2

.NET
Go
Java
Python
TypeScript (