Interface CfnDeliveryStream.ParquetSerDeProperty
- All Superinterfaces:
software.amazon.jsii.JsiiSerializable
- All Known Implementing Classes:
CfnDeliveryStream.ParquetSerDeProperty.Jsii$Proxy
- Enclosing class:
CfnDeliveryStream
For more information, see Apache Parquet .
Example:
// The code below shows an example of how to instantiate this type. // The values are placeholders you should change. import software.amazon.awscdk.services.kinesisfirehose.*; ParquetSerDeProperty parquetSerDeProperty = ParquetSerDeProperty.builder() .blockSizeBytes(123) .compression("compression") .enableDictionaryCompression(false) .maxPaddingBytes(123) .pageSizeBytes(123) .writerVersion("writerVersion") .build();
- See Also:
-
Nested Class Summary
Modifier and TypeInterfaceDescriptionstatic final class
A builder forCfnDeliveryStream.ParquetSerDeProperty
static final class
An implementation forCfnDeliveryStream.ParquetSerDeProperty
-
Method Summary
Modifier and TypeMethodDescriptionbuilder()
default Number
The Hadoop Distributed File System (HDFS) block size.default String
The compression code to use over data blocks.default Object
Indicates whether to enable dictionary compression.default Number
The maximum amount of padding to apply.default Number
The Parquet page size.default String
Indicates the version of row format to output.Methods inherited from interface software.amazon.jsii.JsiiSerializable
$jsii$toJson
-
Method Details
-
getBlockSizeBytes
The Hadoop Distributed File System (HDFS) block size.This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Firehose uses this value for padding calculations.
- See Also:
-
getCompression
The compression code to use over data blocks.The possible values are
UNCOMPRESSED
,SNAPPY
, andGZIP
, with the default beingSNAPPY
. UseSNAPPY
for higher decompression speed. UseGZIP
if the compression ratio is more important than speed.- See Also:
-
getEnableDictionaryCompression
Indicates whether to enable dictionary compression.- See Also:
-
getMaxPaddingBytes
The maximum amount of padding to apply.This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
- See Also:
-
getPageSizeBytes
The Parquet page size.Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
- See Also:
-
getWriterVersion
Indicates the version of row format to output.The possible values are
V1
andV2
. The default isV1
.- See Also:
-
builder
-