Copyright | (c) 2013-2023 Brendan Hay |
---|---|
License | Mozilla Public License, v. 2.0. |
Maintainer | Brendan Hay |
Stability | auto-generated |
Portability | non-portable (GHC extensions) |
Safe Haskell | Safe-Inferred |
Language | Haskell2010 |
Synopsis
- data ParquetSerDe = ParquetSerDe' {}
- newParquetSerDe :: ParquetSerDe
- parquetSerDe_blockSizeBytes :: Lens' ParquetSerDe (Maybe Natural)
- parquetSerDe_compression :: Lens' ParquetSerDe (Maybe ParquetCompression)
- parquetSerDe_enableDictionaryCompression :: Lens' ParquetSerDe (Maybe Bool)
- parquetSerDe_maxPaddingBytes :: Lens' ParquetSerDe (Maybe Natural)
- parquetSerDe_pageSizeBytes :: Lens' ParquetSerDe (Maybe Natural)
- parquetSerDe_writerVersion :: Lens' ParquetSerDe (Maybe ParquetWriterVersion)
Documentation
data ParquetSerDe Source #
A serializer to use for converting data to the Parquet format before storing it in Amazon S3. For more information, see Apache Parquet.
See: newParquetSerDe
smart constructor.
ParquetSerDe' | |
|
Instances
newParquetSerDe :: ParquetSerDe Source #
Create a value of ParquetSerDe
with all optional fields omitted.
Use generic-lens or optics to modify other optional fields.
The following record fields are available, with the corresponding lenses provided for backwards compatibility:
$sel:blockSizeBytes:ParquetSerDe'
, parquetSerDe_blockSizeBytes
- The Hadoop Distributed File System (HDFS) block size. This is useful if
you intend to copy the data from Amazon S3 to HDFS before querying. The
default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses
this value for padding calculations.
$sel:compression:ParquetSerDe'
, parquetSerDe_compression
- The compression code to use over data blocks. The possible values are
UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
.
Use SNAPPY
for higher decompression speed. Use GZIP
if the
compression ratio is more important than speed.
$sel:enableDictionaryCompression:ParquetSerDe'
, parquetSerDe_enableDictionaryCompression
- Indicates whether to enable dictionary compression.
$sel:maxPaddingBytes:ParquetSerDe'
, parquetSerDe_maxPaddingBytes
- The maximum amount of padding to apply. This is useful if you intend to
copy the data from Amazon S3 to HDFS before querying. The default is 0.
$sel:pageSizeBytes:ParquetSerDe'
, parquetSerDe_pageSizeBytes
- The Parquet page size. Column chunks are divided into pages. A page is
conceptually an indivisible unit (in terms of compression and encoding).
The minimum value is 64 KiB and the default is 1 MiB.
$sel:writerVersion:ParquetSerDe'
, parquetSerDe_writerVersion
- Indicates the version of row format to output. The possible values are
V1
and V2
. The default is V1
.
parquetSerDe_blockSizeBytes :: Lens' ParquetSerDe (Maybe Natural) Source #
The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.
parquetSerDe_compression :: Lens' ParquetSerDe (Maybe ParquetCompression) Source #
The compression code to use over data blocks. The possible values are
UNCOMPRESSED
, SNAPPY
, and GZIP
, with the default being SNAPPY
.
Use SNAPPY
for higher decompression speed. Use GZIP
if the
compression ratio is more important than speed.
parquetSerDe_enableDictionaryCompression :: Lens' ParquetSerDe (Maybe Bool) Source #
Indicates whether to enable dictionary compression.
parquetSerDe_maxPaddingBytes :: Lens' ParquetSerDe (Maybe Natural) Source #
The maximum amount of padding to apply. This is useful if you intend to copy the data from Amazon S3 to HDFS before querying. The default is 0.
parquetSerDe_pageSizeBytes :: Lens' ParquetSerDe (Maybe Natural) Source #
The Parquet page size. Column chunks are divided into pages. A page is conceptually an indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and the default is 1 MiB.
parquetSerDe_writerVersion :: Lens' ParquetSerDe (Maybe ParquetWriterVersion) Source #
Indicates the version of row format to output. The possible values are
V1
and V2
. The default is V1
.