Skip to content

ParquetReadSupport

ParquetReadSupport is a concrete ReadSupport (from Apache Parquet) of UnsafeRows.

ParquetReadSupport is <> exclusively when ParquetFileFormat is requested for a data reader (with no support for Vectorized Parquet Decoding and so falling back to parquet-mr).

[[parquet.read.support.class]] ParquetReadSupport is registered as the fully-qualified class name for parquet.read.support.class Hadoop configuration when ParquetFileFormat is requested for a data reader.

[[creating-instance]] [[convertTz]] ParquetReadSupport takes an optional Java TimeZone to be created.

[[logging]] [TIP] ==== Enable ALL logging level for org.apache.spark.sql.execution.datasources.parquet.ParquetReadSupport logger to see what happens inside.

Add the following line to conf/log4j.properties:

log4j.logger.org.apache.spark.sql.execution.datasources.parquet.ParquetReadSupport=ALL

Refer to <>.

=== [[init]] Initializing ReadSupport -- init Method

[source, scala]

init(context: InitContext): ReadContext

NOTE: init is part of the ReadSupport Contract to...FIXME.

init...FIXME

=== [[prepareForRead]] prepareForRead Method

[source, scala]

prepareForRead( conf: Configuration, keyValueMetaData: JMap[String, String], fileSchema: MessageType, readContext: ReadContext): RecordMaterializer[UnsafeRow]


NOTE: prepareForRead is part of the ReadSupport Contract to...FIXME.

prepareForRead...FIXME


Last update: 2020-11-08
Back to top