public class ParquetRecordReader extends CommonParquetRecordReader
CommonParquetRecordReader.Metric
footer, fragmentContext, NUM_RECORDS_TO_READ_NOT_SPECIFIED, operatorContext, parquetReaderStats
DEFAULT_TEXT_COLS_TO_READ
ALLOCATOR_INITIAL_RESERVATION, ALLOCATOR_MAX_RESERVATION
Constructor and Description |
---|
ParquetRecordReader(FragmentContext fragmentContext,
long numRecordsToRead,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.compression.CompressionCodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.compression.CompressionCodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
ParquetRecordReader(FragmentContext fragmentContext,
org.apache.hadoop.fs.Path path,
int rowGroupIndex,
long numRecordsToRead,
org.apache.hadoop.fs.FileSystem fs,
org.apache.parquet.compression.CompressionCodecFactory codecFactory,
org.apache.parquet.hadoop.metadata.ParquetMetadata footer,
List<SchemaPath> columns,
ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) |
Modifier and Type | Method and Description |
---|---|
void |
allocate(Map<String,ValueVector> vectorMap) |
void |
close() |
RecordBatchSizerManager |
getBatchSizesMgr() |
org.apache.parquet.compression.CompressionCodecFactory |
getCodecFactory() |
ParquetReaderUtility.DateCorruptionStatus |
getDateCorruptionStatus()
Flag indicating if the old non-standard data format appears
in this file, see DRILL-4203.
|
protected List<SchemaPath> |
getDefaultColumnsToRead() |
org.apache.hadoop.fs.FileSystem |
getFileSystem() |
FragmentContext |
getFragmentContext() |
org.apache.hadoop.fs.Path |
getHadoopPath() |
OperatorContext |
getOperatorContext() |
ReadState |
getReadState() |
int |
getRowGroupIndex() |
int |
next()
Read the next record batch from the file using the reader and read state
created previously.
|
void |
setup(OperatorContext operatorContext,
OutputMutator output)
Prepare the Parquet reader.
|
String |
toString() |
boolean |
useBulkReader() |
closeStats, handleAndRaise, initNumRecordsToRead, updateRowGroupsStats
getColumns, hasNext, isSkipQuery, isStarQuery, setColumns, transformColumns
public ParquetRecordReader(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, long numRecordsToRead, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetRecordReader(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetRecordReader(FragmentContext fragmentContext, long numRecordsToRead, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus)
public ParquetReaderUtility.DateCorruptionStatus getDateCorruptionStatus()
public org.apache.parquet.compression.CompressionCodecFactory getCodecFactory()
public org.apache.hadoop.fs.Path getHadoopPath()
public org.apache.hadoop.fs.FileSystem getFileSystem()
public int getRowGroupIndex()
public RecordBatchSizerManager getBatchSizesMgr()
public OperatorContext getOperatorContext()
public FragmentContext getFragmentContext()
public boolean useBulkReader()
public ReadState getReadState()
public void setup(OperatorContext operatorContext, OutputMutator output) throws ExecutionSetupException
operatorContext
- operator context for the readeroutput
- The place where output for a particular scan should be written. The record reader is responsible for
mutating the set of schema values for that particular record.ExecutionSetupException
public void allocate(Map<String,ValueVector> vectorMap) throws OutOfMemoryException
allocate
in interface RecordReader
allocate
in class AbstractRecordReader
OutOfMemoryException
public int next()
public void close()
protected List<SchemaPath> getDefaultColumnsToRead()
getDefaultColumnsToRead
in class AbstractRecordReader
public String toString()
toString
in class AbstractRecordReader
Copyright © 1970 The Apache Software Foundation. All rights reserved.