public abstract class ColumnReader<V extends ValueVector> extends Object
Modifier and Type | Field and Description |
---|---|
static Set<org.apache.parquet.column.Encoding> |
DICTIONARY_ENCODINGS |
static Set<org.apache.parquet.column.Encoding> |
VALUE_ENCODINGS |
protected DrillBuf |
vectorData |
Modifier | Constructor and Description |
---|---|
protected |
ColumnReader(ParquetRecordReader parentReader,
org.apache.parquet.column.ColumnDescriptor descriptor,
org.apache.parquet.hadoop.metadata.ColumnChunkMetaData columnChunkMetaData,
boolean fixedLength,
V v,
org.apache.parquet.format.SchemaElement schemaElement) |
Modifier and Type | Method and Description |
---|---|
int |
capacity() |
protected boolean |
checkVectorCapacityReached() |
void |
clear() |
boolean |
determineSize(long recordsReadInCurrentPass)
Determines the size of a single value in a variable column.
|
int |
getRecordsReadInCurrentPass() |
protected void |
hitRowGroupEnd() |
protected void |
postPageRead() |
protected boolean |
processPageData(int recordsToReadInThisPass) |
void |
processPages(long recordsToReadInThisPass) |
Future<Long> |
processPagesAsync(long recordsToReadInThisPass) |
protected abstract void |
readField(long recordsToRead) |
static int |
readIntLittleEndian(DrillBuf in,
int offset)
This is copied out of Parquet library, didn't want to deal with the
unnecessary throws statement they had declared
|
boolean |
readPage()
Read a page.
|
Future<Boolean> |
readPageAsync() |
protected void |
readRecords(int recordsToRead) |
protected Future<Integer> |
readRecordsAsync(int recordsToRead) |
protected int |
readRecordsInBulk(int recordsToReadInThisPass) |
void |
readValues(long recordsToRead) |
protected boolean |
recordsRequireDecoding() |
void |
reset() |
protected int |
totalValuesReadAndReadyToReadInPage() |
void |
updatePosition() |
void |
updateReadyToReadPosition() |
public static final Set<org.apache.parquet.column.Encoding> DICTIONARY_ENCODINGS
public static final Set<org.apache.parquet.column.Encoding> VALUE_ENCODINGS
protected DrillBuf vectorData
protected ColumnReader(ParquetRecordReader parentReader, org.apache.parquet.column.ColumnDescriptor descriptor, org.apache.parquet.hadoop.metadata.ColumnChunkMetaData columnChunkMetaData, boolean fixedLength, V v, org.apache.parquet.format.SchemaElement schemaElement) throws ExecutionSetupException
ExecutionSetupException
public int getRecordsReadInCurrentPass()
public void processPages(long recordsToReadInThisPass) throws IOException
IOException
public void clear()
public void readValues(long recordsToRead)
protected abstract void readField(long recordsToRead)
public boolean determineSize(long recordsReadInCurrentPass) throws IOException
recordsReadInCurrentPass
- records read in current passIOException
protected void readRecords(int recordsToRead)
protected int readRecordsInBulk(int recordsToReadInThisPass) throws IOException
IOException
protected boolean recordsRequireDecoding()
protected boolean processPageData(int recordsToReadInThisPass) throws IOException
IOException
public void updatePosition()
public void updateReadyToReadPosition()
public void reset()
public int capacity()
public boolean readPage() throws IOException
IOException
protected int totalValuesReadAndReadyToReadInPage()
protected void postPageRead()
protected void hitRowGroupEnd()
protected boolean checkVectorCapacityReached()
public static int readIntLittleEndian(DrillBuf in, int offset)
in
- incoming dataoffset
- offsetCopyright © 1970 The Apache Software Foundation. All rights reserved.