Uses of Class
org.apache.drill.common.expression.SchemaPath
Package
Description
Logical expression tree representation.
Visitors for validating or optimizing logical expressions.
Logical operators in a query graph.
TODO - write docs for this package
Drill expression materialization and evaluation facilities.
Defines a revised implementation of the Drill RecordBatch protocol.
Provides run-time semantic analysis of the projection list for the
scan operator.
Provides the "version 3" scan framework (which can also be thought of
as EVF version 2).
Provides run-time semantic analysis of the projection list for the
scan operator.
Drill storage plugin.
Kafka storage plugin.
Defines a mock data source which generates dummy test data for use
in testing.
MongoDB storage plugin.
-
Uses of SchemaPath in org.apache.drill.common.expression
Modifier and TypeMethodDescriptionstatic SchemaPath
SchemaPath.create
(UserBitShared.NamePart namePart) SchemaPath.De.deserialize
(com.fasterxml.jackson.core.JsonParser jp, com.fasterxml.jackson.databind.DeserializationContext ctxt) SchemaPath.getChild
(int index) SchemaPath.getChild
(int index, Object originalValue, TypeProtos.MajorType valueType) SchemaPath.getChild
(String childPath, Object originalValue, TypeProtos.MajorType valueType) static SchemaPath
SchemaPath.getCompoundPath
(int n, String... path) ConstructsSchemaPath
based on givenpath
array up to nth element (inclusively).static SchemaPath
SchemaPath.getCompoundPath
(String... path) TypedFieldExpr.getPath()
static SchemaPath
SchemaPath.getSimplePath
(String name) SchemaPath.getUnIndexed()
Returns schema path with for arrays without index.static SchemaPath
SchemaPath.parseFromString
(String expr) Parses input string using the same rules which are used for the field in the query.Modifier and TypeMethodDescriptionboolean
SchemaPath.contains
(SchemaPath path) ExpressionStringBuilder.visitSchemaPath
(SchemaPath path, StringBuilder sb) ModifierConstructorDescriptionSchemaPath
(SchemaPath path) TypedFieldExpr
(SchemaPath path, TypeProtos.MajorType type) -
Uses of SchemaPath in org.apache.drill.common.expression.visitors
Modifier and TypeMethodDescriptionAbstractExprVisitor.visitSchemaPath
(SchemaPath path, VAL value) AggregateChecker.visitSchemaPath
(SchemaPath path, ErrorCollector errors) ExpressionValidator.visitSchemaPath
(SchemaPath path, ErrorCollector errors) ExprVisitor.visitSchemaPath
(SchemaPath path, VAL value) abstract T
SimpleExprVisitor.visitSchemaPath
(SchemaPath path) SimpleExprVisitor.visitSchemaPath
(SchemaPath path, Void value) -
Uses of SchemaPath in org.apache.drill.common.logical.data
-
Uses of SchemaPath in org.apache.drill.exec.compile.sig
Modifier and TypeMethodDescriptionConstantExpressionIdentifier.visitSchemaPath
(SchemaPath path, IdentityHashMap<LogicalExpression, Object> value) -
Uses of SchemaPath in org.apache.drill.exec.expr
Modifier and TypeMethodDescriptionCloneVisitor.visitSchemaPath
(SchemaPath path, Void value) HashVisitor.visitSchemaPath
(SchemaPath path, Void value) ModifierConstructorDescriptionStatisticsProvider
(Map<SchemaPath, ColumnStatistics<?>> columnStatMap, long rowCount, UdfUtilities udfUtilities) -
Uses of SchemaPath in org.apache.drill.exec.expr.fn.interpreter
Modifier and TypeMethodDescriptionInterpreterEvaluator.EvalVisitor.visitSchemaPath
(SchemaPath path, Integer value) -
Uses of SchemaPath in org.apache.drill.exec.metastore.analyze
Modifier and TypeMethodDescriptionAnalyzeFileInfoProvider.getLocationField
(ColumnNamesOptions columnNamesOptions) AnalyzeInfoProvider.getLocationField
(ColumnNamesOptions columnNamesOptions) Provides schema path to field which will be used as a location for specific table data, for example, for file-based tables, it may be `fqn`.Modifier and TypeMethodDescriptionAnalyzeFileInfoProvider.getProjectionFields
(DrillTable table, MetadataType metadataLevel, ColumnNamesOptions columnNamesOptions) AnalyzeInfoProvider.getProjectionFields
(DrillTable table, MetadataType metadataLevel, ColumnNamesOptions columnNamesOptions) Returns list of fields required for ANALYZE.AnalyzeParquetInfoProvider.getProjectionFields
(DrillTable table, MetadataType metadataLevel, ColumnNamesOptions columnNamesOptions) AnalyzeFileInfoProvider.getSegmentColumns
(DrillTable table, ColumnNamesOptions columnNamesOptions) AnalyzeInfoProvider.getSegmentColumns
(DrillTable table, ColumnNamesOptions columnNamesOptions) Returns list of segment column names for specifiedDrillTable
table.MetadataAggregateContext.interestingColumns()
MetadataControllerContext.interestingColumns()
MetadataAggregateContext.metadataColumns()
Modifier and TypeMethodDescriptionAnalyzeFileInfoProvider.getParentLocationExpression
(SchemaPath locationField) AnalyzeInfoProvider.getParentLocationExpression
(SchemaPath locationField) Returns expression which may be used to determine parent location for specific table data, i.e.Modifier and TypeMethodDescriptionAnalyzeFileInfoProvider.getMetadataInfoCollector
(BasicTablesRequests basicRequests, TableInfo tableInfo, FormatSelection selection, PlannerSettings settings, Supplier<org.apache.calcite.rel.core.TableScan> tableScanSupplier, List<SchemaPath> interestingColumns, MetadataType metadataLevel, int segmentColumnsCount) AnalyzeInfoProvider.getMetadataInfoCollector
(BasicTablesRequests basicRequests, TableInfo tableInfo, FormatSelection selection, PlannerSettings settings, Supplier<org.apache.calcite.rel.core.TableScan> tableScanSupplier, List<SchemaPath> interestingColumns, MetadataType metadataLevel, int segmentColumnsCount) ReturnsMetadataInfoCollector
instance for obtaining information about segments, files, etc.MetadataAggregateContext.MetadataAggregateContextBuilder.interestingColumns
(List<SchemaPath> interestingColumns) MetadataControllerContext.MetadataControllerContextBuilder.interestingColumns
(List<SchemaPath> interestingColumns) MetadataAggregateContext.MetadataAggregateContextBuilder.metadataColumns
(List<SchemaPath> metadataColumns) ModifierConstructorDescriptionFileMetadataInfoCollector
(BasicTablesRequests basicRequests, TableInfo tableInfo, FormatSelection selection, PlannerSettings settings, Supplier<org.apache.calcite.rel.core.TableScan> tableScanSupplier, List<SchemaPath> interestingColumns, MetadataType metadataLevel, int segmentColumnsCount) -
Uses of SchemaPath in org.apache.drill.exec.metastore.store
Modifier and TypeMethodDescriptionMetastoreFileTableMetadataProvider.getPartitionColumns()
SimpleFileTableMetadataProvider.getPartitionColumns()
Modifier and TypeMethodDescriptionMetastoreFileTableMetadataProvider.getPartitionMetadata
(SchemaPath columnName) SimpleFileTableMetadataProvider.getPartitionMetadata
(SchemaPath columnName) -
Uses of SchemaPath in org.apache.drill.exec.physical.base
Modifier and TypeFieldDescriptionstatic final List<SchemaPath>
GroupScan.ALL_COLUMNS
columns list in GroupScan : 1) empty_column is for skipAll query.protected List<SchemaPath>
AbstractGroupScanWithMetadata.columns
protected List<SchemaPath>
AbstractGroupScanWithMetadata.partitionColumns
Modifier and TypeMethodDescriptionAbstractDbGroupScan.getRowKeyPath()
DbGroupScan.getRowKeyPath()
Modifier and TypeMethodDescriptionAbstractGroupScan.getColumns()
AbstractGroupScanWithMetadata.getColumns()
DbGroupScan.getColumns()
GroupScan.getColumns()
Returns a list of columns scanned by this group scanIndexGroupScan.getColumns()
protected <T extends Metadata>
Map<SchemaPath,ColumnStatistics<?>> AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.getImplicitColumnStatistics
(OptionManager optionManager, T metadata, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics) AbstractGroupScan.getPartitionColumns()
AbstractGroupScanWithMetadata.getPartitionColumns()
GroupScan.getPartitionColumns()
Returns a list of columns that can be used for partition pruningModifier and TypeMethodDescriptionlong
AbstractGroupScan.getColumnValueCount
(SchemaPath column) By default, throw exception, since group scan does not have exact column value count.long
AbstractGroupScanWithMetadata.getColumnValueCount
(SchemaPath column) Return column value count for the specified column.long
GroupScan.getColumnValueCount
(SchemaPath column) Return the number of non-null value in the specified column.<T> T
AbstractGroupScanWithMetadata.getPartitionValue
(org.apache.hadoop.fs.Path path, SchemaPath column, Class<T> clazz) AbstractGroupScanWithMetadata.getTypeForColumn
(SchemaPath schemaPath) static boolean
AbstractGroupScanWithMetadata.isImplicitOrPartCol
(SchemaPath schemaPath, OptionManager optionManager) Modifier and TypeMethodDescriptionboolean
AbstractGroupScan.canPushdownProjects
(List<SchemaPath> columns) boolean
GroupScan.canPushdownProjects
(List<SchemaPath> columns) GroupScan should check the list of columns, and see if it could support all the columns in the list.boolean
SchemalessScan.canPushdownProjects
(List<SchemaPath> columns) AbstractGroupScan.clone
(List<SchemaPath> columns) GroupScan.clone
(List<SchemaPath> columns) Returns a clone of GroupScan instance, except that the new GroupScan will use the provided list of columns .SchemalessScan.clone
(List<SchemaPath> columns) AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.filterAndGetMetadata
(Set<SchemaPath> schemaPathsInExpr, Iterable<T> metadataList, FilterPredicate<?> filterPredicate, OptionManager optionManager) Produces filtering of specified metadata using specified filter expression and returns filtered metadata.protected void
AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.filterFileMetadata
(OptionManager optionManager, FilterPredicate<?> filterPredicate, Set<SchemaPath> schemaPathsInExpr) Produces filtering of metadata at file level.protected void
AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.filterPartitionMetadata
(OptionManager optionManager, FilterPredicate<?> filterPredicate, Set<SchemaPath> schemaPathsInExpr) Produces filtering of metadata at partition level.protected void
AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.filterSegmentMetadata
(OptionManager optionManager, FilterPredicate<?> filterPredicate, Set<SchemaPath> schemaPathsInExpr) Produces filtering of metadata at segment level.protected void
AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.filterTableMetadata
(FilterPredicate<?> filterPredicate, Set<SchemaPath> schemaPathsInExpr) Produces filtering of metadata at table level.protected <T extends Metadata>
Map<SchemaPath,ColumnStatistics<?>> AbstractGroupScanWithMetadata.GroupScanWithMetadataFilterer.getImplicitColumnStatistics
(OptionManager optionManager, T metadata, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics) AbstractDbGroupScan.getRestrictedScan
(List<SchemaPath> columns) DbGroupScan.getRestrictedScan
(List<SchemaPath> columns) If this DbGroupScan supports restricted scan, create a restricted scan from this DbGroupScan.void
IndexGroupScan.setColumns
(List<SchemaPath> columns) ModifierConstructorDescriptionprotected
AbstractGroupScanWithMetadata
(String userName, List<SchemaPath> columns, LogicalExpression filter) SchemalessScan
(String userName, org.apache.hadoop.fs.Path selectionRoot, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.physical.config
ModifierConstructorDescriptionFlattenPOP
(PhysicalOperator child, SchemaPath column) UnnestPOP
(PhysicalOperator child, SchemaPath column, String implicitColumn) ModifierConstructorDescriptionLateralJoinPOP
(PhysicalOperator left, PhysicalOperator right, org.apache.calcite.rel.core.JoinRelType joinType, String implicitRIDColumn, List<SchemaPath> excludedColumns) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl
-
Uses of SchemaPath in org.apache.drill.exec.physical.impl.aggregate
-
Uses of SchemaPath in org.apache.drill.exec.physical.impl.orderedpartitioner
Modifier and TypeMethodDescriptionvoid
OrderedPartitionProjector.setup
(FragmentContext context, VectorAccessible incoming, RecordBatch outgoing, List<TransferPair> transfers, VectorContainer partitionVectors, int partitions, SchemaPath outputField) final void
OrderedPartitionProjectorTemplate.setup
(FragmentContext context, VectorAccessible incoming, RecordBatch outgoing, List<TransferPair> transfers, VectorContainer partitionVectors, int partitions, SchemaPath outputField) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.partitionsender
Modifier and TypeMethodDescriptionPartitionerTemplate.OutgoingRecordBatch.getValueVectorId
(SchemaPath path) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.protocol
Modifier and TypeMethodDescriptionBatchAccessor.getValueVectorId
(SchemaPath path) OperatorRecordBatch.getValueVectorId
(SchemaPath path) VectorContainerAccessor.getValueVectorId
(SchemaPath path) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.scan.project
Modifier and TypeFieldDescriptionfinal List<SchemaPath>
ScanSchemaOrchestrator.ScanSchemaOptions.projection
protected final List<SchemaPath>
ScanLevelProjection.projectionList
Modifier and TypeMethodDescriptionScanLevelProjection.Builder.projectionList()
ScanLevelProjection.requestedCols()
Return the set of columns from the SELECT listModifier and TypeMethodDescriptionstatic ScanLevelProjection
ScanLevelProjection.build
(List<SchemaPath> projectionList, List<ScanLevelProjection.ScanProjectionParser> parsers) Builder shortcut, primarily for tests.static ScanLevelProjection
ScanLevelProjection.build
(List<SchemaPath> projectionList, List<ScanLevelProjection.ScanProjectionParser> parsers, TupleMetadata outputSchema) Builder shortcut, primarily for tests.ScanLevelProjection.Builder.projection
(List<SchemaPath> projectionList) Specify the set of columns in the SELECT list.void
ScanSchemaOrchestrator.ScanOrchestratorBuilder.projection
(List<SchemaPath> projection) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.scan.v3
Modifier and TypeMethodDescriptionvoid
ScanLifecycleBuilder.projection
(List<SchemaPath> projection) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.scan.v3.schema
Modifier and TypeMethodDescriptionScanProjectionParser.parse
(Collection<SchemaPath> projList) ScanSchemaConfigBuilder.projection
(List<SchemaPath> projectionList) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.unorderedreceiver
-
Uses of SchemaPath in org.apache.drill.exec.physical.impl.validate
Modifier and TypeMethodDescriptionIteratorValidatorBatchIterator.getValueVectorId
(SchemaPath path) -
Uses of SchemaPath in org.apache.drill.exec.physical.impl.window
-
Uses of SchemaPath in org.apache.drill.exec.physical.impl.xsort
-
Uses of SchemaPath in org.apache.drill.exec.physical.resultSet.project
Modifier and TypeMethodDescriptionstatic RequestedTuple
Projections.parse
(Collection<SchemaPath> projList) Parse a projection list. -
Uses of SchemaPath in org.apache.drill.exec.planner
Modifier and TypeMethodDescriptionFileSystemPartitionDescriptor.getVectorType
(SchemaPath column, PlannerSettings plannerSettings) ParquetPartitionDescriptor.getVectorType
(SchemaPath column, PlannerSettings plannerSettings) PartitionDescriptor.getVectorType
(SchemaPath column, PlannerSettings plannerSettings) Method returns the Major type associated with the given column -
Uses of SchemaPath in org.apache.drill.exec.planner.common
Modifier and TypeMethodDescriptionDrillStatsTable.getColumns()
DrillRelOptUtil.ProjectPushInfo.getFields()
Modifier and TypeMethodDescriptionstatic List<StatisticsHolder<?>>
DrillStatsTable.getEstimatedColumnStats
(DrillStatsTable statsProvider, SchemaPath fieldName) Returns list ofStatisticsKind
and statistics values obtained from specifiedDrillStatsTable
for specified column.DrillStatsTable.getHistogram
(SchemaPath column) Get the histogram of a given column.DrillStatsTable.getNdv
(SchemaPath col) Get the approximate number of distinct values of given column.DrillStatsTable.getNNRowCount
(SchemaPath col) Get non-null rowcount for the column If stats are not present for the given column, a null is returned.void
DrillStatsTable.ColumnStatistics_v1.setName
(SchemaPath name) ModifierConstructorDescriptionDrillScanRelBase
(org.apache.calcite.plan.RelOptCluster cluster, org.apache.calcite.plan.RelTraitSet traits, org.apache.calcite.plan.RelOptTable table, List<SchemaPath> columns) ProjectPushInfo
(List<SchemaPath> fields, Map<String, FieldsReWriterUtil.DesiredField> desiredFields) -
Uses of SchemaPath in org.apache.drill.exec.planner.index
Modifier and TypeMethodDescriptionFunctionalIndexInfo.getNewPath
(SchemaPath path) getNewPath: for an original path, return new rename '$N' path, notice there could be multiple renamed paths if the there are multiple functional indexes refer original path.FunctionalIndexInfo.getNewPathFromExpr
(LogicalExpression expr) return a plain field path if the incoming index expression 'expr' is replaced to be a plain fieldModifier and TypeMethodDescriptionFunctionalIndexInfo.allNewSchemaPaths()
FunctionalIndexInfo.allPathsInFunction()
FunctionalIndexInfo.getPathsInFunctionExpr()
IndexCallContext.getScanColumns()
IndexLogicalPlanCallContext.getScanColumns()
IndexPhysicalPlanCallContext.getScanColumns()
static List<SchemaPath>
IndexPlanUtils.rewriteFunctionColumn
(List<SchemaPath> paths, FunctionalIndexInfo functionInfo, List<SchemaPath> addedPaths) For IndexGroupScan, if a column is only appeared in the should-be-renamed function, this column is to-be-replaced column, we replace that column(schemaPath) from 'a.b' to '$1' in the list of SchemaPath.Modifier and TypeMethodDescriptionFunctionalIndexInfo.getNewPath
(SchemaPath path) getNewPath: for an original path, return new rename '$N' path, notice there could be multiple renamed paths if the there are multiple functional indexes refer original path.boolean
AbstractIndexCollection.isColumnIndexed
(SchemaPath path) boolean
IndexCollection.isColumnIndexed
(SchemaPath path) Check if the field name is the leading key of any of the indexes in this collectionboolean
DrillIndexDefinition.pathExactIn
(SchemaPath path, Collection<LogicalExpression> exprs) static boolean
IndexPlanUtils.pathOnlyInIndexedFunction
(SchemaPath path) org.apache.calcite.rex.RexNode
ExprToRex.visitSchemaPath
(SchemaPath path, Void value) PathInExpr.visitSchemaPath
(SchemaPath path, Void value) Modifier and TypeMethodDescriptionstatic org.apache.calcite.rel.type.RelDataType
FunctionalIndexHelper.rewriteFunctionalRowType
(org.apache.calcite.rel.RelNode origScan, IndexCallContext indexContext, FunctionalIndexInfo functionInfo, Collection<SchemaPath> addedPaths) if a field in rowType serves only the to-be-replaced column(s), we should replace it with new name "$1", otherwise we should keep this dataTypeField and add a new one for "$1"static List<SchemaPath>
IndexPlanUtils.rewriteFunctionColumn
(List<SchemaPath> paths, FunctionalIndexInfo functionInfo, List<SchemaPath> addedPaths) For IndexGroupScan, if a column is only appeared in the should-be-renamed function, this column is to-be-replaced column, we replace that column(schemaPath) from 'a.b' to '$1' in the list of SchemaPath. -
Uses of SchemaPath in org.apache.drill.exec.planner.index.generators
Modifier and TypeMethodDescriptionprotected boolean
AbstractIndexPlanGenerator.checkRowKey
(List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.planner.logical
Modifier and TypeMethodDescriptionDrillScanRel.getColumns()
static List<SchemaPath>
DrillScanRel.getProjectedColumns
(org.apache.calcite.plan.RelOptTable table, boolean isSelectStar) Modifier and TypeMethodDescriptionvoid
ScanFieldDeterminer.FieldList.addProjected
(SchemaPath path) void
ScanFieldDeterminer.FieldList.addReferenced
(SchemaPath path) Modifier and TypeMethodDescriptionvoid
ScanFieldDeterminer.FieldList.addProjected
(Collection<SchemaPath> paths) void
ScanFieldDeterminer.FieldList.addReferenced
(Collection<SchemaPath> paths) ModifierConstructorDescriptionDrillScanRel
(org.apache.calcite.plan.RelOptCluster cluster, org.apache.calcite.plan.RelTraitSet traits, org.apache.calcite.plan.RelOptTable table, org.apache.calcite.rel.type.RelDataType rowType, List<SchemaPath> columns) DrillScanRel
(org.apache.calcite.plan.RelOptCluster cluster, org.apache.calcite.plan.RelTraitSet traits, org.apache.calcite.plan.RelOptTable table, org.apache.calcite.rel.type.RelDataType rowType, List<SchemaPath> columns, boolean partitionFilterPushdown) DrillScanRel
(org.apache.calcite.plan.RelOptCluster cluster, org.apache.calcite.plan.RelTraitSet traits, org.apache.calcite.plan.RelOptTable table, GroupScan groupScan, org.apache.calcite.rel.type.RelDataType rowType, List<SchemaPath> columns) Creates a DrillScanRel for a particular GroupScanDrillScanRel
(org.apache.calcite.plan.RelOptCluster cluster, org.apache.calcite.plan.RelTraitSet traits, org.apache.calcite.plan.RelOptTable table, GroupScan groupScan, org.apache.calcite.rel.type.RelDataType rowType, List<SchemaPath> columns, boolean partitionFilterPushdown) Creates a DrillScanRel for a particular GroupScan -
Uses of SchemaPath in org.apache.drill.exec.planner.sql
Modifier and TypeMethodDescriptionHivePartitionDescriptor.getVectorType
(SchemaPath column, PlannerSettings plannerSettings) -
Uses of SchemaPath in org.apache.drill.exec.planner.sql.parser
-
Uses of SchemaPath in org.apache.drill.exec.record
Modifier and TypeMethodDescriptionstatic List<SchemaPath>
SchemaUtil.getSchemaPaths
(TupleMetadata schema) Returns list ofSchemaPath
for fields taken from specified schema.Modifier and TypeMethodDescriptionHyperVectorWrapper.getFieldIdIfMatches
(int id, SchemaPath expectedPath) SimpleVectorWrapper.getFieldIdIfMatches
(int id, SchemaPath expectedPath) VectorWrapper.getFieldIdIfMatches
(int id, SchemaPath expectedPath) Traverse the object graph and determine whether the provided SchemaPath matches data within the Wrapper.AbstractRecordBatch.getValueVectorId
(SchemaPath path) RecordBatch.getValueVectorId
(SchemaPath path) Gets the value vector type and ID for the given schema path.RecordBatchLoader.getValueVectorId
(SchemaPath path) RecordIterator.getValueVectorId
(SchemaPath path) SchemalessBatch.getValueVectorId
(SchemaPath path) SimpleRecordBatch.getValueVectorId
(SchemaPath path) VectorAccessible.getValueVectorId
(SchemaPath path) Get the value vector type and id for the given schema path.VectorContainer.getValueVectorId
(SchemaPath path) -
Uses of SchemaPath in org.apache.drill.exec.store
Modifier and TypeFieldDescriptionprotected static final List<SchemaPath>
AbstractRecordReader.DEFAULT_TEXT_COLS_TO_READ
Modifier and TypeMethodDescriptionprotected Collection<SchemaPath>
AbstractRecordReader.getColumns()
protected List<SchemaPath>
AbstractRecordReader.getDefaultColumnsToRead()
ColumnExplorer.getTableColumns()
protected Collection<SchemaPath>
AbstractRecordReader.transformColumns
(Collection<SchemaPath> projected) Modifier and TypeMethodDescriptionstatic boolean
ColumnExplorer.isPartitionColumn
(OptionManager optionManager, SchemaPath column) Checks if given column is partition or not.Modifier and TypeMethodDescriptionAbstractStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) AbstractStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options) AbstractStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager metadataProviderManager) StoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) Get the physical scan operator for the particular GroupScan (read) node.StoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options) Get the physical scan operator for the particular GroupScan (read) node.StoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager providerManager) Get the physical scan operator for the particular GroupScan (read) node.protected final void
AbstractRecordReader.setColumns
(Collection<SchemaPath> projected) protected Collection<SchemaPath>
AbstractRecordReader.transformColumns
(Collection<SchemaPath> projected) ModifierConstructorDescriptionColumnExplorer
(OptionManager optionManager, List<SchemaPath> columns) Helper class that encapsulates logic for sorting out columns between actual table columns, partition columns and implicit file columns. -
Uses of SchemaPath in org.apache.drill.exec.store.bson
ModifierConstructorDescriptionBsonRecordReader
(DrillBuf managedBuf, List<SchemaPath> columns, boolean readNumbersAsDouble) -
Uses of SchemaPath in org.apache.drill.exec.store.delta
Modifier and TypeMethodDescriptionboolean
DeltaRowGroupScan.isImplicitColumn
(SchemaPath path, String partitionColumnLabel) Modifier and TypeMethodDescriptionDeltaGroupScan.clone
(List<SchemaPath> columns) DeltaGroupScan.DeltaGroupScanBuilder.columns
(List<SchemaPath> columns) DeltaRowGroupScan.copy
(List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.delta.format
Modifier and TypeMethodDescriptionDeltaFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns) DeltaFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, MetadataProviderManager metadataProviderManager) DeltaFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options) DeltaFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options, MetadataProviderManager metadataProviderManager) -
Uses of SchemaPath in org.apache.drill.exec.store.dfs
Modifier and TypeMethodDescriptionFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns) default AbstractGroupScan
FormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, MetadataProviderManager metadataProviderManager) default AbstractGroupScan
FormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options) default AbstractGroupScan
FormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options, MetadataProviderManager metadataProvider) FileSystemPlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) FileSystemPlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager metadataProviderManager) -
Uses of SchemaPath in org.apache.drill.exec.store.dfs.easy
Modifier and TypeMethodDescriptionboolean
EasyGroupScan.canPushdownProjects
(List<SchemaPath> columns) EasyGroupScan.clone
(List<SchemaPath> columns) EasyFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns) EasyFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, MetadataProviderManager metadataProviderManager) EasyFormatPlugin.getRecordReader
(FragmentContext context, DrillFileSystem dfs, FileWork fileWork, List<SchemaPath> columns, String userName) Return a record reader for the specific file format, when using the originalScanBatch
scanner.ModifierConstructorDescriptionEasyGroupScan
(String userName, FileSelection selection, EasyFormatPlugin<?> formatPlugin, List<SchemaPath> columns, org.apache.hadoop.fs.Path selectionRoot, int minWidth, MetadataProviderManager metadataProvider) EasyGroupScan
(String userName, FileSelection selection, EasyFormatPlugin<?> formatPlugin, List<SchemaPath> columns, org.apache.hadoop.fs.Path selectionRoot, MetadataProviderManager metadataProviderManager) -
Uses of SchemaPath in org.apache.drill.exec.store.direct
Modifier and TypeMethodDescriptionDirectGroupScan.clone
(List<SchemaPath> columns) MetadataDirectGroupScan.clone
(List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.drill.plugin
-
Uses of SchemaPath in org.apache.drill.exec.store.druid
Modifier and TypeMethodDescriptionDruidGroupScan.getColumns()
DruidSubScan.getColumns()
protected Collection<SchemaPath>
DruidRecordReader.transformColumns
(Collection<SchemaPath> projectedColumns) Modifier and TypeMethodDescriptionDruidCompareFunctionProcessor.visitSchemaPath
(SchemaPath path, LogicalExpression valueArg) Modifier and TypeMethodDescriptionboolean
DruidGroupScan.canPushdownProjects
(List<SchemaPath> columns) DruidGroupScan.clone
(List<SchemaPath> columns) protected Collection<SchemaPath>
DruidRecordReader.transformColumns
(Collection<SchemaPath> projectedColumns) ModifierConstructorDescriptionDruidGroupScan
(String userName, DruidScanSpec scanSpec, DruidStoragePluginConfig storagePluginConfig, List<SchemaPath> columns, int maxRecordsToRead, StoragePluginRegistry pluginRegistry) DruidGroupScan
(String userName, DruidStoragePlugin storagePlugin, DruidScanSpec scanSpec, List<SchemaPath> columns, int maxRecordsToRead) DruidRecordReader
(DruidSubScan.DruidSubScanSpec subScanSpec, List<SchemaPath> projectedColumns, int maxRecordsToRead, FragmentContext context, DruidStoragePlugin plugin) DruidSubScan
(StoragePluginRegistry registry, String userName, StoragePluginConfig config, LinkedList<DruidSubScan.DruidSubScanSpec> datasourceScanSpecList, List<SchemaPath> columns, int maxRecordsToRead) -
Uses of SchemaPath in org.apache.drill.exec.store.easy.json
Modifier and TypeMethodDescriptionprotected List<SchemaPath>
JSONRecordReader.getDefaultColumnsToRead()
Deprecated.Modifier and TypeMethodDescriptionJSONFormatPlugin.getRecordReader
(FragmentContext context, DrillFileSystem dfs, FileWork fileWork, List<SchemaPath> columns, String userName) ModifierConstructorDescriptionJSONRecordReader
(FragmentContext fragmentContext, com.fasterxml.jackson.databind.JsonNode embeddedContent, DrillFileSystem fileSystem, List<SchemaPath> columns) Deprecated.Create a new JSON Record Reader that uses an in memory materialized JSON stream.JSONRecordReader
(FragmentContext fragmentContext, List<SchemaPath> columns) Deprecated.Create a JSON Record Reader that uses an InputStream directlyJSONRecordReader
(FragmentContext fragmentContext, org.apache.hadoop.fs.Path inputPath, DrillFileSystem fileSystem, List<SchemaPath> columns, JSONFormatConfig config) Deprecated.Create a JSON Record Reader that uses a file based input stream. -
Uses of SchemaPath in org.apache.drill.exec.store.easy.text
Modifier and TypeMethodDescriptionTextFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, MetadataProviderManager metadataProviderManager) TextFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options, MetadataProviderManager metadataProviderManager) -
Uses of SchemaPath in org.apache.drill.exec.store.enumerable
Modifier and TypeMethodDescriptionEnumerableGroupScan.getColumns()
EnumerableSubScan.getColumns()
ModifierConstructorDescriptionEnumerableGroupScan
(String code, List<SchemaPath> columns, Map<String, Integer> fieldsMap, double rows, TupleMetadata schema, String schemaPath, ColumnConverterFactoryProvider converterFactoryProvider) EnumerableRecordReader
(List<SchemaPath> columns, Map<String, Integer> fieldsMap, String code, String schemaPath, ColumnConverterFactoryProvider factoryProvider) EnumerableSubScan
(String code, List<SchemaPath> columns, Map<String, Integer> fieldsMap, TupleMetadata schema, String schemaPath, ColumnConverterFactoryProvider converterFactoryProvider) -
Uses of SchemaPath in org.apache.drill.exec.store.googlesheets
Modifier and TypeMethodDescriptionGoogleSheetsGroupScan.columns()
GoogleSheetsSubScan.getColumns()
Modifier and TypeMethodDescriptionboolean
GoogleSheetsGroupScan.canPushdownProjects
(List<SchemaPath> columns) GoogleSheetsGroupScan.clone
(List<SchemaPath> columns) GoogleSheetsStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) GoogleSheetsStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager metadataProviderManager) ModifierConstructorDescriptionGoogleSheetsGroupScan
(String userName, GoogleSheetsScanSpec scanSpec, List<SchemaPath> columns, Map<String, ExprNode.ColRelOpConstNode> filters, double selectivity, int maxRecords, StoragePluginRegistry plugins) GoogleSheetsGroupScan
(GoogleSheetsGroupScan that, List<SchemaPath> columns) Constructor for applying columns (Projection pushdown).GoogleSheetsSubScan
(String username, GoogleSheetsStoragePluginConfig config, GoogleSheetsScanSpec scanSpec, List<SchemaPath> columns, Map<String, ExprNode.ColRelOpConstNode> filters, int maxRecords, TupleMetadata schema) -
Uses of SchemaPath in org.apache.drill.exec.store.googlesheets.utils
Modifier and TypeMethodDescriptionstatic boolean
GoogleSheetsUtils.isProjected
(List<SchemaPath> projectedColumns, String columnName) Returns true if the column is projected, false if not. -
Uses of SchemaPath in org.apache.drill.exec.store.hbase
Modifier and TypeMethodDescriptionHBaseGroupScan.getColumns()
HBaseSubScan.getColumns()
protected Collection<SchemaPath>
HBaseRecordReader.transformColumns
(Collection<SchemaPath> columns) Provides the projected columns information to the Hbase Scan instance.Modifier and TypeMethodDescriptionprotected void
CompareFunctionsProcessor.setPath
(SchemaPath path) CompareFunctionsProcessor.visitSchemaPath
(SchemaPath path, LogicalExpression valueArg) Modifier and TypeMethodDescriptionboolean
HBaseGroupScan.canPushdownProjects
(List<SchemaPath> columns) HBaseGroupScan.clone
(List<SchemaPath> columns) protected Collection<SchemaPath>
HBaseRecordReader.transformColumns
(Collection<SchemaPath> columns) Provides the projected columns information to the Hbase Scan instance.static void
HBaseUtils.verifyColumns
(List<SchemaPath> columns, org.apache.hadoop.hbase.HTableDescriptor hTableDesc) Verify the presence of a column family in the schema path of the hbase table or whether the schema path is the row key column.ModifierConstructorDescriptionHBaseGroupScan
(String userName, HBaseScanSpec hbaseScanSpec, HBaseStoragePluginConfig storagePluginConfig, List<SchemaPath> columns, StoragePluginRegistry pluginRegistry) HBaseGroupScan
(String userName, HBaseStoragePlugin storagePlugin, HBaseScanSpec scanSpec, List<SchemaPath> columns) HBaseRecordReader
(org.apache.hadoop.hbase.client.Connection connection, HBaseSubScan.HBaseSubScanSpec subScanSpec, List<SchemaPath> projectedColumns) HBaseSubScan
(StoragePluginRegistry registry, String userName, HBaseStoragePluginConfig hbaseStoragePluginConfig, LinkedList<HBaseSubScan.HBaseSubScanSpec> regionScanSpecList, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.hive
Modifier and TypeMethodDescriptionboolean
HiveScan.canPushdownProjects
(List<SchemaPath> columns) HiveDrillNativeParquetScan.clone
(List<SchemaPath> columns) HiveScan.clone
(List<SchemaPath> columns) HiveDrillNativeParquetRowGroupScan.copy
(List<SchemaPath> columns) HiveStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) HiveStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options) ModifierConstructorDescriptionHiveDrillNativeParquetScan
(String userName, List<SchemaPath> columns, HiveStoragePlugin hiveStoragePlugin, List<HiveMetadataProvider.LogicalInputSplit> logicalInputSplits, Map<String, String> confProperties, ParquetReaderConfig readerConfig) HiveDrillNativeParquetScan
(String userName, List<SchemaPath> columns, HiveStoragePlugin hiveStoragePlugin, List<HiveMetadataProvider.LogicalInputSplit> logicalInputSplits, Map<String, String> confProperties, ParquetReaderConfig readerConfig, LogicalExpression filter) HiveDrillNativeParquetScan
(StoragePluginRegistry engineRegistry, String userName, HiveStoragePluginConfig hiveStoragePluginConfig, List<SchemaPath> columns, List<ReadEntryWithPath> entries, HivePartitionHolder hivePartitionHolder, Map<String, String> confProperties, ParquetReaderConfig readerConfig, LogicalExpression filter, TupleMetadata schema) HiveScan
(String userName, HiveReadEntry hiveReadEntry, HiveStoragePluginConfig hiveStoragePluginConfig, List<SchemaPath> columns, Map<String, String> confProperties, StoragePluginRegistry pluginRegistry) HiveScan
(String userName, HiveReadEntry hiveReadEntry, HiveStoragePlugin hiveStoragePlugin, List<SchemaPath> columns, HiveMetadataProvider metadataProvider, Map<String, String> confProperties) -
Uses of SchemaPath in org.apache.drill.exec.store.hive.readers
ModifierConstructorDescriptionHiveDefaultRecordReader
(HiveTableWithColumnCache table, HivePartition partition, Collection<org.apache.hadoop.mapred.InputSplit> inputSplits, List<SchemaPath> projectedColumns, FragmentContext context, org.apache.hadoop.hive.conf.HiveConf hiveConf, org.apache.hadoop.security.UserGroupInformation proxyUgi) Readers constructor called by initializer.HiveTextRecordReader
(HiveTableWithColumnCache table, HivePartition partition, Collection<org.apache.hadoop.mapred.InputSplit> inputSplits, List<SchemaPath> projectedColumns, FragmentContext context, org.apache.hadoop.hive.conf.HiveConf hiveConf, org.apache.hadoop.security.UserGroupInformation proxyUgi) Constructor matching super. -
Uses of SchemaPath in org.apache.drill.exec.store.http
Modifier and TypeMethodDescriptionboolean
HttpGroupScan.canPushdownProjects
(List<SchemaPath> columns) HttpGroupScan.clone
(List<SchemaPath> columns) HttpStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) HttpStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager metadataProviderManager) ModifierConstructorDescriptionHttpGroupScan
(List<SchemaPath> columns, HttpScanSpec httpScanSpec, Map<String, String> filters, double selectivity, int maxRecords) Deserialize a group scan.HttpGroupScan
(HttpGroupScan that, List<SchemaPath> columns) Applies columns.HttpSubScan
(HttpScanSpec tableSpec, List<SchemaPath> columns, Map<String, String> filters, int maxRecords, TupleMetadata schema) -
Uses of SchemaPath in org.apache.drill.exec.store.iceberg
Modifier and TypeMethodDescriptionIcebergGroupScan.clone
(List<SchemaPath> columns) IcebergGroupScan.IcebergGroupScanBuilder.columns
(List<SchemaPath> columns) IcebergSubScan.IcebergSubScanBuilder.columns
(List<SchemaPath> columns) static org.apache.iceberg.TableScan
IcebergGroupScan.projectColumns
(org.apache.iceberg.TableScan tableScan, List<SchemaPath> columns) ModifierConstructorDescriptionIcebergGroupScan
(String userName, StoragePluginConfig storageConfig, FormatPluginConfig formatConfig, List<SchemaPath> columns, TupleMetadata schema, String path, LogicalExpression condition, Integer maxRecords, StoragePluginRegistry pluginRegistry) IcebergSubScan
(String userName, StoragePluginConfig storageConfig, FormatPluginConfig formatConfig, List<SchemaPath> columns, String path, List<IcebergWork> workList, TupleMetadata schema, LogicalExpression condition, Integer maxRecords, StoragePluginRegistry pluginRegistry) -
Uses of SchemaPath in org.apache.drill.exec.store.iceberg.format
Modifier and TypeMethodDescriptionIcebergFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns) IcebergFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, MetadataProviderManager metadataProviderManager) -
Uses of SchemaPath in org.apache.drill.exec.store.ischema
Modifier and TypeMethodDescriptionInfoSchemaFilterBuilder.visitSchemaPath
(SchemaPath path, Void value) Modifier and TypeMethodDescriptionInfoSchemaGroupScan.clone
(List<SchemaPath> columns) InfoSchemaStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.jdbc
ModifierConstructorDescriptionJdbcBatchReader
(DataSource source, String sql, List<SchemaPath> columns) JdbcGroupScan
(String sql, List<SchemaPath> columns, JdbcStorageConfig config, double rows, String username, StoragePluginRegistry plugins) JdbcSubScan
(String sql, List<SchemaPath> columns, StoragePluginConfig config, String username, StoragePluginRegistry plugins) -
Uses of SchemaPath in org.apache.drill.exec.store.kafka
Modifier and TypeMethodDescriptionboolean
KafkaGroupScan.canPushdownProjects
(List<SchemaPath> columns) KafkaGroupScan.clone
(List<SchemaPath> columns) ModifierConstructorDescriptionKafkaGroupScan
(String userName, KafkaStoragePluginConfig kafkaStoragePluginConfig, List<SchemaPath> columns, int records, KafkaScanSpec scanSpec, StoragePluginRegistry pluginRegistry) KafkaGroupScan
(String userName, KafkaStoragePlugin kafkaStoragePlugin, List<SchemaPath> columns, int records, KafkaScanSpec kafkaScanSpec) KafkaGroupScan
(KafkaGroupScan that, List<SchemaPath> columns) KafkaGroupScan
(KafkaStoragePlugin kafkaStoragePlugin, KafkaScanSpec kafkaScanSpec, List<SchemaPath> columns, int records) KafkaSubScan
(String userName, KafkaStoragePlugin kafkaStoragePlugin, List<SchemaPath> columns, int records, List<KafkaPartitionScanSpec> partitionSubScanSpecList) KafkaSubScan
(StoragePluginRegistry registry, String userName, KafkaStoragePluginConfig kafkaStoragePluginConfig, List<SchemaPath> columns, int records, LinkedList<KafkaPartitionScanSpec> partitionSubScanSpecList) -
Uses of SchemaPath in org.apache.drill.exec.store.kudu
Modifier and TypeMethodDescriptionboolean
KuduGroupScan.canPushdownProjects
(List<SchemaPath> columns) KuduGroupScan.clone
(List<SchemaPath> columns) ModifierConstructorDescriptionKuduGroupScan
(KuduScanSpec kuduScanSpec, KuduStoragePluginConfig kuduStoragePluginConfig, List<SchemaPath> columns, StoragePluginRegistry pluginRegistry) KuduGroupScan
(KuduStoragePlugin kuduStoragePlugin, KuduScanSpec kuduScanSpec, List<SchemaPath> columns) KuduRecordReader
(org.apache.kudu.client.KuduClient client, KuduSubScan.KuduSubScanSpec subScanSpec, List<SchemaPath> projectedColumns) KuduSubScan
(StoragePluginRegistry registry, KuduStoragePluginConfig kuduStoragePluginConfig, LinkedList<KuduSubScan.KuduSubScanSpec> tabletScanSpecList, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.mock
Modifier and TypeMethodDescriptionboolean
MockGroupScanPOP.canPushdownProjects
(List<SchemaPath> columns) MockGroupScanPOP.clone
(List<SchemaPath> columns) MockStorageEngine.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.mongo
Modifier and TypeMethodDescriptionMongoGroupScan.getColumns()
MongoSubScan.getColumns()
protected Collection<SchemaPath>
MongoRecordReader.transformColumns
(Collection<SchemaPath> projectedColumns) Modifier and TypeMethodDescriptionMongoCompareFunctionProcessor.visitSchemaPath
(SchemaPath path, LogicalExpression valueArg) Modifier and TypeMethodDescriptionboolean
MongoGroupScan.canPushdownProjects
(List<SchemaPath> columns) MongoGroupScan.clone
(List<SchemaPath> columns) protected Collection<SchemaPath>
MongoRecordReader.transformColumns
(Collection<SchemaPath> projectedColumns) ModifierConstructorDescriptionMongoGroupScan
(String userName, MongoScanSpec scanSpec, MongoStoragePluginConfig storagePluginConfig, List<SchemaPath> columns, boolean useAggregate, StoragePluginRegistry pluginRegistry) MongoGroupScan
(String userName, MongoStoragePlugin storagePlugin, MongoScanSpec scanSpec, List<SchemaPath> columns, boolean useAggregate) MongoRecordReader
(BaseMongoSubScanSpec subScanSpec, List<SchemaPath> projectedColumns, FragmentContext context, MongoStoragePlugin plugin) MongoSubScan
(StoragePluginRegistry registry, String userName, StoragePluginConfig mongoPluginConfig, LinkedList<BaseMongoSubScanSpec> chunkScanSpecList, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.openTSDB
Modifier and TypeMethodDescriptionboolean
OpenTSDBGroupScan.canPushdownProjects
(List<SchemaPath> columns) OpenTSDBGroupScan.clone
(List<SchemaPath> columns) ModifierConstructorDescriptionOpenTSDBGroupScan
(OpenTSDBScanSpec openTSDBScanSpec, OpenTSDBStoragePluginConfig openTSDBStoragePluginConfig, List<SchemaPath> columns, StoragePluginRegistry pluginRegistry) OpenTSDBGroupScan
(OpenTSDBStoragePlugin storagePlugin, OpenTSDBScanSpec scanSpec, List<SchemaPath> columns) OpenTSDBRecordReader
(Service client, OpenTSDBSubScan.OpenTSDBSubScanSpec subScanSpec, List<SchemaPath> projectedColumns) OpenTSDBSubScan
(StoragePluginRegistry registry, OpenTSDBStoragePluginConfig storage, LinkedList<OpenTSDBSubScan.OpenTSDBSubScanSpec> tabletScanSpecList, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.store.parquet
Modifier and TypeFieldDescriptionprotected final List<SchemaPath>
AbstractParquetRowGroupScan.columns
Modifier and TypeMethodDescriptionstatic Map<SchemaPath,
ColumnStatistics<?>> ParquetTableMetadataUtils.addImplicitColumnsStatistics
(Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, List<SchemaPath> columns, List<String> partitionValues, OptionManager optionManager, org.apache.hadoop.fs.Path location, boolean supportsFileImplicitColumns) Creates new map based on specifiedcolumnStatistics
with added statistics for implicit and partition (dir) columns.AbstractParquetRowGroupScan.getColumns()
static Map<SchemaPath,
ColumnStatistics<?>> ParquetTableMetadataUtils.getColumnStatistics
(TupleMetadata schema, DrillStatsTable statistics) Returns map with schema path andColumnStatistics
obtained from specifiedDrillStatsTable
for all columns from specifiedBaseTableMetadata
.static Map<SchemaPath,
TypeProtos.MajorType> ParquetTableMetadataUtils.getFileFields
(MetadataBase.ParquetTableMetadataBase parquetTableMetadata, MetadataBase.ParquetFileMetadata file) Returns map of column names with their drill types for specifiedfile
.static Map<SchemaPath,
TypeProtos.MajorType> ParquetTableMetadataUtils.getIntermediateFields
(MetadataBase.ParquetTableMetadataBase parquetTableMetadata, MetadataBase.RowGroupMetadata rowGroup) Returns map of column names with their Drill types for everyNameSegment
inSchemaPath
in specifiedrowGroup
.BaseParquetMetadataProvider.getPartitionColumns()
ParquetGroupScanStatistics.getPartitionColumns()
static Map<SchemaPath,
ColumnStatistics<?>> ParquetTableMetadataUtils.getRowGroupColumnStatistics
(MetadataBase.ParquetTableMetadataBase tableMetadata, MetadataBase.RowGroupMetadata rowGroupMetadata) Converts specifiedMetadataBase.RowGroupMetadata
into the map ofColumnStatistics
instances with column names as keys.static Map<SchemaPath,
TypeProtos.MajorType> ParquetTableMetadataUtils.getRowGroupFields
(MetadataBase.ParquetTableMetadataBase parquetTableMetadata, MetadataBase.RowGroupMetadata rowGroup) Returns map of column names with their drill types for specifiedrowGroup
.FilterEvaluatorUtils.FieldReferenceFinder.visitSchemaPath
(SchemaPath path, Void value) FilterEvaluatorUtils.FieldReferenceFinder.visitUnknown
(LogicalExpression e, Void value) Modifier and TypeMethodDescriptionlong
ParquetGroupScanStatistics.getColumnValueCount
(SchemaPath column) BaseParquetMetadataProvider.getPartitionMetadata
(SchemaPath columnName) static PartitionMetadata
ParquetTableMetadataUtils.getPartitionMetadata
(SchemaPath partitionColumn, List<FileMetadata> files) ReturnsPartitionMetadata
instance received by merging specifiedFileMetadata
list.ParquetGroupScanStatistics.getPartitionPaths
(SchemaPath column) ParquetGroupScanStatistics.getPartitionValue
(org.apache.hadoop.fs.Path path, SchemaPath column) ParquetGroupScanStatistics.getTypeForColumn
(SchemaPath schemaPath) boolean
AbstractParquetRowGroupScan.isImplicitColumn
(SchemaPath path, String partitionColumnLabel) FilterEvaluatorUtils.FieldReferenceFinder.visitSchemaPath
(SchemaPath path, Void value) Modifier and TypeMethodDescriptionstatic Map<SchemaPath,
ColumnStatistics<?>> ParquetTableMetadataUtils.addImplicitColumnsStatistics
(Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, List<SchemaPath> columns, List<String> partitionValues, OptionManager optionManager, org.apache.hadoop.fs.Path location, boolean supportsFileImplicitColumns) Creates new map based on specifiedcolumnStatistics
with added statistics for implicit and partition (dir) columns.static Map<SchemaPath,
ColumnStatistics<?>> ParquetTableMetadataUtils.addImplicitColumnsStatistics
(Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, List<SchemaPath> columns, List<String> partitionValues, OptionManager optionManager, org.apache.hadoop.fs.Path location, boolean supportsFileImplicitColumns) Creates new map based on specifiedcolumnStatistics
with added statistics for implicit and partition (dir) columns.boolean
AbstractParquetGroupScan.canPushdownProjects
(List<SchemaPath> columns) ParquetReaderUtility.checkForCorruptDateValuesInStatistics
(org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, boolean autoCorrectCorruptDates) Detect corrupt date values by looking at the min/max values in the metadata.ParquetGroupScan.clone
(List<SchemaPath> columns) static boolean
ParquetReaderUtility.containsComplexColumn
(org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns) Check whether any of columns in the given list is either nested or repetitive.abstract AbstractParquetRowGroupScan
AbstractParquetRowGroupScan.copy
(List<SchemaPath> columns) ParquetRowGroupScan.copy
(List<SchemaPath> columns) ParquetReaderUtility.detectCorruptDates
(org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, boolean autoCorrectCorruptDates) Check for corrupted dates in a parquet file.protected void
AbstractParquetGroupScan.RowGroupScanFilterer.filterFileMetadata
(OptionManager optionManager, FilterPredicate<?> filterPredicate, Set<SchemaPath> schemaPathsInExpr) Produces filtering of metadata at file level.ParquetFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns) ParquetFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options) ParquetFormatPlugin.getGroupScan
(String userName, FileSelection selection, List<SchemaPath> columns, OptionManager options, MetadataProviderManager metadataProviderManager) static RowsMatch
FilterEvaluatorUtils.matches
(LogicalExpression expr, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, TupleMetadata schema, long rowCount, UdfUtilities udfUtilities, FunctionLookupContext functionImplementationRegistry, Set<SchemaPath> schemaPathsInExpr) static RowsMatch
FilterEvaluatorUtils.matches
(LogicalExpression expr, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, TupleMetadata schema, long rowCount, UdfUtilities udfUtilities, FunctionLookupContext functionImplementationRegistry, Set<SchemaPath> schemaPathsInExpr) static <T extends Comparable<T>>
RowsMatchFilterEvaluatorUtils.matches
(FilterPredicate<T> parquetPredicate, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, long rowCount, TupleMetadata fileMetadata, Set<SchemaPath> schemaPathsInExpr, UdfUtilities udfUtilities) static <T extends Comparable<T>>
RowsMatchFilterEvaluatorUtils.matches
(FilterPredicate<T> parquetPredicate, Map<SchemaPath, ColumnStatistics<?>> columnsStatistics, long rowCount, TupleMetadata fileMetadata, Set<SchemaPath> schemaPathsInExpr, UdfUtilities udfUtilities) ModifierConstructorDescriptionprotected
AbstractParquetGroupScan
(String userName, List<SchemaPath> columns, List<ReadEntryWithPath> entries, ParquetReaderConfig readerConfig, LogicalExpression filter) ParquetGroupScan
(String userName, FileSelection selection, ParquetFormatPlugin formatPlugin, List<SchemaPath> columns, ParquetReaderConfig readerConfig, LogicalExpression filter, MetadataProviderManager metadataProviderManager) ParquetGroupScan
(String userName, FileSelection selection, ParquetFormatPlugin formatPlugin, List<SchemaPath> columns, ParquetReaderConfig readerConfig, MetadataProviderManager metadataProviderManager) ParquetRowGroupScan
(StoragePluginRegistry registry, String userName, StoragePluginConfig storageConfig, FormatPluginConfig formatConfig, LinkedList<RowGroupReadEntry> rowGroupReadEntries, List<SchemaPath> columns, ParquetReaderConfig readerConfig, org.apache.hadoop.fs.Path selectionRoot, LogicalExpression filter, TupleMetadata schema) -
Uses of SchemaPath in org.apache.drill.exec.store.parquet.columnreaders
Modifier and TypeMethodDescriptionprotected List<SchemaPath>
ParquetRecordReader.getDefaultColumnsToRead()
ModifierConstructorDescriptionParquetRecordReader
(FragmentContext fragmentContext, long numRecordsToRead, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) ParquetRecordReader
(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, long numRecordsToRead, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) ParquetRecordReader
(FragmentContext fragmentContext, org.apache.hadoop.fs.Path path, int rowGroupIndex, org.apache.hadoop.fs.FileSystem fs, org.apache.parquet.compression.CompressionCodecFactory codecFactory, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, List<SchemaPath> columns, ParquetReaderUtility.DateCorruptionStatus dateCorruptionStatus) ParquetSchema
(OptionManager options, int rowGroupIndex, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, Collection<SchemaPath> selectedCols) Build the Parquet schema. -
Uses of SchemaPath in org.apache.drill.exec.store.parquet.metadata
Modifier and TypeMethodDescriptionstatic void
Metadata.createMeta
(org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path path, ParquetReaderConfig readerConfig, boolean allColumnsInteresting, Set<SchemaPath> columnSet) Create the parquet metadata file for the directory at the given path, and for any subdirectories.Metadata.getParquetFileMetadata_v4
(Metadata_V4.ParquetTableMetadata_v4 parquetTableMetadata, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, org.apache.hadoop.fs.FileStatus file, org.apache.hadoop.fs.FileSystem fs, boolean allColumnsInteresting, boolean skipNonInteresting, Set<SchemaPath> columnSet, ParquetReaderConfig readerConfig) Get the file metadata for a single fileModifierConstructorDescriptionColumnMetadata_v1
(SchemaPath name, org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName primitiveType, org.apache.parquet.schema.OriginalType originalType, Object max, Object min, Long nulls) Key
(SchemaPath name) Key
(SchemaPath name) ModifierConstructorDescriptionFileMetadataCollector
(org.apache.parquet.hadoop.metadata.ParquetMetadata metadata, org.apache.hadoop.fs.FileStatus file, org.apache.hadoop.fs.FileSystem fs, boolean allColumnsInteresting, boolean skipNonInteresting, Set<SchemaPath> columnSet, ParquetReaderConfig readerConfig) -
Uses of SchemaPath in org.apache.drill.exec.store.parquet2
ModifierConstructorDescriptionDrillParquetGroupConverter
(OutputMutator mutator, BaseWriter baseWriter, org.apache.parquet.schema.GroupType schema, Collection<SchemaPath> columns, OptionManager options, ParquetReaderUtility.DateCorruptionStatus containsCorruptedDates, boolean skipRepeated, String parentName) The constructor is responsible for creation of converters tree and may invoke itself for creation of child converters when nested field is group type field too.DrillParquetReader
(FragmentContext fragmentContext, org.apache.parquet.hadoop.metadata.ParquetMetadata footer, RowGroupReadEntry entry, List<SchemaPath> columns, DrillFileSystem fileSystem, ParquetReaderUtility.DateCorruptionStatus containsCorruptedDates, long recordsToRead) DrillParquetRecordMaterializer
(OutputMutator mutator, org.apache.parquet.schema.MessageType schema, Collection<SchemaPath> columns, OptionManager options, ParquetReaderUtility.DateCorruptionStatus containsCorruptedDates) -
Uses of SchemaPath in org.apache.drill.exec.store.phoenix
ModifierConstructorDescriptionPhoenixGroupScan
(String user, String sql, List<SchemaPath> columns, PhoenixScanSpec scanSpec, PhoenixStoragePlugin plugin) PhoenixGroupScan
(String userName, String sql, List<SchemaPath> columns, PhoenixScanSpec scanSpec, PhoenixStoragePluginConfig config, StoragePluginRegistry plugins) PhoenixGroupScan
(PhoenixGroupScan scan, List<SchemaPath> columns) PhoenixSubScan
(String userName, String sql, List<SchemaPath> columns, PhoenixScanSpec scanSpec, StoragePluginConfig config, StoragePluginRegistry registry) PhoenixSubScan
(String userName, String sql, List<SchemaPath> columns, PhoenixScanSpec scanSpec, PhoenixStoragePlugin plugin) -
Uses of SchemaPath in org.apache.drill.exec.store.splunk
Modifier and TypeMethodDescriptionboolean
SplunkGroupScan.canPushdownProjects
(List<SchemaPath> columns) SplunkGroupScan.clone
(List<SchemaPath> columns) SplunkStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) SplunkStoragePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns, SessionOptionManager options, MetadataProviderManager metadataProviderManager) ModifierConstructorDescriptionSplunkGroupScan
(SplunkGroupScan that, List<SchemaPath> columns) Applies columns.SplunkGroupScan
(SplunkPluginConfig config, List<SchemaPath> columns, SplunkScanSpec splunkScanSpec, Map<String, ExprNode.ColRelOpConstNode> filters, double selectivity, int maxRecords) Deserialize a group scan.SplunkSubScan
(SplunkPluginConfig config, SplunkScanSpec splunkScanSpec, List<SchemaPath> columns, Map<String, ExprNode.ColRelOpConstNode> filters, int maxRecords) -
Uses of SchemaPath in org.apache.drill.exec.store.sys
Modifier and TypeMethodDescriptionSystemTableScan.clone
(List<SchemaPath> columns) SystemTablePlugin.getPhysicalScan
(String userName, JSONOptions selection, List<SchemaPath> columns) -
Uses of SchemaPath in org.apache.drill.exec.util
Modifier and TypeMethodDescriptionstatic Collection<SchemaPath>
EncodedSchemaPathSet.decode
(Collection<SchemaPath> encodedPaths) Returns the decoded Collection of SchemaPath from the input which may contain a mix of encoded and non-encoded SchemaPaths.Modifier and TypeMethodDescriptionstatic boolean
EncodedSchemaPathSet.isEncodedSchemaPath
(SchemaPath schemaPath) Modifier and TypeMethodDescriptionstatic Collection<SchemaPath>
EncodedSchemaPathSet.decode
(Collection<SchemaPath> encodedPaths) Returns the decoded Collection of SchemaPath from the input which may contain a mix of encoded and non-encoded SchemaPaths.static boolean
Utilities.isStarQuery
(Collection<SchemaPath> projected) Return true if list of schema path has star column. -
Uses of SchemaPath in org.apache.drill.exec.vector.complex
Modifier and TypeMethodDescriptionstatic TypedFieldId
FieldIdUtil.getFieldId
(ValueVector vector, int id, SchemaPath expectedPath, boolean hyper) -
Uses of SchemaPath in org.apache.drill.exec.vector.complex.fn
Modifier and TypeMethodDescriptionstatic void
JsonReaderUtils.ensureAtLeastOneField
(BaseWriter.ComplexWriter writer, Collection<SchemaPath> columns, boolean allTextMode, List<BaseWriter.ListWriter> emptyArrayWriters) static FieldSelection
FieldSelection.getFieldSelection
(List<SchemaPath> fields) Generates a field selection based on a list of fields.JsonReader.Builder.schemaPathColumns
(List<SchemaPath> columns) static void
JsonReaderUtils.writeColumnsUsingSchema
(BaseWriter.ComplexWriter writer, Collection<SchemaPath> columns, TupleMetadata schema, boolean allTextMode) Creates writers which correspond to the specified schema for specified root writer. -
Uses of SchemaPath in org.apache.drill.metastore.metadata
Modifier and TypeFieldDescriptionprotected Map<SchemaPath,
ColumnStatistics<?>> BaseMetadata.BaseMetadataBuilder.columnsStatistics
protected final Map<SchemaPath,
ColumnStatistics<?>> BaseMetadata.columnsStatistics
Modifier and TypeMethodDescriptionPartitionMetadata.getColumn()
It allows to obtain the column path for this partitionSegmentMetadata.getColumn()
Modifier and TypeMethodDescriptionBaseMetadata.getColumnsStatistics()
Metadata.getColumnsStatistics()
Returns statistics stored in current metadata represented as Map of columnSchemaPath
s and correspondingColumnStatistics
.NonInterestingColumnsMetadata.getColumnsStatistics()
BaseTableMetadata.getInterestingColumns()
TableMetadata.getInterestingColumns()
TableMetadataProvider.getPartitionColumns()
Returns list of partition columns for table from thisTableMetadataProvider
.Modifier and TypeMethodDescriptionPartitionMetadata.PartitionMetadataBuilder.column
(SchemaPath column) SegmentMetadata.SegmentMetadataBuilder.column
(SchemaPath column) BaseMetadata.getColumn
(SchemaPath name) Metadata.getColumn
(SchemaPath name) Returns metadata description for the specified columnNonInterestingColumnsMetadata.getColumn
(SchemaPath name) BaseMetadata.getColumnStatistics
(SchemaPath columnName) Metadata.getColumnStatistics
(SchemaPath columnName) Returns statistics for specified column stored in current metadata.NonInterestingColumnsMetadata.getColumnStatistics
(SchemaPath columnName) TableMetadataProvider.getPartitionMetadata
(SchemaPath columnName) Returns list ofPartitionMetadata
instances which corresponds to partitions for specified column and provides metadata for specific partitions and its columns.<V> V
BaseMetadata.getStatisticsForColumn
(SchemaPath columnName, StatisticsKind<V> statisticsKind) <V> V
Metadata.getStatisticsForColumn
(SchemaPath columnName, StatisticsKind<V> statisticsKind) Returns value of column statistics which corresponds to specifiedStatisticsKind
for column with specifiedcolumnName
.<V> V
NonInterestingColumnsMetadata.getStatisticsForColumn
(SchemaPath columnName, StatisticsKind<V> statisticsKind) Modifier and TypeMethodDescriptionBaseTableMetadata.cloneWithStats
(Map<SchemaPath, ColumnStatistics<?>> columnStatistics, List<StatisticsHolder<?>> tableStatistics) TableMetadata.cloneWithStats
(Map<SchemaPath, ColumnStatistics<?>> columnStatistics, List<StatisticsHolder<?>> tableStatistics) BaseMetadata.BaseMetadataBuilder.columnsStatistics
(Map<SchemaPath, ColumnStatistics<?>> columnsStatistics) BaseTableMetadata.BaseTableMetadataBuilder.interestingColumns
(List<SchemaPath> interestingColumns) ModifierConstructorDescriptionNonInterestingColumnsMetadata
(Map<SchemaPath, ColumnStatistics<?>> columnsStatistics) -
Uses of SchemaPath in org.apache.drill.metastore.util
Modifier and TypeMethodDescriptionstatic <T extends BaseMetadata>
Map<SchemaPath,ColumnStatistics<?>> TableMetadataUtils.mergeColumnsStatistics
(Collection<T> metadataList, Set<SchemaPath> columns, List<CollectableColumnStatisticsKind<?>> statisticsToCollect) Merges list of specified metadata into the map ofColumnStatistics
with columns as keys.Modifier and TypeMethodDescriptionstatic void
SchemaPathUtils.addColumnMetadata
(TupleMetadata schema, SchemaPath schemaPath, TypeProtos.MajorType type, Map<SchemaPath, TypeProtos.MajorType> types) Adds column with specified schema path and type into specifiedTupleMetadata schema
.static ColumnMetadata
SchemaPathUtils.getColumnMetadata
(SchemaPath schemaPath, TupleMetadata schema) ReturnsColumnMetadata
instance obtained from specifiedTupleMetadata schema
which corresponds to the specified column schema path.static boolean
SchemaPathUtils.isFieldNestedInDictOrRepeatedMap
(SchemaPath schemaPath, TupleMetadata schema) Checks if field identified by the schema path is child in eitherDICT
orREPEATED MAP
.Modifier and TypeMethodDescriptionstatic void
SchemaPathUtils.addColumnMetadata
(TupleMetadata schema, SchemaPath schemaPath, TypeProtos.MajorType type, Map<SchemaPath, TypeProtos.MajorType> types) Adds column with specified schema path and type into specifiedTupleMetadata schema
.static <T extends BaseMetadata>
Map<SchemaPath,ColumnStatistics<?>> TableMetadataUtils.mergeColumnsStatistics
(Collection<T> metadataList, Set<SchemaPath> columns, List<CollectableColumnStatisticsKind<?>> statisticsToCollect) Merges list of specified metadata into the map ofColumnStatistics
with columns as keys.