public class OrcShimV200 extends Object implements OrcShim<org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch>
| Constructor and Description |
|---|
OrcShimV200() |
| Modifier and Type | Method and Description |
|---|---|
static boolean[] |
computeProjectionMask(org.apache.orc.TypeDescription schema,
int[] selectedFields)
Computes the ORC projection mask of the fields to include from the selected
fields.rowOrcInputFormat.nextRecord(null).
|
HiveOrcBatchWrapper |
createBatchWrapper(org.apache.orc.TypeDescription schema,
int batchSize) |
protected org.apache.orc.Reader |
createReader(org.apache.hadoop.fs.Path path,
org.apache.hadoop.conf.Configuration conf) |
org.apache.orc.RecordReader |
createRecordReader(org.apache.hadoop.conf.Configuration conf,
org.apache.orc.TypeDescription schema,
int[] selectedFields,
List<OrcFilters.Predicate> conjunctPredicates,
org.apache.flink.core.fs.Path path,
long splitStart,
long splitLength)
Create orc
RecordReader from conf, schema and etc... |
protected org.apache.orc.RecordReader |
createRecordReader(org.apache.orc.Reader reader,
org.apache.orc.Reader.Options options) |
static org.apache.flink.api.java.tuple.Tuple2<Long,Long> |
getOffsetAndLengthForSplit(long splitStart,
long splitLength,
List<org.apache.orc.StripeInformation> stripes) |
boolean |
nextBatch(org.apache.orc.RecordReader reader,
org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch rowBatch)
Read the next row batch.
|
protected org.apache.orc.Reader.Options |
readOrcConf(org.apache.orc.Reader.Options options,
org.apache.hadoop.conf.Configuration conf) |
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, waitcreateShim, defaultShimprotected org.apache.orc.Reader createReader(org.apache.hadoop.fs.Path path,
org.apache.hadoop.conf.Configuration conf)
throws IOException
IOExceptionprotected org.apache.orc.RecordReader createRecordReader(org.apache.orc.Reader reader,
org.apache.orc.Reader.Options options)
throws IOException
IOExceptionprotected org.apache.orc.Reader.Options readOrcConf(org.apache.orc.Reader.Options options,
org.apache.hadoop.conf.Configuration conf)
public org.apache.orc.RecordReader createRecordReader(org.apache.hadoop.conf.Configuration conf,
org.apache.orc.TypeDescription schema,
int[] selectedFields,
List<OrcFilters.Predicate> conjunctPredicates,
org.apache.flink.core.fs.Path path,
long splitStart,
long splitLength)
throws IOException
OrcShimRecordReader from conf, schema and etc...createRecordReader in interface OrcShim<org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch>IOExceptionpublic HiveOrcBatchWrapper createBatchWrapper(org.apache.orc.TypeDescription schema, int batchSize)
createBatchWrapper in interface OrcShim<org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch>public boolean nextBatch(org.apache.orc.RecordReader reader,
org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch rowBatch)
throws IOException
OrcShimnextBatch in interface OrcShim<org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch>IOException@VisibleForTesting public static org.apache.flink.api.java.tuple.Tuple2<Long,Long> getOffsetAndLengthForSplit(long splitStart, long splitLength, List<org.apache.orc.StripeInformation> stripes)
public static boolean[] computeProjectionMask(org.apache.orc.TypeDescription schema,
int[] selectedFields)
Copyright © 2014–2023 The Apache Software Foundation. All rights reserved.