Created
November 24, 2015 19:30
-
-
Save omalley/7a53cb3ae91fa4c22023 to your computer and use it in GitHub Desktop.
An example ORC writer using a dynamic schema in Hive 2.0
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import org.apache.hadoop.conf.Configuration; | |
import org.apache.hadoop.fs.Path; | |
import org.apache.hadoop.hive.ql.exec.vector.BytesColumnVector; | |
import org.apache.hadoop.hive.ql.exec.vector.LongColumnVector; | |
import org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch; | |
import org.apache.orc.CompressionKind; | |
import org.apache.orc.TypeDescription; | |
import org.apache.orc.OrcFile; | |
import org.apache.orc.Writer; | |
import java.io.IOException; | |
public class OrcWriter2 { | |
public static void main(String[] args) throws IOException, | |
InterruptedException { | |
String path = "/tmp/file2.orc"; | |
TypeDescription schema = TypeDescription.createStruct() | |
.addField("field1", TypeDescription.createInt()) | |
.addField("field2", TypeDescription.createString()) | |
.addField("field3", TypeDescription.createString()); | |
Configuration conf = new Configuration(); | |
Writer writer = OrcFile.createWriter(new Path(path), | |
OrcFile.writerOptions(conf) | |
.setSchema(schema) | |
.stripeSize(100000) | |
.bufferSize(10000) | |
.compress(CompressionKind.ZLIB) | |
.version(OrcFile.Version.V_0_12)); | |
VectorizedRowBatch batch = schema.createRowBatch(); | |
batch.size = 1; | |
((LongColumnVector) batch.cols[0]).vector[0] = 1; | |
((BytesColumnVector) batch.cols[1]).setVal(0, "hello".getBytes()); | |
((BytesColumnVector) batch.cols[2]).setVal(0, "orcFile".getBytes()); | |
writer.addRowBatch(batch); | |
writer.close(); | |
} | |
} |
use hive2.3.3,core-orc1.3.3
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
why i write this into hive's hdfs filepath,but select result is null