val sqlContext = new org.apache.spark.sql.SQLContext(sc)
val df = sqlContext.read.parquet("/path/to/file/without/hdfs://")
df.printSchema
df.count
import java.util
import org.apache.hadoop.io._
import org.apache.hadoop.hbase.util._
val filePath = "path/to/file"
spark.sparkContext.sequenceFile(filePath, classOf[BytesWritable], classOf[BytesWritable])
.map { case (rowKey, value) =>
val rowKeyBytes = util.Arrays.copyOf(rowKey.getBytes, rowKey.getLength)
val msgBytes = util.Arrays.copyOf(value.getBytes, value.getLength)
val keyValuePair = (Bytes.toString(rowKeyBytes), Bytes.toString(msgBytes))
val rowKeyStr = keyValuePair._1
keyValuePair
}.collect.foreach(println)