spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sadhan Sood <>
Subject Adding partitions to parquet data
Date Thu, 20 Nov 2014 18:33:28 GMT
We are loading parquet data as temp tables but wondering if there is a way
to add a partition to the data without going through hive (we still want to
use spark's parquet serde as compared to hive). The data looks like ->

/date1/file1, /date1/file2 ... , /date2/file1,

and we are loading it like:
val parquetFileRDD = sqlContext.parquetFile(comma separated parquet file

but it would be nice to able to add a partition and provide date in the
query parameter.

View raw message