In Spark 2.1 we've added a from_json function that I think will do what you want.

On Fri, Nov 18, 2016 at 2:29 AM, kant kodali <kanth909@gmail.com> wrote:
This seem to work 

import org.apache.spark.sql._
val rdd = df2.rdd.map { case Row(j: String) => j }
spark.read.json(rdd).show()
However I wonder if this any inefficiency here ? since I have to apply this function for billion rows.