Hi All I have few tables in hive and I wanted to run query against them with spark as execution engine.
Can I direct;y load these tables in spark shell and run query?
I tried with
1.val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc)
2.qlContext.sql("FROM event_impressions select count(*)") where event_impressions is the table name.
It give me error saying "org.apache.spark.sql.AnalysisException: no such table event_impressions; line 1 pos 5"
Does anybody hit similar issues?