Hi All I have few tables in hive and I wanted to run query against them with spark as execution engine.

Can I direct;y load these tables in spark shell and run query?

I tried with 
1.val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc)
2.qlContext.sql("FROM event_impressions select count(*)") where event_impressions is the table name.

It give me error saying "org.apache.spark.sql.AnalysisException: no such table event_impressions; line 1 pos 5"

Does anybody hit similar issues?