spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From HARIPRIYA AYYALASOMAYAJULA <aharipriy...@gmail.com>
Subject Re: Problem connecting to HBase
Date Sun, 15 Mar 2015 17:04:29 GMT
Hello all,

Thank you for your responses. I did try to include the
zookeeper.znode.parent property in the hbase-site.xml. It still continues
to give the same error.

I am using Spark 1.2.0 and hbase 0.98.9.

Could you please suggest what else could be done?


On Fri, Mar 13, 2015 at 10:25 PM, Ted Yu <yuzhihong@gmail.com> wrote:

> In HBaseTest.scala:
>     val conf = HBaseConfiguration.create()
> You can add some log (for zookeeper.znode.parent, e.g.) to see if the
> values from hbase-site.xml are picked up correctly.
>
> Please use pastebin next time you want to post errors.
>
> Which Spark release are you using ?
> I assume it contains SPARK-1297
>
> Cheers
>
> On Fri, Mar 13, 2015 at 7:47 PM, HARIPRIYA AYYALASOMAYAJULA <
> aharipriya92@gmail.com> wrote:
>
>>
>> Hello,
>>
>> I am running a HBase test case. I am using the example from the following:
>>
>> https://github.com/apache/spark/blob/master/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala
>>
>> I created a very small HBase table with 5 rows and 2 columns.
>> I have attached a screenshot of the error log. I believe it is a problem
>> where the driver program is unable to establish connection to the hbase.
>>
>> The following is my simple.sbt:
>>
>> name := "Simple Project"
>>
>> version := "1.0"
>>
>> scalaVersion := "2.10.4"
>>
>> libraryDependencies ++= Seq(
>>
>>  "org.apache.spark" %% "spark-core" % "1.2.0",
>>
>>  "org.apache.hbase" % "hbase" % "0.98.9-hadoop2" % "provided",
>>
>>  "org.apache.hbase" % "hbase-client" % "0.98.9-hadoop2" % "provided",
>>
>>  "org.apache.hbase" % "hbase-server" % "0.98.9-hadoop2" % "provided",
>>
>>  "org.apache.hbase" % "hbase-common" % "0.98.9-hadoop2" % "provided"
>> )
>>
>> I am using a 23 node cluster, did copy hbase-site.xml into /spark/conf
>> folder
>> and set spark.executor.extraClassPath pointing to the /hbase/ folder in
>> the spark-defaults.conf
>>
>> Also, while submitting the spark job I am including the required jars :
>>
>> spark-submit --class "HBaseTest" --master yarn-cluster
>> --driver-class-path
>>  /opt/hbase/0.98.9/lib/hbase-server-0.98.9-hadoop2.jar:/opt/hbase/0.98.9/lib/hbase-protocol-0.98.9-hadoop2.jar:/opt/hbase/0.98.9/lib/hbase-hadoop2-compat-0.98.9-hadoop2.jar:/opt/hbase/0.98.9/lib/hbase-client-0.98.9-hadoop2.jar:/opt/hbase/0.98.9/lib/hbase-common-0.98.9-hadoop2.jar:/opt/hbase/0.98.9/lib/htrace-core-2.04.jar
>>  /home/priya/usingHBase/Spark/target/scala-2.10/simple-project_2.10-1.0.jar
>> /Priya/sparkhbase-test1
>>
>> It would be great if you could point where I am going wrong, and what
>> could be done to correct it.
>>
>> Thank you for your time.
>> --
>> Regards,
>> Haripriya Ayyalasomayajula
>> Graduate Student
>> Department of Computer Science
>> University of Houston
>> Contact : 650-796-7112
>>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
>> For additional commands, e-mail: user-help@spark.apache.org
>>
>
>


-- 
Regards,
Haripriya Ayyalasomayajula
Graduate Student
Department of Computer Science
University of Houston
Contact : 650-796-7112

Mime
View raw message