Just to follow up on this(so if someone searches and hits this post),
This was solved now. Basically, I found hadoop dfs examples, found out
it was hadoop and DID need to fix /etc/hosts AND needed to move some
info from hdfs-site.xml to core-site.xml. Once the hadoop example
worked, the hbase examples started working.
Thanks,
Dean
-----Original Message-----
From: Buttler, David [mailto:buttler1@llnl.gov]
Sent: Wednesday, December 08, 2010 2:50 PM
To: user@hbase.apache.org
Subject: RE: hbase/hadoop config(do I need to change
/etc/hosts???)-fixed now but using local file system..grrrr
Are you sure you have the distributed option set to true?
Dave
-----Original Message-----
From: Hiller, Dean (Contractor) [mailto:dean.hiller@broadridge.com]
Sent: Wednesday, December 08, 2010 1:15 PM
To: user@hbase.apache.org
Subject: RE: hbase/hadoop config(do I need to change
/etc/hosts???)-fixed now but using local file system..grrrr
So /etc/hosts changes fixed it so that I have no 127.0.0.1 in the logs
anywhere anymore. (odd I didn't see that in the docs...I must have
missed that...too bad they don't use NetworkInterfaces.getInterfaces
instead of InetAddress.getLocalHost...can find the real ip
address easier independent of /etc/hosts).
Now, my client or hbase(not sure which) is still using the local file
system L. What config am I missing for this to work..(running the
performance example with just 10 rows to see if it works).
10/12/08 07:03:53 INFO zookeeper.ClientCnxn: Priming connection to
java.nio.chan
nels.SocketChannel[connected local=/206.88.43.168:49811
remote=/206.88.43.164:21
81]
10/12/08 07:03:53 INFO zookeeper.ClientCnxn: Server connection
successful
10/12/08 07:03:53 WARN mapred.JobClient: Use GenericOptionsParser for
parsing th
e arguments. Applications should implement Tool for the same.
10/12/08 07:03:53 INFO input.FileInputFormat: Total input paths to
process : 1
10/12/08 07:03:53 INFO hbase.PerformanceEvaluation: Total # of splits:
20
org.apache.hadoop.ipc.RemoteException: java.io.FileNotFoundException:
File file:
/tmp/hadoop-root/mapred/system/job_201012080654_0002/job.xml does not
exist.
at
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSys
tem.java:361)
at
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.
java:245)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:192)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:142)
at
org.apache.hadoop.fs.LocalFileSystem.copyToLocalFile(LocalFileSystem.
java:61)
at
org.apache.hadoop.fs.FileSystem.copyToLocalFile(FileSystem.java:1197)
at
org.apache.hadoop.mapred.JobInProgress.<init>(JobInProgress.java:257)
From: Hiller, Dean (Contractor)
Sent: Wednesday, December 08, 2010 1:34 PM
To: 'user@hbase.apache.org'
Subject: hbase/hadoop config(do I need to change /etc/hosts???)-still
getting 127.0.0.1 in logs
I am trying to run a system with no DNS(they don't have one in this lab
L ). Hopefully hbase/hadoop should still work fine, correct?
I am having a lot of trouble tracking down why 127.0.0.1 is being used
and why this shows up in hbase master logs...
2010-12-08 04:38:26,354 DEBUG
org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper
: Wrote master address 127.0.0.1:60000 to ZooKeeper
And
2010-12-08 04:38:29,899 INFO
org.apache.hadoop.hbase.master.ServerManager: Recei
ved start message from: localhost.localdomain,60020,1291808309865
2010-12-08 04:38:29,910 DEBUG
org.apache.hadoop.hbase.zookeeper.ZooKeeperWrapper
: Updated ZNode /hbase/rs/1291808309865 with data 127.0.0.1:60020
I am wondering if I need to tweak my /etc/hosts file? I am not sure
what java call is being made inside hbase there though.
My /etc/hosts file has(which MATCHES above exactly!!!!
"localhost.localdomain")
127.0.0.1 localhost.localdomain localhost
::1 localhost6.localdomain6 localhost6
My conf files are pretty simple(and currently-though I would be bad in
production-I am sharing all config files through
a mount so all nodes mount the one config location...this is just for
ease of changing the files though). What is wrong with
my config. (I also share this same config with the client I run who
keeps looking up 127.0.0.1 as well as the master). Am I
missing some kind of property override that comes from the defaults? (I
mean, I see 0.0.0.0 but I tend to think hbase may
be doing a lookup and I need to fix my /etc/hosts file?)
Hdfs-site.xml
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/data/hadoop</value>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://206.88.43.8:54310</value>
</property>
<property>
<name>dfs.replication</name>
<value>8</value>
</property>
</configuration>
Mapred-site.xml
<configuration>
<property>
<name>mapred.job.tracker</name>
<value>hdfs://206.88.43.4:54311</value>
</property>
<property>
<name>mapred.child.java.opts</name>
<value>-Xmx512m</value>
</property>
</configuration>
<configuration>
<property>
<name>hbase.rootdir</name>
<value>hdfs://206.88.43.8:54310/hbase</value>
<description>The directory shared by region servers.
Should be fully-qualified to include the filesystem to use.
E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
</description>
</property>
<property>
<name>hbase.master</name>
<value>206.88.43.8:60000</value>
<description>The host and port that the HBase master runs at.
</description>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
<description>The mode the cluster will be in. Possible values are
false: standalone and pseudo-distributed setups with managed
Zookeeper
true: fully-distributed with unmanaged Zookeeper Quorum (see
hbase-env.sh)
</description>
</property>
<property>
<name>hbase.zookeeper.quorum</name>
<value>206.88.43.8,206.88.43.4</value>
<description>Comma separated list of servers in the ZooKeeper
Quorum.
For example,
"host1.mydomain.com,host2.mydomain.com,host3.mydomain.com".
By default this is set to localhost for local and
pseudo-distributed modes
of operation. For a fully-distributed setup, this should be set to
a full
list of ZooKeeper quorum servers. If HBASE_MANAGES_ZK is set in
hbase-env.
sh
this is the list of servers which we will start/stop ZooKeeper on.
</description>
</property>
<property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/opt/data/zookeeper</value>
<description>Property from ZooKeeper's config zoo.cfg.
The directory where the snapshot is stored.
</description>
</property>
</configuration>
This message and any attachments are intended only for the use of the
addressee and
may contain information that is privileged and confidential. If the
reader of the
message is not the intended recipient or an authorized representative of
the
intended recipient, you are hereby notified that any dissemination of
this
communication is strictly prohibited. If you have received this
communication in
error, please notify us immediately by e-mail and delete the message and
any
attachments from your system.
This message and any attachments are intended only for the use of the addressee and
may contain information that is privileged and confidential. If the reader of the
message is not the intended recipient or an authorized representative of the
intended recipient, you are hereby notified that any dissemination of this
communication is strictly prohibited. If you have received this communication in
error, please notify us immediately by e-mail and delete the message and any
attachments from your system.
|