hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Tatsuya Kawano <tatsuy...@snowcocoa.info>
Subject Re: HBase 0.20.1 on Ubuntu 9.04: master fails to start
Date Tue, 27 Oct 2009 08:01:04 GMT
Hi Artyom,

Your configuration files look just fine.


>> 2009-10-26 13:34:30,031 WARN org.apache.hadoop.hdfs.dfsclient: datastreamer
>> exception: org.apache.hadoop.ipc.remoteexcep tion: java.io.ioexception: file
>> /hbase.version could only be replicated to 0 nodes, instead of 1

I'm not totally sure, but I think this exception occurs when there is
no HDFS data node available in the cluster.

Can you access to the HDFS name node status screen at
<http://servers-ip:50070/> from a web browser to see if there is a
data node available?

Thanks,

-- 
Tatsuya Kawano (Mr.)
Tokyo, Japan


On Tue, Oct 27, 2009 at 11:24 AM, Artyom Shvedchikov <shoolc@gmail.com> wrote:
> Hello.
>
> We are testing the latest HBase 0.20.1 in pseudo-distributed mode with
> Hadoop 0.20.1 on such environment:
> *h/w*: Intel C2D 1.86 GHz, RAM 2 Gb 667 MHz, HDD 1TB Seagate SATA2 7200 Rpm
> *s/w*: Ubuntu 9.04, Filesystem type is *ext3*, Java  1.6.0_16-b01, Hadoop
> 0.20.1, HBase 0.20.1
>
> File */etc/hosts*
>
>> 127.0.0.1       localhost
>>
>> # The following lines are desirable for IPv6 capable hosts
>> ::1     localhost ip6-localhost ip6-loopback
>> fe00::0 ip6-localnet
>> ff00::0 ip6-mcastprefix
>> ff02::1 ip6-allnodes
>> ff02::2 ip6-allrouters
>> ff02::3 ip6-allhosts
>>
> Hadoop and HBase are running in pseudo-distributed mode:
> Two options added to *hadoop-env.sh*:
>
>> export JAVA_HOME=/usr/lib/jvm/java-6-sun
>> export HADOOP_OPTS=-Djava.net.preferIPv4Stack=true
>>
> *core-site.xml*:
>
>> <configuration>
>> <property>
>>   <name>fs.default.name</name>
>>   <value>hdfs://127.0.0.1:9000</value>
>> </property>
>> <property>
>>   <name>hadoop.tmp.dir</name>
>>   <value>/hadoop/tmp/hadoop-${user.name}</value>
>>   <description>A base for other temporary directories.</description>
>> </property>
>> </configuration>
>>
> *hdfs-site.xml*:
>
>> <configuration>
>>   <property>
>>     <name>dfs.replication</name>
>>     <value>1</value>
>>   </property>
>> <property>
>>   <name>dfs.name.dir</name>
>>   <value>/hadoop/hdfs/name</value>
>> </property>
>> <property>
>>   <name>dfs.data.dir</name>
>>   <value>/hadoop/hdfs/data</value>
>> </property>
>> <property>
>>   <name>dfs.datanode.socket.write.timeout</name>
>>   <value>0</value>
>> </property>
>> <property>
>>    <name>dfs.datanode.max.xcievers</name>
>>    <value>1023</value>
>> </property>
>> </configuration>
>>
> *marped-site.xml:*
>
>> <configuration>
>> <property>
>>   <name>mapred.job.tracker</name>
>>   <value>127.0.0.1:9001</value>
>> </property>
>> </configuration>
>>
> *hbase-site.xml:*
>
>> <configuration>
>>   <property>
>>     <name>hbase.rootdir</name>
>>     <value>hdfs://localhost:9000/</value>
>>     <description>The directory shared by region servers.
>>     Should be fully-qualified to include the filesystem to use.
>>     E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR
>>     </description>
>>   </property>
>>   <property>
>>     <name>hbase.master</name>
>>     <value>127.0.0.1:60000</value>
>>     <description>The host and port that the HBase master runs at.
>>     </description>
>>   </property>
>>   <property>
>>      <name>hbase.tmp.dir</name>
>>      <value>/hadoop/tmp/hbase-${user.name}</value>
>>      <description>Temporary directory on the local
>> filesystem.</description>
>>   </property>
>>     <property>
>>         <name>hbase.zookeeper.quorum</name>
>>         <value>127.0.0.1</value>
>>         <description>The directory shared by region servers.
>>         </description>
>>     </property>
>> </configuration>
>>
>  Hadoop and HBase are running under *hbase *user, all necessary directories
> are owned by *hbase *user (I mean */hadoop* directory and all its
> subdirectories).
>
> First launch was successfull, but after several days of work we trapt in
> problem that hbase master was down, then we tried to restart it (*
> stop-hbase.sh*, then *start-hbase.sh*) - restart fails with error:
>
>> 2009-10-26 13:34:30,031 WARN org.apache.hadoop.hdfs.dfsclient: datastreamer
>> exception: org.apache.hadoop.ipc.remoteexcep tion: java.io.ioexception: file
>> /hbase.version could only be replicated to 0 nodes, instead of 1         at
>> org.apache.hadoop.hdfs.server.namenode.fsnamesystem.getadditionalblock(fsnamesystem.java:1267)
>> at
>> org.apache.hadoop.hdfs.server.namenode.namenode.addblock(namenode.java:422)
>>
>
> Then I tried to reformat hdfs (then, also remove all hadoop and hbase data,
> then format hdfs again) and start hadoop and hbase again, but HBase master
> fails to start with the same error.
>
> Could someone revise our configuration and tell us what is the reason for
> such HBase master instance behaviour?
>
> Thanks in advance, Artyom
> -------------------------------------------------
> Best wishes, Artyom Shvedchikov
>

Mime
View raw message