kylin-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "fengYu (JIRA)" <>
Subject [jira] [Commented] (KYLIN-1280) Convert Cuboid Data to HFile failed when hbase in different HDFS
Date Fri, 22 Jan 2016 14:04:39 GMT


fengYu commented on KYLIN-1280:

[]This is caused by KYLIN-957, because I deploy kylin with two HDFS,
one for hadoop(hdfs://A) and another is for hbase(hdfs://B), in step 'Convert Cuboid Data
to HFile', the mr job input is at hdfs://A and output is to hdfs://B, But in most hadoop cluster,
it can not recognize hdfs://B, so the job always failed.
One way to solve it just redeploy hadoop cluster and make it recognize hdfs://B, or use this
patch to change name service to current master name node before execute the job.

> Convert Cuboid Data to HFile failed when hbase in different HDFS
> ----------------------------------------------------------------
>                 Key: KYLIN-1280
>                 URL:
>             Project: Kylin
>          Issue Type: Bug
>    Affects Versions: 2.0
>            Reporter: fengYu
>         Attachments: 0001-transform-path-in-other-HDFS-to-real-name-node-path.patch
> I deploy kylin-2.0 with hbase which rely on a different HDFS with hadoop cluster, so
I config this property 'kylin.hbase.cluster.fs' = hdfs://A, the name service is different
with 'fs.defaultFS' in hadoop cluster which is hdfs://B.
> In the step 'Convert Cuboid Data to HFile' execute failed, error log is :
> Failed to run job : Unable to map logical nameservice URI 'hdfs://A'
to a NameNode. Local configuration does not have a failover proxy provide
> r configured.
>         at org.apache.hadoop.mapred.YARNRunner.submitJob(
>         at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(
>         at org.apache.hadoop.mapreduce.Job$
>         at org.apache.hadoop.mapreduce.Job$
>         at Method)
>         at
>         at
>         at org.apache.hadoop.mapreduce.Job.submit(
>         at
>         at
>         at
>         at
>         at
>         at org.apache.kylin.job.execution.AbstractExecutable.execute(
>         at org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(
>         at org.apache.kylin.job.execution.AbstractExecutable.execute(
>         at org.apache.kylin.job.impl.threadpool.DefaultScheduler$
>         at java.util.concurrent.ThreadPoolExecutor.runWorker(
>         at java.util.concurrent.ThreadPoolExecutor$
>         at
> I think it is because node manager in hadoop cluster can not recognition hdfs://A in
they config. So, I have to tranform the path hdfs://A/path/to/hfile to hdfs://namenode_ip:port/path/to/hfile
before execute this step. and it works for me.  
> Here is my patch.

This message was sent by Atlassian JIRA

View raw message