spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From MEETHU MATHEW <meethu2...@yahoo.co.in>
Subject Re: org.jboss.netty.channel.ChannelException: Failed to bind to: master/1xx.xx..xx:0
Date Tue, 01 Jul 2014 06:59:44 GMT
Hi,

I did netstat -na | grep 192.168.125.174 and its showing 192.168.125.174:7077 LISTEN(after
starting master)

I tried to execute the following script from the slaves manually but it ends up with the same
exception and log.This script is internally executing the java command.
 /usr/local/spark-1.0.0/sbin/start-slave.sh 1 spark://192.168.125.174:7077
In this case netstat is showing any connection established to master:7077.

When we manually execute the java command,the connection is getting established to master.

Thanks & Regards, 
Meethu M


On Monday, 30 June 2014 6:38 PM, Akhil Das <akhil@sigmoidanalytics.com> wrote:
 


Are you sure you have this ip 192.168.125.174 bind for that machine? (netstat -na | grep 192.168.125.174)


Thanks
Best Regards


On Mon, Jun 30, 2014 at 5:34 PM, MEETHU MATHEW <meethu2006@yahoo.co.in> wrote:

Hi all,
>
>
>I reinstalled spark,reboot the system,but still I am not able to start the workers.Its
throwing the following exception:
>
>
>Exception in thread "main" org.jboss.netty.channel.ChannelException: Failed to bind to:
master/192.168.125.174:0
>
>
>I doubt the problem is with 192.168.125.174:0. Eventhough the command contains master:7077,why
its showing 0 in the log.
>
>
>java -cp ::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar
-XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m org.apache.spark.deploy.worker.Worker
spark://master:7077
>
>
>Can somebody tell me  a solution.
> 
>Thanks & Regards, 
>Meethu M
>
>
>
>On Friday, 27 June 2014 4:28 PM, MEETHU MATHEW <meethu2006@yahoo.co.in> wrote:
> 
>
>
>Hi,
>ya I tried setting another PORT also,but the same problem..
>master is set in etc/hosts
> 
>Thanks & Regards, 
>Meethu M
>
>
>
>On Friday, 27 June 2014 3:23 PM, Akhil Das <akhil@sigmoidanalytics.com> wrote:
> 
>
>
>tha's strange, did you try setting the master port to something else (use SPARK_MASTER_PORT).
>
>
>Also you said you are able to start it from the java commandline
>
>
>java -cp ::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar
-XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m org.apache.spark.deploy.worker.Worker
spark://:master:7077
>
>
>
>What is the master ip specified here? is it like you have entry for master in the /etc/hosts? 
>
>
>Thanks
>Best Regards
>
>
>On Fri, Jun 27, 2014 at 3:09 PM, MEETHU MATHEW <meethu2006@yahoo.co.in> wrote:
>
>Hi Akhil,
>>
>>
>>I am running it in a LAN itself..The IP of the master is given correctly.
>> 
>>Thanks & Regards, 
>>Meethu M
>>
>>
>>
>>On Friday, 27 June 2014 2:51 PM, Akhil Das <akhil@sigmoidanalytics.com> wrote:
>> 
>>
>>
>>why is it binding to port 0? 192.168.125.174:0 :/
>>
>>
>>Check the ip address of that master machine (ifconfig) looks like the ip address has
been changed (hoping you are running this machines on a LAN)
>>
>>
>>Thanks
>>Best Regards
>>
>>
>>On Fri, Jun 27, 2014 at 12:00 PM, MEETHU MATHEW <meethu2006@yahoo.co.in> wrote:
>>
>>Hi all,
>>>
>>>
>>>My Spark(Standalone mode) was running fine till yesterday.But now I am getting
 the following exeception when I am running start-slaves.sh or start-all.sh
>>>
>>>
>>>slave3: failed to launch org.apache.spark.deploy.worker.Worker:
>>>slave3:   at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
>>>slave3:   at java.lang.Thread.run(Thread.java:662)
>>>
>>>
>>>The log files has the following lines.
>>>
>>>
>>>14/06/27 11:06:30 INFO SecurityManager: Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
>>>14/06/27 11:06:30 INFO SecurityManager: Changing view acls to: hduser
>>>14/06/27 11:06:30 INFO SecurityManager: SecurityManager: authentication disabled;
ui acls disabled; users with view permissions: Set(hduser)
>>>14/06/27 11:06:30 INFO Slf4jLogger: Slf4jLogger started
>>>14/06/27 11:06:30 INFO Remoting: Starting remoting
>>>Exception in thread "main" org.jboss.netty.channel.ChannelException: Failed to
bind to: master/192.168.125.174:0
>>>at org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272)
>>>...
>>>Caused by: java.net.BindException: Cannot assign requested address
>>>...
>>>I saw the same error reported before and have tried the following solutions.
>>>
>>>
>>>Set the variable SPARK_LOCAL_IP ,Changed the SPARK_MASTER_PORT to a different
number..But nothing is working.
>>>
>>>
>>>When I try to start the worker from the respective machines using the following
java command,its running without any exception
>>>
>>>
>>>java -cp ::/usr/local/spark-1.0.0/conf:/usr/local/spark-1.0.0/assembly/target/scala-2.10/spark-assembly-1.0.0-hadoop1.2.1.jar
-XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true -Xms512m -Xmx512m org.apache.spark.deploy.worker.Worker
spark://:master:7077
>>>
>>>
>>>
>>>Somebody please give a solution
>>> 
>>>Thanks & Regards, 
>>>Meethu M
>>
>>
>>
>
>
>
>
>
Mime
View raw message