spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Akhil Das <ak...@sigmoidanalytics.com>
Subject Re: issue while running the code in standalone mode: "Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory"
Date Mon, 24 Nov 2014 10:59:47 GMT
This can happen mainly because of the following:

- Wrong master url (Make sure you give the master url which is listed on
top left corner of the webui - running on 8080)

- Allocated more memory/cores while creating the sparkContext.



Thanks
Best Regards

On Mon, Nov 24, 2014 at 4:13 PM, vdiwakar.malladi <
vdiwakar.malladi@gmail.com> wrote:

> Hi,
>
> When i trying to execute the program from my laptop by connecting to HDP
> environment (on which Spark also configured), i'm getting the warning
> ("Initial job has not accepted any resources; check your cluster UI to
> ensure that workers are registered and have sufficient memory") and Job is
> being terminated. My console has following log statements.
>
> Note: I could able to run the same client program by using spark-submit
> command. Whatever parameters i passed to spark-submit command, i passed the
> same to to SparkConf object. But still getting the same error. Any clue on
> this?
>
> 14/11/24 16:07:09 INFO scheduler.DAGScheduler: Submitting 2 missing tasks
> from Stage 0 (MappedRDD[4] at map at JavaSchemaRDD.scala:42)
> 14/11/24 16:07:09 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0
> with
> 2 tasks
> 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
> app-20141124023636-0004/0 is now EXITED (Command exited with code 1)
> 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20141124023636-0004/0 removed: Command exited with code 1
> 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added:
> app-20141124023636-0004/2 on worker-20141124021958-STI-SM-DEV-SYS4-51561
> (STI-SM-DEV-SYS4:51561) with 4 cores
> 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20141124023636-0004/2 on hostPort STI-SM-DEV-SYS4:51561 with 4
> cores,
> 8.0 GB RAM
> 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
> app-20141124023636-0004/1 is now EXITED (Command exited with code 1)
> 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20141124023636-0004/1 removed: Command exited with code 1
> 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added:
> app-20141124023636-0004/3 on worker-20141124022001-STI-SM-DEV-SYS5-50404
> (STI-SM-DEV-SYS5:50404) with 4 cores
> 14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted
> executor
> ID app-20141124023636-0004/3 on hostPort STI-SM-DEV-SYS5:50404 with 4
> cores,
> 8.0 GB RAM
> 14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
> app-20141124023636-0004/2 is now RUNNING
> 14/11/24 16:07:10 INFO client.AppClient$ClientActor: Executor updated:
> app-20141124023636-0004/3 is now RUNNING
> 14/11/24 16:07:24 WARN scheduler.TaskSchedulerImpl: Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient memory
> 14/11/24 16:07:39 WARN scheduler.TaskSchedulerImpl: Initial job has not
> accepted any resources; check your cluster UI to ensure that workers are
> registered and have sufficient memory
> 14/11/24 16:07:43 INFO client.AppClient$ClientActor: Executor updated:
> app-20141124023636-0004/3 is now EXITED (Command exited with code 1)
> 14/11/24 16:07:43 INFO cluster.SparkDeploySchedulerBackend: Executor
> app-20141124023636-0004/3 removed: Command exited with code 1
>
> Thanks in advance.
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/issue-while-running-the-code-in-standalone-mode-Initial-job-has-not-accepted-any-resources-check-you-tp19628.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
> For additional commands, e-mail: user-help@spark.apache.org
>
>

Mime
View raw message