Wouldn't it likely be the opposite? Too much memory / too many cores being requested relative to the resource that YARN makes available?

On Nov 24, 2014 11:00 AM, "Akhil Das" <akhil@sigmoidanalytics.com> wrote:
This can happen mainly because of the following:

- Wrong master url (Make sure you give the master url which is listed on top left corner of the webui - running on 8080)

- Allocated more memory/cores while creating the sparkContext.



Thanks
Best Regards

On Mon, Nov 24, 2014 at 4:13 PM, vdiwakar.malladi <vdiwakar.malladi@gmail.com> wrote:
Hi,

When i trying to execute the program from my laptop by connecting to HDP
environment (on which Spark also configured), i'm getting the warning
("Initial job has not accepted any resources; check your cluster UI to
ensure that workers are registered and have sufficient memory") and Job is
being terminated. My console has following log statements.

Note: I could able to run the same client program by using spark-submit
command. Whatever parameters i passed to spark-submit command, i passed the
same to to SparkConf object. But still getting the same error. Any clue on
this?

14/11/24 16:07:09 INFO scheduler.DAGScheduler: Submitting 2 missing tasks
from Stage 0 (MappedRDD[4] at map at JavaSchemaRDD.scala:42)
14/11/24 16:07:09 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with
2 tasks
14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
app-20141124023636-0004/0 is now EXITED (Command exited with code 1)
14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20141124023636-0004/0 removed: Command exited with code 1
14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added:
app-20141124023636-0004/2 on worker-20141124021958-STI-SM-DEV-SYS4-51561
(STI-SM-DEV-SYS4:51561) with 4 cores
14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20141124023636-0004/2 on hostPort STI-SM-DEV-SYS4:51561 with 4 cores,
8.0 GB RAM
14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
app-20141124023636-0004/1 is now EXITED (Command exited with code 1)
14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20141124023636-0004/1 removed: Command exited with code 1
14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor added:
app-20141124023636-0004/3 on worker-20141124022001-STI-SM-DEV-SYS5-50404
(STI-SM-DEV-SYS5:50404) with 4 cores
14/11/24 16:07:09 INFO cluster.SparkDeploySchedulerBackend: Granted executor
ID app-20141124023636-0004/3 on hostPort STI-SM-DEV-SYS5:50404 with 4 cores,
8.0 GB RAM
14/11/24 16:07:09 INFO client.AppClient$ClientActor: Executor updated:
app-20141124023636-0004/2 is now RUNNING
14/11/24 16:07:10 INFO client.AppClient$ClientActor: Executor updated:
app-20141124023636-0004/3 is now RUNNING
14/11/24 16:07:24 WARN scheduler.TaskSchedulerImpl: Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient memory
14/11/24 16:07:39 WARN scheduler.TaskSchedulerImpl: Initial job has not
accepted any resources; check your cluster UI to ensure that workers are
registered and have sufficient memory
14/11/24 16:07:43 INFO client.AppClient$ClientActor: Executor updated:
app-20141124023636-0004/3 is now EXITED (Command exited with code 1)
14/11/24 16:07:43 INFO cluster.SparkDeploySchedulerBackend: Executor
app-20141124023636-0004/3 removed: Command exited with code 1

Thanks in advance.



--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/issue-while-running-the-code-in-standalone-mode-Initial-job-has-not-accepted-any-resources-check-you-tp19628.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org