I would check the queue you are submitting job, assuming it is yarn...

On Tue, Sep 26, 2017 at 11:40 PM, JG Perrin <jperrin@lumeris.com> wrote:



I get the infamous:

Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources


I run the app via Eclipse, connecting:

        SparkSession spark = SparkSession.builder()

                .appName("Converter - Benchmark")


                .config("spark.cores.max", "4")

                .config("spark.executor.memory", "16g")




Everything seems ok on the cluster side:



I probably missed something super obvious, but can’t find it…


Any help/hint is welcome! - TIA






Best Regards,
Ayan Guha