Please consider the following scenario.
I've started the spark master by invoking the org.apache.spark.deploy.master.Master.startSystemAndActor method in a java code and connected a worker to it using the org.apache.spark.deploy.worker.Worker.startSystemAndActor method. and then I have successfully created a java spark & SQL contexts and performed SQL queries.
My question is, can I change this order?
Can I start the master first, then create a spark context... and later on connect a worker to the master?
While trying out this scenario, I have successfully started the master. Please see the screenshot here.
But when I create an spark context, it terminates automatically. is it because the master not being connected to a worker?