spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Adaryl Wakefield <adaryl.wakefi...@hotmail.com>
Subject RE: finding Spark Master
Date Wed, 08 Mar 2017 01:53:24 GMT
Ah so I see setMaster(‘yarn-client’). Hmm.

What I was ultimately trying to do was develop with Eclipse on my windows box and have the
code point to my cluster so it executes there instead of my local windows machine. Perhaps
I’m going about this wrong.

Adaryl "Bob" Wakefield, MBA
Principal
Mass Street Analytics, LLC
913.938.6685
www.massstreet.net<http://www.massstreet.net>
www.linkedin.com/in/bobwakefieldmba<http://www.linkedin.com/in/bobwakefieldmba>
Twitter: @BobLovesData

From: Koert Kuipers [mailto:koert@tresata.com]
Sent: Tuesday, March 7, 2017 7:47 PM
To: Adaryl Wakefield <adaryl.wakefield@hotmail.com>
Cc: user@spark.apache.org
Subject: Re: finding Spark Master

assuming this is running on yarn there is really spark-master. every job created its own "master"
within a yarn application.

On Tue, Mar 7, 2017 at 6:27 PM, Adaryl Wakefield <adaryl.wakefield@hotmail.com<mailto:adaryl.wakefield@hotmail.com>>
wrote:
I’m running a three node cluster along with Spark along with Hadoop as part of a HDP stack.
How do I find my Spark Master? I’m just seeing the clients. I’m trying to figure out what
goes in setMaster() aside from local[*].

Adaryl "Bob" Wakefield, MBA
Principal
Mass Street Analytics, LLC
913.938.6685<tel:(913)%20938-6685>
www.massstreet.net<http://www.massstreet.net>
www.linkedin.com/in/bobwakefieldmba<http://www.linkedin.com/in/bobwakefieldmba>
Twitter: @BobLovesData


Mime
View raw message