Are you referring to have spark picking up a new jar build?  If so, you can probably script that on bash.

Thank You,

Irving Duran


On Wed, Nov 28, 2018 at 12:44 PM Mina Aslani <aslanimina@gmail.com> wrote:
Hi,

I have a question for you.
Do we need to kill a spark job every time we change and deploy it to cluster? Or, is there a way for Spark to automatically pick up the recent jar version?

Best regards,
Mina