something like this import org.apache.spark.TaskContext ds.map(r => { val taskContext = TaskContext.get() if (taskContext.partitionId == 1000) { throw new RuntimeException } r }) On Mon, Feb 11, 2019 at 8:41 AM Serega Sheypak wrote: > > I need to crash task which does repartition. > > пн, 11 февр. 2019 г. в 10:37, Gabor Somogyi : >> >> What blocks you to put if conditions inside the mentioned map function? >> >> On Mon, Feb 11, 2019 at 10:31 AM Serega Sheypak wrote: >>> >>> Yeah, but I don't need to crash entire app, I want to fail several tasks or executors and then wait for completion. >>> >>> вс, 10 февр. 2019 г. в 21:49, Gabor Somogyi : >>>> >>>> Another approach is adding artificial exception into the application's source code like this: >>>> >>>> val query = input.toDS.map(_ / 0).writeStream.format("console").start() >>>> >>>> G >>>> >>>> >>>> On Sun, Feb 10, 2019 at 9:36 PM Serega Sheypak wrote: >>>>> >>>>> Hi BR, >>>>> thanks for your reply. I want to mimic the issue and kill tasks at a certain stage. Killing executor is also an option for me. >>>>> I'm curious how do core spark contributors test spark fault tolerance? >>>>> >>>>> >>>>> вс, 10 февр. 2019 г. в 16:57, Gabor Somogyi : >>>>>> >>>>>> Hi Serega, >>>>>> >>>>>> If I understand your problem correctly you would like to kill one executor only and the rest of the app has to be untouched. >>>>>> If that's true yarn -kill is not what you want because it stops the whole application. >>>>>> >>>>>> I've done similar thing when tested/testing Spark's HA features. >>>>>> - jps -vlm | grep "org.apache.spark.executor.CoarseGrainedExecutorBackend.*applicationid" >>>>>> - kill -9 pidofoneexecutor >>>>>> >>>>>> Be aware if it's a multi-node cluster check whether at least one process runs on a specific node(it's not required). >>>>>> Happy killing... >>>>>> >>>>>> BR, >>>>>> G >>>>>> >>>>>> >>>>>> On Sun, Feb 10, 2019 at 4:19 PM Jörn Franke wrote: >>>>>>> >>>>>>> yarn application -kill applicationid ? >>>>>>> >>>>>>> > Am 10.02.2019 um 13:30 schrieb Serega Sheypak : >>>>>>> > >>>>>>> > Hi there! >>>>>>> > I have weird issue that appears only when tasks fail at specific stage. I would like to imitate failure on my own. >>>>>>> > The plan is to run problematic app and then kill entire executor or some tasks when execution reaches certain stage. >>>>>>> > >>>>>>> > Is it do-able? >>>>>>> >>>>>>> --------------------------------------------------------------------- >>>>>>> To unsubscribe e-mail: user-unsubscribe@spark.apache.org >>>>>>> -- Sent from my iPhone --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscribe@spark.apache.org