spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Gourav Sengupta <gourav.sengu...@gmail.com>
Subject Re: Structured Streaming: How to add a listener for when a batch is complete
Date Wed, 04 Sep 2019 15:02:55 GMT
Hi Natalie,
in case we are using Python then listener is not available, we can use the
SPARK UI REST API's .

For that to happen we have to start the SPARK session with following
settings: spark.sql("SET spark.sql.streaming.metricsEnabled=true")
For  the UI from which we can use the JSON results from the below:
https://spark.apache.org/docs/latest/monitoring.html#rest-api

In case you are working on AWS EMR, the UI URL is a bit different, let me
know in that case and I will email to you.


Hi TD,
What will be the best way to stop a long running SPARK streaming job
gracefully? I will be really grateful if you could kindly help us out.


Regards,
Gourav Sengupta

On Wed, Sep 4, 2019 at 2:28 AM Tathagata Das <tdas@databricks.com> wrote:

>
> https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html#reporting-metrics-programmatically-using-asynchronous-apis
>
> On Tue, Sep 3, 2019, 3:26 PM Natalie Ruiz
> <Natalie.Ruiz@microsoft.com.invalid> wrote:
>
>> Hello all,
>>
>>
>>
>> I’m a beginner, new to Spark and wanted to know if there was an
>> equivalent to Spark Streaming’s StreamingListenerBatchCompleted in
>> Structured Streaming? I want to add a listener for when a batch is complete
>> but the documentation and examples I find are for Spark Streaming and not
>> Structured Streaming and registering it to a StreamingContext, which from
>> my understanding is just for Spark Streaming
>>
>>
>>
>> Thanks!
>>
>

Mime
View raw message