Are you looking for RDD.wholeTextFiles?

On 3 August 2015 at 10:57, Spark Enthusiast <> wrote:
All  examples of Spark Stream programming that I see assume streams of lines that are then tokenised and acted upon (like the WordCount example).

How do I process Streams that span multiple lines? Are there examples that I can use?