spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From A Shaikh <>
Subject Re: Running a spark code using submit job in google cloud platform
Date Thu, 12 Jan 2017 17:39:46 GMT
You may have tested this code on Spark version on your local machine
version of which may be different to whats in Google Cloud Storage.
You need to select appropraite Spark version when you submit your job.

On 12 January 2017 at 15:51, Anahita Talebi <>

> Dear all,
> I am trying to run a .jar file as a job using submit job in google cloud
> console.
> I actually ran the spark code on my local computer to generate a .jar
> file. Then in the Argument folder, I give the value of the arguments that I
> used in the spark code. One of the argument is training data set that I put
> in the same bucket that I save my .jar file. In the bucket, I put only the
> .jar file, training dataset and testing dataset.
> Main class or jar
> gs://Anahita/test.jar
> Arguments
> --lambda=.001
> --eta=1.0
> --trainFile=gs://Anahita/small_train.dat
> --testFile=gs://Anahita/small_test.dat
> The problem is that when I run the job I get the following error and
> actually it cannot read  my training and testing data sets.
> Exception in thread "main" java.lang.NoSuchMethodError: org.apache.spark.rdd.RDD.coalesce(IZLscala/math/Ordering;)Lorg/apache/spark/rdd/RDD;
> Can anyone help me how I can solve this problem?
> Thanks,
> Anahita

View raw message