spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Matei Zaharia <matei.zaha...@gmail.com>
Subject Re: NO SUCH METHOD EXCEPTION
Date Tue, 11 Mar 2014 19:04:07 GMT
Since it’s from Scala, it might mean you’re running with a different version of Scala than
you compiled Spark with. Spark 0.8 and earlier use Scala 2.9, while Spark 0.9 uses Scala 2.10.

Matei

On Mar 11, 2014, at 8:19 AM, Jeyaraj, Arockia R (Arockia) <arockia.r.jeyaraj@verizon.com>
wrote:

> Hi,
>  
> Can anyone help me to resolve this issue? Why am I getting NoSuchMethod exception?
>  
> 14/03/11 09:56:11 ERROR executor.Executor: Exception in task ID 0
> java.lang.NoSuchMethodError: scala.Predef$.augmentString(Ljava/lang/String;)Lsca
> la/collection/immutable/StringOps;
> at kafka.utils.VerifiableProperties.getIntInRange(VerifiableProperties.s
> cala:75)
> at kafka.utils.VerifiableProperties.getInt(VerifiableProperties.scala:58
> )
> at kafka.utils.ZKConfig.<init>(ZkUtils.scala:837)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:73)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:77)
> at org.apache.spark.streaming.kafka.KafkaReceiver.onStart(KafkaInputDStr
> eam.scala:98)
> at org.apache.spark.streaming.dstream.NetworkReceiver.start(NetworkInput
> DStream.scala:126)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:173)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:169)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109)
> at org.apache.spark.scheduler.Task.run(Task.scala:53)
> at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mc
> V$sp(Executor.scala:213)
> at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.sca
> la:49)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
> 
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExec
> utor.java:886)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor
> .java:908)
> at java.lang.Thread.run(Thread.java:619)
> 14/03/11 09:56:11 WARN scheduler.TaskSetManager: Lost TID 0 (task 0.0:0)
> 14/03/11 09:56:11 WARN scheduler.TaskSetManager: Loss was due to java.lang.NoSuc
> hMethodError
> java.lang.NoSuchMethodError: scala.Predef$.augmentString(Ljava/lang/String;)Lsca
> la/collection/immutable/StringOps;
> at kafka.utils.VerifiableProperties.getIntInRange(VerifiableProperties.s
> cala:75)
> at kafka.utils.VerifiableProperties.getInt(VerifiableProperties.scala:58
> )
> at kafka.utils.ZKConfig.<init>(ZkUtils.scala:837)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:73)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:77)
> at org.apache.spark.streaming.kafka.KafkaReceiver.onStart(KafkaInputDStr
> eam.scala:98)
> at org.apache.spark.streaming.dstream.NetworkReceiver.start(NetworkInput
> DStream.scala:126)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:173)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:169)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109)
> at org.apache.spark.scheduler.Task.run(Task.scala:53)
> at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mc
> V$sp(Executor.scala:213)
> at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.sca
> la:49)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
> 
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExec
> utor.java:886)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor
> .java:908)
> at java.lang.Thread.run(Thread.java:619)
> 14/03/11 09:56:11 ERROR scheduler.TaskSetManager: Task 0.0:0 failed 1 times; abo
> rting job
> 14/03/11 09:56:11 INFO scheduler.TaskSchedulerImpl: Remove TaskSet 0.0 from pool
> 
> 14/03/11 09:56:11 INFO scheduler.DAGScheduler: Failed to run runJob at NetworkIn
> putTracker.scala:182
> [error] (Thread-34) org.apache.spark.SparkException: Job aborted: Task 0.0:0 fai
> 
>  
>  
> Thanks
> Arockia Raja


Mime
View raw message