spark-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From 周千昊 <qhz...@apache.org>
Subject Re: please help with ClassNotFoundException
Date Fri, 14 Aug 2015 06:21:53 GMT
Hi, Sea
     Problem solved, it turn out to be that I have updated spark cluster to
1.4.1, however the client has not been updated.
     Thank you so much.

Sea <261810726@qq.com>于2015年8月14日周五 下午1:01写道:

> I have no idea... We use scala. You upgrade to 1.4 so quickly...,  are you
> using spark in production?  Spark 1.3 is better than spark1.4.
>
> ------------------ 原始邮件 ------------------
> *发件人:* "周千昊";<z.qianhao@gmail.com>;
> *发送时间:* 2015年8月14日(星期五) 中午11:14
> *收件人:* "Sea"<261810726@qq.com>; "dev@spark.apache.org"<
> dev@spark.apache.org>;
> *主题:* Re: please help with ClassNotFoundException
>
> Hi Sea
>      I have updated spark to 1.4.1, however the problem still exists, any
> idea?
>
> Sea <261810726@qq.com>于2015年8月14日周五 上午12:36写道:
>
>> Yes, I guess so. I see this bug before.
>>
>>
>> ------------------ 原始邮件 ------------------
>> *发件人:* "周千昊";<z.qianhao@gmail.com>;
>> *发送时间:* 2015年8月13日(星期四) 晚上9:30
>> *收件人:* "Sea"<261810726@qq.com>; "dev@spark.apache.org"<
>> dev@spark.apache.org>;
>> *主题:* Re: please help with ClassNotFoundException
>>
>> Hi sea
>>     Is it the same issue as
>> https://issues.apache.org/jira/browse/SPARK-8368
>>
>> Sea <261810726@qq.com>于2015年8月13日周四 下午6:52写道:
>>
>>> Are you using 1.4.0?  If yes, use 1.4.1
>>>
>>>
>>> ------------------ 原始邮件 ------------------
>>> *发件人:* "周千昊";<qhzhou@apache.org>;
>>> *发送时间:* 2015年8月13日(星期四) 晚上6:04
>>> *收件人:* "dev"<dev@spark.apache.org>;
>>> *主题:* please help with ClassNotFoundException
>>>
>>> Hi,
>>>     I am using spark 1.4 when an issue occurs to me.
>>>     I am trying to use the aggregate function:
>>>     JavaRdd<String> rdd = some rdd;
>>>     HashMap<Long, TypeA> zeroValue = new HashMap();
>>>     // add initial key-value pair for zeroValue
>>>     rdd.aggregate(zeroValue,
>>>                    new Function2<HashMap<Long, TypeA>,
>>>                         String,
>>>                         HashMap<Long, TypeA>>(){//implementation},
>>>                    new Function2<HashMap<Long, TypeA>,
>>>                         String,
>>>                         HashMap<Long, TypeA>(){//implementation})
>>>
>>>     here is the stack trace when i run the application:
>>>
>>> Caused by: java.lang.ClassNotFoundException: TypeA
>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
>>> at java.security.AccessController.doPrivileged(Native Method)
>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>> at java.lang.Class.forName0(Native Method)
>>> at java.lang.Class.forName(Class.java:274)
>>> at
>>> org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:66)
>>> at
>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612)
>>> at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>> at
>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>> at java.util.HashMap.readObject(HashMap.java:1180)
>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> at java.lang.reflect.Method.invoke(Method.java:606)
>>> at
>>> java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
>>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
>>> at
>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>> at
>>> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:69)
>>> at
>>> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:89)
>>> at org.apache.spark.util.Utils$.clone(Utils.scala:1458)
>>> at org.apache.spark.rdd.RDD$$anonfun$aggregate$1.apply(RDD.scala:1049)
>>> at
>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:148)
>>> at
>>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:109)
>>> at org.apache.spark.rdd.RDD.withScope(RDD.scala:286)
>>> at org.apache.spark.rdd.RDD.aggregate(RDD.scala:1047)
>>> at
>>> org.apache.spark.api.java.JavaRDDLike$class.aggregate(JavaRDDLike.scala:413)
>>> at
>>> org.apache.spark.api.java.AbstractJavaRDDLike.aggregate(JavaRDDLike.scala:47)
>>>      *however I have checked that TypeA is in the jar file which is in
>>> the classpath*
>>> *    And when I use an empty HashMap as the zeroValue, the exception has
>>> gone*
>>> *    Does anyone meet the same problem, or can anyone help me with it?*
>>>
>> --
>> Best Regard
>> ZhouQianhao
>>
> --
> Best Regard
> ZhouQianhao
>

Mime
View raw message