kuromatsu nobuyuki created SPARK-4132:
-----------------------------------------
Summary: Spark uses incompatible HDFS API
Key: SPARK-4132
URL: https://issues.apache.org/jira/browse/SPARK-4132
Project: Spark
Issue Type: Bug
Components: Spark Core
Affects Versions: 1.1.0
Environment: Spark1.1.0 on Hadoop1.2.1
CentOS 6.3 64bit
Reporter: kuromatsu nobuyuki
Priority: Minor
When I enable event logging and set it to output to HDFS, initialization fails with 'java.lang.ClassNotFoundException'
(see trace below).
I found that an API incompatibility in org.apache.hadoop.fs.permission.FsPermission between
Hadoop 1.0.4 and Hadoop 1.1.0 (and above) causes this error (org.apache.hadoop.fs.permission.FsPermission$2
is used in 1.0.4 but doesn't exist in my 1.2.1 environment).
I think that the Spark jar file pre-built for Hadoop1.X should be built on Hadoop Stable version(Hadoop
1.2.1).
2014-10-24 10:43:22,893 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 9000:
readAndProcess threw exception java.lang.RuntimeException:
readObject can't find class org.apache.hadoop.fs.permission.FsPermission$2. Count of bytes
read: 0
java.lang.RuntimeException: readObject can't find class org.apache.hadoop.fs.permission.FsPermission$2
at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:233)
at org.apache.hadoop.ipc.RPC$Invocation.readFields(RPC.java:106)
at org.apache.hadoop.ipc.Server$Connection.processData(Server.java:1347)
at org.apache.hadoop.ipc.Server$Connection.processOneRpc(Server.java:1326)
at org.apache.hadoop.ipc.Server$Connection.readAndProcess(Server.java:1226)
at org.apache.hadoop.ipc.Server$Listener.doRead(Server.java:577)
at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:384)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1146)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:701)
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.fs.permission.FsPermission$2
at java.net.URLClassLoader$1.run(URLClassLoader.java:217)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:205)
at java.lang.ClassLoader.loadClass(ClassLoader.java:323)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:294)
at java.lang.ClassLoader.loadClass(ClassLoader.java:268)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:270)
at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:810)
at org.apache.hadoop.io.ObjectWritable.readObject(ObjectWritable.java:231)
... 9 more
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org
|