spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Chris Westin <cwes...@yahoo.com.INVALID>
Subject Job submission failure exception - unserializable TaskEndReason
Date Thu, 10 Mar 2016 20:41:39 GMT
I'm getting an exception when I try to submit a job (through prediction.io, if you know it):
[INFO] [Runner$] Submission command: /home/pio/PredictionIO/vendors/spark-1.5.1/bin/spark-submit
--class io.prediction.tools.imprt.FileToEvents --files file:/home/pio/PredictionIO/conf/log4j.properties,file:/home/pio/PredictionIO/vendors/hbase-1.0.0/conf/hbase-site.xml
--driver-class-path /home/pio/PredictionIO/conf:/home/pio/PredictionIO/vendors/hbase-1.0.0/conf
file:/home/pio/PredictionIO/lib/pio-assembly-0.9.5.jar --appid 1 --input file:/home/pio/pio-ur/my_events.json
--env PIO_STORAGE_SOURCES_HBASE_TYPE=hbase,PIO_ENV_LOADED=1,PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta,PIO_FS_BASEDIR=/home/pio/.pio_store,PIO_STORAGE_SOURCES_HBASE_HOME=/home/pio/PredictionIO/vendors/hbase-1.0.0,PIO_HOME=/home/pio/PredictionIO,PIO_FS_ENGINESDIR=/home/pio/.pio_store/engines,PIO_STORAGE_SOURCES_LOCALFS_PATH=/home/pio/.pio_store/models,PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch,PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICSEARCH,PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=LOCALFS,PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event,PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/home/pio/PredictionIO/vendors/elasticsearch-1.4.4,PIO_FS_TMPDIR=/home/pio/.pio_store/tmp,PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model,PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE,PIO_CONF_DIR=/home/pio/PredictionIO/conf,PIO_STORAGE_SOURCES_LOCALFS_TYPE=localfs[INFO]
[Remoting] Starting remoting[INFO] [Remoting] Remoting started; listening on addresses :[akka.tcp://sparkDriver@192.168.56.150:52838][WARN]
[MetricsSystem] Using default name DAGScheduler for source because spark.app.id is not set.[ERROR]
[Executor] Exception in task 0.0 in stage 0.0 (TID 0)[WARN] [ThrowableSerializationWrapper]
Task exception could not be deserialized[ERROR] [TaskResultGetter] Could not deserialize TaskEndReason:
ClassNotFound with classloader org.apache.spark.util.MutableURLClassLoader@59e5ddf[WARN] [TaskSetManager]
Lost task 0.0 in stage 0.0 (TID 0, localhost): UnknownReason[ERROR] [TaskSetManager] Task
0 in stage 0.0 failed 1 times; aborting jobException in thread "main" org.apache.spark.SparkException:
Job aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent failure:
Lost task 0.0 in stage 0.0 (TID 0, localhost): UnknownReasonDriver stacktrace: at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270) at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at scala.Option.foreach(Option.scala:236) at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1822) at org.apache.spark.SparkContext.runJob(SparkContext.scala:1835)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1912) at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1055)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:306) at org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:998)
at io.prediction.data.storage.hbase.HBPEvents.write(HBPEvents.scala:108) at io.prediction.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:90)
at io.prediction.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:63) at
scala.Option.map(Option.scala:145) at io.prediction.tools.imprt.FileToEvents$.main(FileToEvents.scala:63)
at io.prediction.tools.imprt.FileToEvents.main(FileToEvents.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)[pio@allinone pio-ur]$
Hoping to get some hints what's causing this, or at least how to get the TaskEndReason class
added to the class path in case it can provide more information.
Thanks!


Mime
View raw message