spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From WangTaoTheTonic <barneystin...@aliyun.com>
Subject Corrupted Exception while deserialize task
Date Thu, 25 Dec 2014 14:27:30 GMT
Hi Guys, 

I found an excetpion while running application using 1.2.0-snapshot version.
It shows like this:

2014-12-23 07:45:36,333 | ERROR | [Executor task launch worker-0] |
Exception in task 0.0 in stage 0.0 (TID 0) |
org.apache.spark.Logging$class.logError(Logging.scala:96)
java.io.StreamCorruptedException: invalid stream header: 00546864
	at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:804)
	at java.io.ObjectInputStream.<init>(ObjectInputStream.java:299)
	at
org.apache.spark.serializer.JavaDeserializationStream$$anon$1.<init>(JavaSerializer.scala:57)
	at
org.apache.spark.serializer.JavaDeserializationStream.<init>(JavaSerializer.scala:57)
	at
org.apache.spark.serializer.JavaSerializerInstance.deserializeStream(JavaSerializer.scala:99)
	at
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:86)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:163)
	at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:745)
2014-12-23 07:45:36,357 | INFO  |
[sparkExecutor-akka.actor.default-dispatcher-3] | Got assigned task 1 |
org.apache.spark.Logging$class.logInfo(Logging.scala:59)
2014-12-23 07:45:36,358 | INFO  | [Executor task launch worker-0] | Running
task 1.0 in stage 0.0 (TID 1) |
org.apache.spark.Logging$class.logInfo(Logging.scala:59)
2014-12-23 07:45:36,414 | ERROR | [Executor task launch worker-0] |
Exception in task 1.0 in stage 0.0 (TID 1) |
org.apache.spark.Logging$class.logError(Logging.scala:96)
java.io.StreamCorruptedException: invalid stream header: 00546864
	at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:804)
	at java.io.ObjectInputStream.<init>(ObjectInputStream.java:299)
	at
org.apache.spark.serializer.JavaDeserializationStream$$anon$1.<init>(JavaSerializer.scala:57)
	at
org.apache.spark.serializer.JavaDeserializationStream.<init>(JavaSerializer.scala:57)
	at
org.apache.spark.serializer.JavaSerializerInstance.deserializeStream(JavaSerializer.scala:99)
	at
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:86)
	at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:163)
	at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:745)

I know it happened while executor deserialize task. But after checking the
spark code, I found components of one task is very simple: its files, jars
and an Task object contains stageId and partitionId.

I cann't confirm what cause this issue and it is hard to reproduce it.

But I think the application code does not make difference as code segment
here is gransparent to users.

Anyone have some ideas? Thanks for offering help.

P.S. This error occured in every executor of this application.



--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Corrupted-Exception-while-deserialize-task-tp20857.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscribe@spark.apache.org
For additional commands, e-mail: user-help@spark.apache.org


Mime
View raw message