spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Tathagata Das <tathagata.das1...@gmail.com>
Subject Re: Build error
Date Sat, 31 Jan 2015 05:53:53 GMT
That is a known issue uncovered last week. It fails on certain
environments, not on Jenkins which is our testing environment.
There is already a PR up to fix it. For now you can build using "mvn
package -DskipTests"
TD

On Fri, Jan 30, 2015 at 8:59 PM, Andrew Musselman <
andrew.musselman@gmail.com> wrote:

> Off master, got this error; is that typical?
>
> -------------------------------------------------------
>  T E S T S
> -------------------------------------------------------
> Running org.apache.spark.streaming.mqtt.JavaMQTTStreamSuite
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.495 sec
> - in org.apache.spark.streaming.mqtt.JavaMQTTStreamSuite
>
> Results :
>
>
>
>
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0
>
> [INFO]
> [INFO] --- scalatest-maven-plugin:1.0:test (test) @
> spark-streaming-mqtt_2.10 ---
> Discovery starting.
> Discovery completed in 498 milliseconds.
> Run starting. Expected test count is: 1
> MQTTStreamSuite:
> - mqtt input stream *** FAILED ***
>   org.eclipse.paho.client.mqttv3.MqttException: Too many publishes in
> progress
>   at
> org.eclipse.paho.client.mqttv3.internal.ClientState.send(ClientState.java:432)
>   at
> org.eclipse.paho.client.mqttv3.internal.ClientComms.internalSend(ClientComms.java:121)
>   at
> org.eclipse.paho.client.mqttv3.internal.ClientComms.sendNoWait(ClientComms.java:139)
>   at org.eclipse.paho.client.mqttv3.MqttTopic.publish(MqttTopic.java:107)
>   at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$publishData$1.apply(MQTTStreamSuite.scala:125)
>   at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$publishData$1.apply(MQTTStreamSuite.scala:124)
>   at scala.collection.immutable.Range.foreach(Range.scala:141)
>   at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite.publishData(MQTTStreamSuite.scala:124)
>   at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$3.apply$mcV$sp(MQTTStreamSuite.scala:78)
>   at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$3.apply(MQTTStreamSuite.scala:66)
>   ...
> Exception in thread "Thread-20" org.apache.spark.SparkException: Job
> cancelled because SparkContext was shut down
>     at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:690)
>     at
> org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:689)
>     at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
>     at
> org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:689)
>     at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onStop(DAGScheduler.scala:1384)
>     at org.apache.spark.util.EventLoop.stop(EventLoop.scala:81)
>     at
> org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1319)
>     at org.apache.spark.SparkContext.stop(SparkContext.scala:1250)
>     at
> org.apache.spark.streaming.StreamingContext.stop(StreamingContext.scala:510)
>     at
> org.apache.spark.streaming.StreamingContext.stop(StreamingContext.scala:485)
>     at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$2.apply$mcV$sp(MQTTStreamSuite.scala:59)
>     at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$2.apply(MQTTStreamSuite.scala:57)
>     at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite$$anonfun$2.apply(MQTTStreamSuite.scala:57)
>     at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:210)
>     at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite.runTest(MQTTStreamSuite.scala:38)
>     at
> org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
>     at
> org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
>     at
> org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
>     at
> org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
>     at scala.collection.immutable.List.foreach(List.scala:318)
>     at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
>     at org.scalatest.SuperEngine.org
> $scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
>     at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
>     at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
>     at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
>     at org.scalatest.Suite$class.run(Suite.scala:1424)
>     at org.scalatest.FunSuite.org
> $scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
>     at
> org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
>     at
> org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
>     at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
>     at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
>     at org.apache.spark.streaming.mqtt.MQTTStreamSuite.org
> $scalatest$BeforeAndAfter$$super$run(MQTTStreamSuite.scala:38)
>     at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:241)
>     at
> org.apache.spark.streaming.mqtt.MQTTStreamSuite.run(MQTTStreamSuite.scala:38)
>     at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
>     at
> org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
>     at
> org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
>     at
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>     at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
>     at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
>     at
> org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
>     at org.scalatest.Suite$class.run(Suite.scala:1421)
>     at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
>     at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
>     at
> org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
>     at
> org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
>     at scala.collection.immutable.List.foreach(List.scala:318)
>     at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
>     at
> org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
>     at
> org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
>     at
> org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
>     at
> org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
>     at org.scalatest.tools.Runner$.main(Runner.scala:860)
>     at org.scalatest.tools.Runner.main(Runner.scala)
> Run completed in 16 seconds, 340 milliseconds.
> Total number of tests run: 1
> Suites: completed 2, aborted 0
> Tests: succeeded 0, failed 1, canceled 0, ignored 0, pending 0
> *** 1 TEST FAILED ***
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] Spark Project Parent POM .......................... SUCCESS [4.123s]
> [INFO] Spark Project Networking .......................... SUCCESS [4.948s]
> [INFO] Spark Project Shuffle Streaming Service ........... SUCCESS [4.045s]
> [INFO] Spark Project Core ................................ SUCCESS
> [1:58:19.003s]
> [INFO] Spark Project Bagel ............................... SUCCESS
> [28.989s]
> [INFO] Spark Project GraphX .............................. SUCCESS
> [2:09.803s]
> [INFO] Spark Project Streaming ........................... SUCCESS
> [5:27.931s]
> [INFO] Spark Project Catalyst ............................ SUCCESS
> [2:23.327s]
> [INFO] Spark Project SQL ................................. SUCCESS
> [2:43.832s]
> [INFO] Spark Project ML Library .......................... SUCCESS
> [9:31.367s]
> [INFO] Spark Project Tools ............................... SUCCESS
> [10.819s]
> [INFO] Spark Project Hive ................................ SUCCESS
> [3:45.769s]
> [INFO] Spark Project REPL ................................ SUCCESS
> [1:41.802s]
> [INFO] Spark Project Assembly ............................ SUCCESS [1.822s]
> [INFO] Spark Project External Twitter .................... SUCCESS
> [21.989s]
> [INFO] Spark Project External Flume Sink ................. SUCCESS
> [24.345s]
> [INFO] Spark Project External Flume ...................... SUCCESS
> [39.686s]
> [INFO] Spark Project External MQTT ....................... FAILURE
> [37.047s]
> [INFO] Spark Project External ZeroMQ ..................... SKIPPED
> [INFO] Spark Project External Kafka ...................... SKIPPED
> [INFO] Spark Project Examples ............................ SKIPPED
> [INFO]
> ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO]
> ------------------------------------------------------------------------
> [INFO] Total time: 2:29:01.232s
> [INFO] Finished at: Fri Jan 30 19:16:43 PST 2015
> [INFO] Final Memory: 104M/1005M
> [INFO]
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal
> org.scalatest:scalatest-maven-plugin:1.0:test (test) on project
> spark-streaming-mqtt_2.10: There are test failures -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the
> -e switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions,
> please read the following articles:
> [ERROR] [Help 1]
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the
> command
> [ERROR]   mvn <goals> -rf :spark-streaming-mqtt_2.10
>
>

Mime
View raw message