carbondata-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: carbondata-master-spark-2.1 » Apache CarbonData :: Examples #3284
Date Wed, 02 Jan 2019 08:02:13 GMT
See <https://builds.apache.org/job/carbondata-master-spark-2.1/org.apache.carbondata$carbondata-examples/3284/display/redirect>

------------------------------------------
[...truncated 271.49 KB...]
|      40|
+--------+

2019-01-02 08:00:46 AUDIT audit:72 - {"time":"January 2, 2019 12:00:46 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173127523117489","opStatus":"START"}
2019-01-02 08:00:46 ERROR DataLoadExecutor:55 - Data Load is partially success for table origin_table
2019-01-02 08:00:47 AUDIT audit:93 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173127523117489","opStatus":"SUCCESS","opTime":"187 ms","table":"default.origin_table","extraInfo":{"SegmentId":"4","DataSize":"2.85KB","IndexSize":"1.38KB"}}
2019-01-02 08:00:47 AUDIT audit:72 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173127738583917","opStatus":"START"}
2019-01-02 08:00:47 ERROR DataLoadExecutor:55 - Data Load is partially success for table origin_table
2019-01-02 08:00:47 AUDIT audit:93 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173127738583917","opStatus":"SUCCESS","opTime":"186 ms","table":"default.origin_table","extraInfo":{"SegmentId":"5","DataSize":"2.85KB","IndexSize":"1.38KB"}}
+--------+
|count(1)|
+--------+
|      60|
+--------+

2019-01-02 08:00:47 AUDIT audit:72 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"DROP
TABLE","opId":"6173127984431170","opStatus":"START"}
2019-01-02 08:00:47 AUDIT audit:93 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"DROP
TABLE","opId":"6173127984431170","opStatus":"SUCCESS","opTime":"62 ms","table":"default.origin_table","extraInfo":{}}
- ExternalTableExample
2019-01-02 08:00:47 ERROR DataLoadExecutor:55 - Data Load is partially success for table _tempTable_6173128178460653

Data:
java.lang.NullPointerException
	at org.apache.carbondata.examples.sdk.CarbonReaderExample.main(CarbonReaderExample.java:118)
	at org.apache.carbondata.examplesCI.RunExamples$$anonfun$19.apply$mcV$sp(RunExamples.scala:126)
	at org.apache.carbondata.examplesCI.RunExamples$$anonfun$19.apply(RunExamples.scala:126)
	at org.apache.carbondata.examplesCI.RunExamples$$anonfun$19.apply(RunExamples.scala:126)
	at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
	at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
	at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
	at org.scalatest.Transformer.apply(Transformer.scala:22)
	at org.scalatest.Transformer.apply(Transformer.scala:20)
	at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
	at org.apache.spark.sql.test.util.CarbonFunSuite.withFixture(CarbonFunSuite.scala:41)
	at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
	at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
	at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
	at scala.collection.immutable.List.foreach(List.scala:381)
	at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
	at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
	at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
	at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
	at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
	at org.scalatest.Suite$class.run(Suite.scala:1424)
	at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
	at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
	at org.apache.carbondata.examplesCI.RunExamples.org$scalatest$BeforeAndAfterAll$$super$run(RunExamples.scala:35)
	at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
	at org.apache.carbondata.examplesCI.RunExamples.run(RunExamples.scala:35)
	at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
	at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
	at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
	at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
	at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
	at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
	at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
	at org.scalatest.Suite$class.run(Suite.scala:1421)
	at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
	at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
	at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
	at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
	at scala.collection.immutable.List.foreach(List.scala:381)
	at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
	at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
	at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
	at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
	at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
	at org.scalatest.tools.Runner$.main(Runner.scala:860)
	at org.scalatest.tools.Runner.main(Runner.scala)
null
- CarbonReaderExample
2019-01-02 08:00:47 AUDIT audit:72 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"CREATE
TABLE","opId":"6173128343752924","opStatus":"START"}
2019-01-02 08:00:47 AUDIT audit:93 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"CREATE
TABLE","opId":"6173128343752924","opStatus":"SUCCESS","opTime":"125 ms","table":"default.hive_carbon_example","extraInfo":{"bad_record_path":"","local_dictionary_enable":"true","external":"false","sort_columns":"","comment":""}}
2019-01-02 08:00:47 AUDIT audit:72 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173128497629888","opStatus":"START"}
2019-01-02 08:00:47 AUDIT audit:93 - {"time":"January 2, 2019 12:00:47 AM PST","username":"jenkins","opName":"LOAD
DATA","opId":"6173128497629888","opStatus":"SUCCESS","opTime":"200 ms","table":"default.hive_carbon_example","extraInfo":{"SegmentId":"0","DataSize":"921.0B","IndexSize":"498.0B"}}
+---+---------+--------+
| id|     name|  salary|
+---+---------+--------+
|  1|  'liang'|200000.0|
|  2|'anubhav'| 20000.0|
+---+---------+--------+

OK
OK
OK
OK
+---++-------++--------------+
| ID|| NAME || SALARY        |
+---++-------++--------------+
| 1 || 'liang' || 200000.0  |
+---++-------++--------------+
| 2 || 'anubhav' || 20000.0   |
+---++-------++--------------+
******Total Number Of Rows Fetched ****** 2
OK
+--------------+
| NAME         |
+---++---------+
| 'liang'    |
+---++---------+
| 'anubhav'      |
+---++---------+
 ********** Total Rows Fetched When Quering The Individual Columns **********2
OK
+---++-------++--------------+
| Salary|| ID || NAME        |
+---++-------++--------------+
| 200000.0 || 1 || 'liang'  |
+---++-------++--------------+
| 20000.0 || 2 || 'anubhav'   |
+---++-------++--------------+
 ********** Total Rows Fetched When Quering The Out Of Order Columns **********2
- HiveExample
2019-01-02 08:01:27 ERROR OperationLog:166 - Failed to remove corresponding log file of operation:
OperationHandle [opType=EXECUTE_STATEMENT, getHandleIdentifier()=f410ec33-0b23-49f9-8436-d1decc26b62e]
java.io.FileNotFoundException: File does not exist: /tmp/jenkins/operation_logs/934121b5-9057-408f-93be-eb9c6780537b/f410ec33-0b23-49f9-8436-d1decc26b62e
	at org.apache.commons.io.FileUtils.forceDelete(FileUtils.java:2275)
	at org.apache.hadoop.hive.ql.session.OperationLog$LogFile.remove(OperationLog.java:163)
	at org.apache.hadoop.hive.ql.session.OperationLog.close(OperationLog.java:121)
	at org.apache.hive.service.cli.operation.Operation.cleanupOperationLog(Operation.java:269)
	at org.apache.hive.service.cli.operation.SQLOperation.close(SQLOperation.java:317)
	at org.apache.hive.service.cli.operation.OperationManager.closeOperation(OperationManager.java:204)
	at org.apache.hive.service.cli.session.HiveSessionImpl.close(HiveSessionImpl.java:557)
	at org.apache.hive.service.cli.session.HiveSessionImplwithUGI.close(HiveSessionImplwithUGI.java:110)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
	at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
	at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
	at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
	at com.sun.proxy.$Proxy39.close(Unknown Source)
	at org.apache.hive.service.cli.session.SessionManager.closeSession(SessionManager.java:292)
	at org.apache.hive.service.cli.CLIService.closeSession(CLIService.java:237)
	at org.apache.hive.service.cli.thrift.ThriftCLIService$1.deleteContext(ThriftCLIService.java:120)
	at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:300)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
2019-01-02 08:01:27 ERROR HiveSessionImpl:591 - Failed to cleanup session log dir: SessionHandle
[934121b5-9057-408f-93be-eb9c6780537b]
java.io.FileNotFoundException: File does not exist: /tmp/jenkins/operation_logs/934121b5-9057-408f-93be-eb9c6780537b
	at org.apache.commons.io.FileUtils.forceDelete(FileUtils.java:2275)
	at org.apache.hive.service.cli.session.HiveSessionImpl.cleanupSessionLogDir(HiveSessionImpl.java:589)
	at org.apache.hive.service.cli.session.HiveSessionImpl.close(HiveSessionImpl.java:561)
	at org.apache.hive.service.cli.session.HiveSessionImplwithUGI.close(HiveSessionImplwithUGI.java:110)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78)
	at org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36)
	at org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
	at org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59)
	at com.sun.proxy.$Proxy39.close(Unknown Source)
	at org.apache.hive.service.cli.session.SessionManager.closeSession(SessionManager.java:292)
	at org.apache.hive.service.cli.CLIService.closeSession(CLIService.java:237)
	at org.apache.hive.service.cli.thrift.ThriftCLIService$1.deleteContext(ThriftCLIService.java:120)
	at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:300)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Exception encountered when invoking run on a nested suite - Cannot call methods on a
stopped SparkContext.
This stopped SparkContext was created at:

org.apache.spark.sql.test.util.QueryTest.<init>(QueryTest.scala:115)
org.apache.carbondata.examplesCI.RunExamples.<init>(RunExamples.scala:35)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:423)
java.lang.Class.newInstance(Class.java:442)
org.scalatest.tools.DiscoverySuite$.getSuiteInstance(DiscoverySuite.scala:69)
org.scalatest.tools.DiscoverySuite$$anonfun$1.apply(DiscoverySuite.scala:38)
org.scalatest.tools.DiscoverySuite$$anonfun$1.apply(DiscoverySuite.scala:37)
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
scala.collection.Iterator$class.foreach(Iterator.scala:893)
scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
scala.collection.AbstractIterable.foreach(Iterable.scala:54)
scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
scala.collection.AbstractTraversable.map(Traversable.scala:104)
org.scalatest.tools.DiscoverySuite.<init>(DiscoverySuite.scala:37)
org.scalatest.tools.Runner$.genDiscoSuites$1(Runner.scala:2390)

The currently active SparkContext was created at:

(No active SparkContext.)
          *** ABORTED ***
  java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext.
This stopped SparkContext was created at:

org.apache.spark.sql.test.util.QueryTest.<init>(QueryTest.scala:115)
org.apache.carbondata.examplesCI.RunExamples.<init>(RunExamples.scala:35)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:423)
java.lang.Class.newInstance(Class.java:442)
org.scalatest.tools.DiscoverySuite$.getSuiteInstance(DiscoverySuite.scala:69)
org.scalatest.tools.DiscoverySuite$$anonfun$1.apply(DiscoverySuite.scala:38)
org.scalatest.tools.DiscoverySuite$$anonfun$1.apply(DiscoverySuite.scala:37)
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
scala.collection.Iterator$class.foreach(Iterator.scala:893)
scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
scala.collection.AbstractIterable.foreach(Iterable.scala:54)
scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
scala.collection.AbstractTraversable.map(Traversable.scala:104)
org.scalatest.tools.DiscoverySuite.<init>(DiscoverySuite.scala:37)
org.scalatest.tools.Runner$.genDiscoSuites$1(Runner.scala:2390)

The currently active SparkContext was created at:

(No active SparkContext.)
  at org.apache.spark.SparkContext.assertNotStopped(SparkContext.scala:100)
  at org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:716)
  at org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:715)
  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
  at org.apache.spark.SparkContext.withScope(SparkContext.scala:701)
  at org.apache.spark.SparkContext.parallelize(SparkContext.scala:715)
  at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
  at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
  at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
  ...
Run completed in 3 minutes, 6 seconds.
Total number of tests run: 20
Suites: completed 1, aborted 1
Tests: succeeded 20, failed 0, canceled 0, ignored 0, pending 0
*** 1 SUITE ABORTED ***
[JENKINS] Recording test results

Mime
View raw message