carbondata-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: carbondata-master-spark-2.2 #626
Date Mon, 09 Jul 2018 08:00:48 GMT
See <https://builds.apache.org/job/carbondata-master-spark-2.2/626/display/redirect>

------------------------------------------
[...truncated 48.89 MB...]
	at org.apache.spark.sql.execution.command.DropTableCommand.run(ddl.scala:203)
	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:58)
	at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:56)
	at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:67)
	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:183)
	at org.apache.spark.sql.CarbonSession$$anonfun$sql$1.apply(CarbonSession.scala:107)
	at org.apache.spark.sql.CarbonSession$$anonfun$sql$1.apply(CarbonSession.scala:96)
	at org.apache.spark.sql.CarbonSession.withProfiler(CarbonSession.scala:154)
	at org.apache.spark.sql.CarbonSession.sql(CarbonSession.scala:94)
	at org.apache.spark.sql.test.Spark2TestQueryExecutor.sql(Spark2TestQueryExecutor.scala:35)
	at org.apache.spark.sql.test.util.QueryTest.sql(QueryTest.scala:113)
	at org.apache.carbondata.spark.testsuite.standardpartition.StandardPartitionGlobalSortTestCase$$anonfun$30.apply$mcV$sp(StandardPartitionGlobalSortTestCase.scala:684)
	at org.apache.carbondata.spark.testsuite.standardpartition.StandardPartitionGlobalSortTestCase$$anonfun$30.apply(StandardPartitionGlobalSortTestCase.scala:680)
	at org.apache.carbondata.spark.testsuite.standardpartition.StandardPartitionGlobalSortTestCase$$anonfun$30.apply(StandardPartitionGlobalSortTestCase.scala:680)
	at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
	at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
	at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
	at org.scalatest.Transformer.apply(Transformer.scala:22)
	at org.scalatest.Transformer.apply(Transformer.scala:20)
	at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
	at org.apache.spark.sql.test.util.CarbonFunSuite.withFixture(CarbonFunSuite.scala:41)
	at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
	at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
	at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
	at scala.collection.immutable.List.foreach(List.scala:381)
	at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
	at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
	at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
	at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
	at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
	at org.scalatest.Suite$class.run(Suite.scala:1424)
	at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
	at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
	at org.apache.carbondata.spark.testsuite.standardpartition.StandardPartitionGlobalSortTestCase.org$scalatest$BeforeAndAfterAll$$super$run(StandardPartitionGlobalSortTestCase.scala:31)
	at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
	at org.apache.carbondata.spark.testsuite.standardpartition.StandardPartitionGlobalSortTestCase.run(StandardPartitionGlobalSortTestCase.scala:31)
	at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
	at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
	at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
	at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
	at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
	at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
	at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
	at org.scalatest.Suite$class.run(Suite.scala:1421)
	at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
	at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
	at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
	at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
	at scala.collection.immutable.List.foreach(List.scala:381)
	at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
	at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
	at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
	at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
	at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
	at org.scalatest.tools.Runner$.main(Runner.scala:860)
	at org.scalatest.tools.Runner.main(Runner.scala)
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_database: default
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO CarbonSparkSqlParser: Parsing command: create table partdatecarb4_hive(name
string, age Int) partitioned by(country string, state string, city string)
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
skip CarbonOptimizer
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
Skip CarbonOptimizer
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_database: default
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_database: default
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO HiveMetaStore: 0: get_database: default
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
18/07/09 01:00:22 INFO HiveMetaStore: 0: create_table: Table(tableName:partdatecarb4_hive,
dbName:default, owner:jenkins, createTime:1531123222, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:name,
type:string, comment:null), FieldSchema(name:age, type:int, comment:null)], location:file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive,
inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat,
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe,
parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[FieldSchema(name:country,
type:string, comment:null), FieldSchema(name:state, type:string, comment:null), FieldSchema(name:city,
type:string, comment:null)], parameters:{spark.sql.sources.schema.numPartCols=3, spark.sql.sources.schema.part.0={"type":"struct","fields":[{"name":"name","type":"string","nullable":true,"metadata":{}},{"name":"age","type":"integer","nullable":true,"metadata":{}},{"name":"country","type":"string","nullable":true,"metadata":{}},{"name":"state","type":"string","nullable":true,"metadata":{}},{"name":"city","type":"string","nullable":true,"metadata":{}}]},
spark.sql.sources.schema.partCol.0=country, spark.sql.sources.schema.partCol.2=city, spark.sql.sources.schema.partCol.1=state,
spark.sql.sources.schema.numParts=1}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE,
privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null, rolePrivileges:null))
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=create_table: Table(tableName:partdatecarb4_hive,
dbName:default, owner:jenkins, createTime:1531123222, lastAccessTime:0, retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:name,
type:string, comment:null), FieldSchema(name:age, type:int, comment:null)], location:file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive,
inputFormat:org.apache.hadoop.mapred.TextInputFormat, outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat,
compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe,
parameters:{serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, skewedInfo:SkewedInfo(skewedColNames:[],
skewedColValues:[], skewedColValueLocationMaps:{})), partitionKeys:[FieldSchema(name:country,
type:string, comment:null), FieldSchema(name:state, type:string, comment:null), FieldSchema(name:city,
type:string, comment:null)], parameters:{spark.sql.sources.schema.numPartCols=3, spark.sql.sources.schema.part.0={"type":"struct","fields":[{"name":"name","type":"string","nullable":true,"metadata":{}},{"name":"age","type":"integer","nullable":true,"metadata":{}},{"name":"country","type":"string","nullable":true,"metadata":{}},{"name":"state","type":"string","nullable":true,"metadata":{}},{"name":"city","type":"string","nullable":true,"metadata":{}}]},
spark.sql.sources.schema.partCol.0=country, spark.sql.sources.schema.partCol.2=city, spark.sql.sources.schema.partCol.1=state,
spark.sql.sources.schema.numParts=1}, viewOriginalText:null, viewExpandedText:null, tableType:MANAGED_TABLE,
privileges:PrincipalPrivilegeSet(userPrivileges:{}, groupPrivileges:null, rolePrivileges:null))

18/07/09 01:00:22 WARN HiveMetaStore: Location: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive
specified for non-external table:partdatecarb4_hive
18/07/09 01:00:22 INFO FileUtils: Creating directory if it doesn't exist: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive
18/07/09 01:00:22 INFO CarbonSparkSqlParser: Parsing command: insert into partdatecarb4_hive
partition(state,city,country='india') select 'name1',12,'KA', 'BGLR'
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: int
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
skip CarbonOptimizer
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
Skip CarbonOptimizer
18/07/09 01:00:22 INFO FileUtils: Creating directory if it doesn't exist: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/.hive-staging_hive_2018-07-09_01-00-22_656_4233094735015688534-1
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
skip CarbonOptimizer
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
Skip CarbonOptimizer
18/07/09 01:00:22 INFO FileOutputCommitter: File Output Committer Algorithm version is 1
18/07/09 01:00:22 INFO CarbonSQLHadoopMapReduceCommitProtocol: Using output committer class
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
18/07/09 01:00:22 INFO CodeGenerator: Code generated in 6.178569 ms
18/07/09 01:00:22 INFO SparkContext: Starting job: sql at StandardPartitionGlobalSortTestCase.scala:688
18/07/09 01:00:22 INFO DAGScheduler: Got job 2350 (sql at StandardPartitionGlobalSortTestCase.scala:688)
with 1 output partitions
18/07/09 01:00:22 INFO DAGScheduler: Final stage: ResultStage 3234 (sql at StandardPartitionGlobalSortTestCase.scala:688)
18/07/09 01:00:22 INFO DAGScheduler: Parents of final stage: List()
18/07/09 01:00:22 INFO DAGScheduler: Missing parents: List()
18/07/09 01:00:22 INFO DAGScheduler: Submitting ResultStage 3234 (MapPartitionsRDD[12179]
at sql at StandardPartitionGlobalSortTestCase.scala:688), which has no missing parents
18/07/09 01:00:22 INFO MemoryStore: Block broadcast_3759 stored as values in memory (estimated
size 155.8 KB, free 1444.0 MB)
18/07/09 01:00:22 INFO MemoryStore: Block broadcast_3759_piece0 stored as bytes in memory
(estimated size 56.2 KB, free 1443.9 MB)
18/07/09 01:00:22 INFO BlockManagerInfo: Added broadcast_3759_piece0 in memory on 67.195.81.140:40222
(size: 56.2 KB, free: 1454.5 MB)
18/07/09 01:00:22 INFO SparkContext: Created broadcast 3759 from broadcast at DAGScheduler.scala:1006
18/07/09 01:00:22 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 3234 (MapPartitionsRDD[12179]
at sql at StandardPartitionGlobalSortTestCase.scala:688) (first 15 tasks are for partitions
Vector(0))
18/07/09 01:00:22 INFO TaskSchedulerImpl: Adding task set 3234.0 with 1 tasks
18/07/09 01:00:22 INFO TaskSetManager: Starting task 0.0 in stage 3234.0 (TID 49800, localhost,
executor driver, partition 0, PROCESS_LOCAL, 5039 bytes)
18/07/09 01:00:22 INFO Executor: Running task 0.0 in stage 3234.0 (TID 49800)
18/07/09 01:00:22 INFO CodeGenerator: Code generated in 7.373409 ms
18/07/09 01:00:22 INFO FileOutputCommitter: File Output Committer Algorithm version is 1
18/07/09 01:00:22 INFO CarbonSQLHadoopMapReduceCommitProtocol: Using output committer class
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
18/07/09 01:00:22 INFO CodeGenerator: Code generated in 4.568556 ms
18/07/09 01:00:22 INFO CodeGenerator: Code generated in 12.041288 ms
18/07/09 01:00:22 INFO CodeGenerator: Code generated in 4.46988 ms
18/07/09 01:00:22 INFO FileOutputCommitter: Saved output of task 'attempt_20180709010022_3234_m_000000_0'
to file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/.hive-staging_hive_2018-07-09_01-00-22_656_4233094735015688534-1/-ext-10000/_temporary/0/task_20180709010022_3234_m_000000
18/07/09 01:00:22 INFO SparkHadoopMapRedUtil: attempt_20180709010022_3234_m_000000_0: Committed
18/07/09 01:00:22 INFO Executor: Finished task 0.0 in stage 3234.0 (TID 49800). 1829 bytes
result sent to driver
18/07/09 01:00:22 INFO TaskSetManager: Finished task 0.0 in stage 3234.0 (TID 49800) in 58
ms on localhost (executor driver) (1/1)
18/07/09 01:00:22 INFO TaskSchedulerImpl: Removed TaskSet 3234.0, whose tasks have all completed,
from pool 
18/07/09 01:00:22 INFO DAGScheduler: ResultStage 3234 (sql at StandardPartitionGlobalSortTestCase.scala:688)
finished in 0.059 s
18/07/09 01:00:22 INFO DAGScheduler: Job 2350 finished: sql at StandardPartitionGlobalSortTestCase.scala:688,
took 0.079669 s
18/07/09 01:00:22 INFO FileFormatWriter: Job null committed.
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: int
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO HiveMetaStore: 0: partition_name_has_valid_characters
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=partition_name_has_valid_characters

18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO HiveMetaStore: 0: get_partition_with_auth : db=default tbl=partdatecarb4_hive[india,KA,BGLR]
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_partition_with_auth :
db=default tbl=partdatecarb4_hive[india,KA,BGLR]	
18/07/09 01:00:22 INFO FileUtils: Creating directory if it doesn't exist: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/country=india/state=KA/city=BGLR
18/07/09 01:00:22 INFO SessionState: Could not get hdfsEncryptionShim, it is only applicable
to hdfs filesystem.
18/07/09 01:00:22 INFO Hive: Renaming src: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/.hive-staging_hive_2018-07-09_01-00-22_656_4233094735015688534-1/-ext-10000/state=KA/city=BGLR/part-00000-257f05b2-933d-4ba2-bbd1-632be779d663.c000,
dest: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/country=india/state=KA/city=BGLR/part-00000-257f05b2-933d-4ba2-bbd1-632be779d663.c000,
Status:true
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_partition_with_auth : db=default tbl=partdatecarb4_hive[india,KA,BGLR]
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_partition_with_auth :
db=default tbl=partdatecarb4_hive[india,KA,BGLR]	
18/07/09 01:00:22 INFO HiveMetaStore: 0: append_partition : db=default tbl=partdatecarb4_hive[india,KA,BGLR]
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=append_partition : db=default
tbl=partdatecarb4_hive[india,KA,BGLR]	
18/07/09 01:00:22 WARN log: Updating partition stats fast for: partdatecarb4_hive
18/07/09 01:00:22 WARN log: Updated size to 9
18/07/09 01:00:22 INFO Hive: New loading path = file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/.hive-staging_hive_2018-07-09_01-00-22_656_4233094735015688534-1/-ext-10000/state=KA/city=BGLR
with partSpec {country=india, state=KA, city=BGLR}
18/07/09 01:00:22 INFO CarbonSparkSqlParser: Parsing command: `default`.`partdatecarb4_hive`
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_database: default
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: int
18/07/09 01:00:22 INFO CarbonSparkSqlParser: Parsing command: insert into partdatecarb4_hive
partition(state,city,country='india') select 'name1',12,'KA', 'BGLR'
18/07/09 01:00:22 INFO HiveMetaStore: 0: get_table : db=default tbl=partdatecarb4_hive
18/07/09 01:00:22 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=partdatecarb4_hive

18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: string
18/07/09 01:00:22 INFO CatalystSqlParser: Parsing command: int
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
skip CarbonOptimizer
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
Skip CarbonOptimizer
18/07/09 01:00:22 INFO FileUtils: Creating directory if it doesn't exist: file:/home/jenkins/jenkins-slave/workspace/carbondata-master-spark-2.2/integration/spark-common/target/warehouse/partdatecarb4_hive/.hive-staging_hive_2018-07-09_01-00-22_884_659282070237940617-1
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
skip CarbonOptimizer
18/07/09 01:00:22 INFO CarbonLateDecodeRule: ScalaTest-main-running-StandardPartitionGlobalSortTestCase
Skip CarbonOptimizer
18/07/09 01:00:22 INFO FileOutputCommitter: File Output Committer Algorithm version is 1
18/07/09 01:00:22 INFO CarbonSQLHadoopMapReduceCommitProtocol: Using output committer class
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: Failed to parse POMs
java.io.IOException: Backing channel 'H20' is disconnected.
	at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:214)
	at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:283)
	at com.sun.proxy.$Proxy152.isAlive(Unknown Source)
	at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1137)
	at hudson.maven.ProcessCache$MavenProcess.call(ProcessCache.java:166)
	at hudson.maven.MavenModuleSetBuild$MavenModuleSetBuildExecution.doRun(MavenModuleSetBuild.java:879)
	at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:504)
	at hudson.model.Run.execute(Run.java:1794)
	at hudson.maven.MavenModuleSetBuild.run(MavenModuleSetBuild.java:543)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
Caused by: java.io.IOException: Unexpected termination of the channel
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:77)
Caused by: java.io.EOFException
	at java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2679)
	at java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3154)
	at java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:862)
	at java.io.ObjectInputStream.<init>(ObjectInputStream.java:358)
	at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:48)
	at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:36)
	at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:63)
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9
ERROR: H20 is offline; cannot locate JDK 1.8 (latest)
ERROR: H20 is offline; cannot locate Maven 3.3.9

Mime
View raw message