lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject [JENKINS] Lucene-Solr-NightlyTests-5.3 - Build # 17 - Still Failing
Date Fri, 22 Jan 2016 19:00:10 GMT
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-5.3/17/

1 tests failed.
FAILED:  org.apache.solr.cloud.hdfs.StressHdfsTest.test

Error Message:
No registered leader was found after waiting for 30000ms , collection: delete_data_dir slice: shard4

Stack Trace:
org.apache.solr.common.SolrException: No registered leader was found after waiting for 30000ms , collection: delete_data_dir slice: shard4
	at __randomizedtesting.SeedInfo.seed([7141961A233E85D7:F915A9C08DC2E82F]:0)
	at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:637)
	at org.apache.solr.cloud.hdfs.StressHdfsTest.createAndDeleteCollection(StressHdfsTest.java:159)
	at org.apache.solr.cloud.hdfs.StressHdfsTest.test(StressHdfsTest.java:98)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:963)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:938)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
	at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 10969 lines...]
   [junit4] JVM J1: stdout was not empty, see: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/temp/junit4-J1-20160122_142625_681.sysout
   [junit4] >>> JVM J1: stdout (verbatim) ----
   [junit4] java.lang.OutOfMemoryError: GC overhead limit exceeded
   [junit4] Dumping heap to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/heapdumps/java_pid2447.hprof ...
   [junit4] Heap dump file created [608644680 bytes in 24.688 secs]
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/temp/junit4-J1-20160122_142625_681.syserr
   [junit4] >>> JVM J1: stderr (verbatim) ----
   [junit4] WARN: Unhandled exception in event serialization. -> java.lang.OutOfMemoryError: GC overhead limit exceeded
   [junit4] 	at java.nio.CharBuffer.wrap(CharBuffer.java:369)
   [junit4] 	at sun.nio.cs.StreamEncoder.implWrite(StreamEncoder.java:265)
   [junit4] 	at sun.nio.cs.StreamEncoder.write(StreamEncoder.java:125)
   [junit4] 	at sun.nio.cs.StreamEncoder.write(StreamEncoder.java:113)
   [junit4] 	at java.io.OutputStreamWriter.write(OutputStreamWriter.java:194)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.stream.JsonWriter.string(JsonWriter.java:535)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.stream.JsonWriter.value(JsonWriter.java:364)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.bind.TypeAdapters$22.write(TypeAdapters.java:626)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.bind.TypeAdapters$22.write(TypeAdapters.java:578)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.Streams.write(Streams.java:67)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.GsonToMiniGsonTypeAdapterFactory$3.write(GsonToMiniGsonTypeAdapterFactory.java:98)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.bind.TypeAdapterRuntimeTypeWrapper.write(TypeAdapterRuntimeTypeWrapper.java:66)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$1.write(ReflectiveTypeAdapterFactory.java:82)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.internal.bind.ReflectiveTypeAdapterFactory$Adapter.write(ReflectiveTypeAdapterFactory.java:194)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.dependencies.com.google.gson.Gson.toJson(Gson.java:512)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.events.Serializer.serialize(Serializer.java:87)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.slave.SlaveMain$4.write(SlaveMain.java:410)
   [junit4] 	at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
   [junit4] 	at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
   [junit4] 	at java.io.PrintStream.flush(PrintStream.java:338)
   [junit4] 	at java.io.FilterOutputStream.flush(FilterOutputStream.java:140)
   [junit4] 	at java.io.PrintStream.write(PrintStream.java:482)
   [junit4] 	at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
   [junit4] 	at sun.nio.cs.StreamEncoder.implFlushBuffer(StreamEncoder.java:291)
   [junit4] 	at sun.nio.cs.StreamEncoder.implFlush(StreamEncoder.java:295)
   [junit4] 	at sun.nio.cs.StreamEncoder.flush(StreamEncoder.java:141)
   [junit4] 	at java.io.OutputStreamWriter.flush(OutputStreamWriter.java:229)
   [junit4] 	at org.apache.log4j.helpers.QuietWriter.flush(QuietWriter.java:59)
   [junit4] 	at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:324)
   [junit4] 	at org.apache.log4j.WriterAppender.append(WriterAppender.java:162)
   [junit4] 	at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251)
   [junit4] 	at org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66)
   [junit4] <<< JVM J1: EOF ----

[...truncated 641 lines...]
   [junit4] Suite: org.apache.solr.cloud.hdfs.StressHdfsTest
   [junit4]   2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/init-core-data-001
   [junit4]   2> 9642014 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /ab_m/wo
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 9642100 WARN  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 9642115 WARN  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 9642118 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 9642158 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_34535_hdfs____y9njeo/webapp
   [junit4]   2> 9642523 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log NO JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 9643008 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34535
   [junit4]   2> 9643297 WARN  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 9643299 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 9643329 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_56341_datanode____ld00ly/webapp
   [junit4]   2> 9643693 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log NO JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 9644097 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:56341
   [junit4]   2> 9644513 WARN  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 9644514 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 9644567 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.6.0-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_34204_datanode____.ck5otg/webapp
   [junit4]   2> 9644653 INFO  (IPC Server handler 4 on 43763) [    ] BlockStateChange BLOCK* processReport: from storage DS-5943d48a-93bf-4f50-acb0-9efc92284236 node DatanodeRegistration(127.0.0.1, datanodeUuid=e880bdf6-3cde-4875-8e90-fa54f39c1df3, infoPort=56341, ipcPort=39414, storageInfo=lv=-56;cid=testClusterID;nsid=1875309937;c=0), blocks: 0, hasStaleStorages: true, processing time: 0 msecs
   [junit4]   2> 9644654 INFO  (IPC Server handler 4 on 43763) [    ] BlockStateChange BLOCK* processReport: from storage DS-923ed43a-9fca-4afa-bb50-775d2d9b1f76 node DatanodeRegistration(127.0.0.1, datanodeUuid=e880bdf6-3cde-4875-8e90-fa54f39c1df3, infoPort=56341, ipcPort=39414, storageInfo=lv=-56;cid=testClusterID;nsid=1875309937;c=0), blocks: 0, hasStaleStorages: false, processing time: 0 msecs
   [junit4]   2> 9644995 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log NO JSP Support for /, did not find org.apache.jasper.servlet.JspServlet
   [junit4]   2> 9645281 INFO  (SUITE-StressHdfsTest-seed#[7141961A233E85D7]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34204
   [junit4]   2> 9645602 INFO  (IPC Server handler 6 on 43763) [    ] BlockStateChange BLOCK* processReport: from storage DS-8c9d9fdc-d28a-4c28-8e3b-9a66704a8388 node DatanodeRegistration(127.0.0.1, datanodeUuid=6cc931c2-449b-4cb1-8e87-8c39e4e40057, infoPort=34204, ipcPort=33010, storageInfo=lv=-56;cid=testClusterID;nsid=1875309937;c=0), blocks: 0, hasStaleStorages: true, processing time: 0 msecs
   [junit4]   2> 9645603 INFO  (IPC Server handler 6 on 43763) [    ] BlockStateChange BLOCK* processReport: from storage DS-e5a97b2e-64d8-470d-a759-43bbbfc49217 node DatanodeRegistration(127.0.0.1, datanodeUuid=6cc931c2-449b-4cb1-8e87-8c39e4e40057, infoPort=34204, ipcPort=33010, storageInfo=lv=-56;cid=testClusterID;nsid=1875309937;c=0), blocks: 0, hasStaleStorages: false, processing time: 0 msecs
   [junit4]   2> 9645720 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 9645727 INFO  (Thread-103172) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 9645727 INFO  (Thread-103172) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 9645823 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkTestServer start zk server on port:60619
   [junit4]   2> 9645824 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9645855 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9645899 INFO  (zkCallback-2848-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@5bbff98b name:ZooKeeperConnection Watcher:127.0.0.1:60619 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9645900 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9645900 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9645900 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /solr
   [junit4]   2> 9645903 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9645932 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9645971 INFO  (zkCallback-2849-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@16c3a841 name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9645979 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9645980 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9645980 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /collections/collection1
   [junit4]   2> 9645981 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /collections/collection1/shards
   [junit4]   2> 9645982 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /collections/control_collection
   [junit4]   2> 9645983 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /collections/control_collection/shards
   [junit4]   2> 9645985 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 9645985 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/solrconfig.xml
   [junit4]   2> 9645987 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml
   [junit4]   2> 9645987 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/schema.xml
   [junit4]   2> 9645988 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 9645989 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 9645990 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 9645990 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/stopwords.txt
   [junit4]   2> 9645991 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 9645991 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/protwords.txt
   [junit4]   2> 9645993 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 9645993 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/currency.xml
   [junit4]   2> 9645994 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
   [junit4]   2> 9645994 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/enumsConfig.xml
   [junit4]   2> 9645996 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 9645996 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/open-exchange-rates.json
   [junit4]   2> 9645997 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 9645997 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 9645998 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 9645999 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/old_synonyms.txt
   [junit4]   2> 9646000 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 9646000 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient makePath: /configs/conf1/synonyms.txt
   [junit4]   2> 9646452 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.SolrTestCaseJ4 Writing core.properties file to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1
   [junit4]   2> 9646459 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.Server jetty-9.2.11.v20150529
   [junit4]   2> 9646555 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3a3aaed8{/ab_m/wo,null,AVAILABLE}
   [junit4]   2> 9646556 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.ServerConnector Started ServerConnector@62ad04c5{HTTP/1.1}{127.0.0.1:37899}
   [junit4]   2> 9646556 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.Server Started @9649713ms
   [junit4]   2> 9646556 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostPort=37899, solr.data.dir=hdfs://localhost:43763/hdfs__localhost_43763__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-5.3_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001_tempDir-002_control_data, hostContext=/ab_m/wo, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores}
   [junit4]   2> 9646557 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.s.SolrDispatchFilter SolrDispatchFilter.init(): sun.misc.Launcher$AppClassLoader@7b3cb2c6
   [junit4]   2> 9646557 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrResourceLoader new SolrResourceLoader for directory: '/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/'
   [junit4]   2> 9646580 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9646611 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9646643 INFO  (zkCallback-2850-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@67fbefe2 name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9646644 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9646644 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9646645 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 9646645 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/solr.xml
   [junit4]   2> 9646662 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoresLocator Config-defined core root directory: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores
   [junit4]   2> 9646663 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer New CoreContainer 1046115022
   [junit4]   2> 9646663 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer Loading cores into CoreContainer [instanceDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/]
   [junit4]   2> 9646663 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer loading shared library: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/lib
   [junit4]   2> 9646667 WARN  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrResourceLoader Can't find (or read) directory to add to classloader: lib (resolved as: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/lib).
   [junit4]   2> 9646683 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.h.c.HttpShardHandlerFactory created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 9646685 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 9646685 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.l.LogWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 9646685 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.l.LogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 9646685 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:60619/solr
   [junit4]   2> 9646686 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkController zkHost includes chroot
   [junit4]   2> 9646686 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9646716 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9646745 INFO  (zkCallback-2852-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@5d8afdc3 name:ZooKeeperConnection Watcher:127.0.0.1:60619 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9646747 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9646747 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9646844 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9646855 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@1949bd0b name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9646856 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9646857 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/queue
   [junit4]   2> 9646858 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/collection-queue-work
   [junit4]   2> 9646860 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/collection-map-running
   [junit4]   2> 9646861 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/collection-map-completed
   [junit4]   2> 9646863 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/collection-map-failure
   [junit4]   2> 9646864 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /live_nodes
   [junit4]   2> 9646865 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /aliases.json
   [junit4]   2> 9646866 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /clusterstate.json
   [junit4]   2> 9646867 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /security.json
   [junit4]   2> 9646868 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:37899_ab_m%2Fwo
   [junit4]   2> 9646868 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /live_nodes/127.0.0.1:37899_ab_m%2Fwo
   [junit4]   2> 9646870 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer_elect
   [junit4]   2> 9646870 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer_elect/election
   [junit4]   2> 9646871 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 9646872 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.LeaderElector Joined leadership election with path: /overseer_elect/election/95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000
   [junit4]   2> 9646873 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:37899_ab_m%2Fwo
   [junit4]   2> 9646873 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer_elect/leader
   [junit4]   2> 9646874 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer Overseer (id=95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) starting
   [junit4]   2> 9646875 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /overseer/queue-work
   [junit4]   2> 9646879 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.OverseerAutoReplicaFailoverThread Starting OverseerAutoReplicaFailoverThread autoReplicaFailoverWorkLoopDelay=10000 autoReplicaFailoverWaitAfterExpiration=10000 autoReplicaFailoverBadNodeExpiration=60000
   [junit4]   2> 9646894 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer Starting to work on the main queue
   [junit4]   2> 9646907 INFO  (OverseerCollectionProcessor-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.OverseerCollectionProcessor Process current queue of collection creations
   [junit4]   2> 9646922 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updating cluster state from ZooKeeper... 
   [junit4]   2> 9646946 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.CoreContainer Security conf doesn't exist. Skipping setup for authorization module.
   [junit4]   2> 9646946 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.CoreContainer No authentication plugin used.
   [junit4]   2> 9646947 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.CoresLocator Looking for core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores
   [junit4]   2> 9646948 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.SolrCore Created CoreDescriptor: {instanceDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1, dataDir=data/, configSetProperties=configsetprops.json, absoluteInstDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1/, name=collection1, shard=, transient=false, coreNodeName=, collection=control_collection, config=solrconfig.xml, loadOnStartup=true, schema=schema.xml}
   [junit4]   2> 9646948 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.CoresLocator Found core collection1 in /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1/
   [junit4]   2> 9646948 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.CoresLocator Found 1 core definitions
   [junit4]   2> 9646949 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController publishing state=down
   [junit4]   2> 9646949 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController numShards not found on descriptor - reading it from system property
   [junit4]   2> 9646949 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.s.SolrDispatchFilter user.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0
   [junit4]   2> 9646949 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.s.SolrDispatchFilter SolrDispatchFilter.init() done
   [junit4]   2> 9646950 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController look for our core node name
   [junit4]   2> 9646950 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.DistributedQueue NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 9646951 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":null,
   [junit4]   2>   "operation":"state",
   [junit4]   2>   "core":"collection1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "node_name":"127.0.0.1:37899_ab_m%2Fwo",
   [junit4]   2>   "base_url":"http://127.0.0.1:37899/ab_m/wo"} current state version: 0
   [junit4]   2> 9646951 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.o.ReplicaMutator Update state numShards=1 message={
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":null,
   [junit4]   2>   "operation":"state",
   [junit4]   2>   "core":"collection1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "node_name":"127.0.0.1:37899_ab_m%2Fwo",
   [junit4]   2>   "base_url":"http://127.0.0.1:37899/ab_m/wo"}
   [junit4]   2> 9646952 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.o.ClusterStateMutator building a new cName: control_collection
   [junit4]   2> 9646952 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.o.ReplicaMutator Assigning new node to shard shard=shard1
   [junit4]   2> 9646953 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 9646957 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updated cluster state version to 1
   [junit4]   2> 9647950 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController waiting to find shard id in clusterstate for collection1
   [junit4]   2> 9647950 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController Check for collection zkNode:control_collection
   [junit4]   2> 9647951 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ZkController Collection zkNode exists
   [junit4]   2> 9647951 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.c.ZkStateReader Load collection config from:/collections/control_collection
   [junit4]   2> 9647952 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.c.ZkStateReader path=/collections/control_collection configName=conf1 specified config exists in ZooKeeper
   [junit4]   2> 9647952 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.SolrResourceLoader new SolrResourceLoader for directory: '/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1/'
   [junit4]   2> 9647980 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.Config loaded config solrconfig.xml with version 0 
   [junit4]   2> 9647997 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.SolrConfig current version of requestparams : -1
   [junit4]   2> 9648019 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 5.3.2
   [junit4]   2> 9648071 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.Config Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 9648080 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.IndexSchema Reading Solr Schema from /configs/conf1/schema.xml
   [junit4]   2> 9648087 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.IndexSchema [collection1] Schema name=test
   [junit4]   2> 9649051 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 9649131 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.IndexSchema default search field in schema is text
   [junit4]   2> 9649228 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.IndexSchema unique key field: id
   [junit4]   2> 9649265 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.FileExchangeRateProvider Reloading exchange rates from file currency.xml
   [junit4]   2> 9649268 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.FileExchangeRateProvider Reloading exchange rates from file currency.xml
   [junit4]   2> 9649285 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 9649286 WARN  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Unknown key IMPORTANT NOTE
   [junit4]   2> 9649287 WARN  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Expected key, got STRING
   [junit4]   2> 9649287 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 9649287 WARN  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Unknown key IMPORTANT NOTE
   [junit4]   2> 9649287 WARN  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.s.OpenExchangeRatesOrgProvider Expected key, got STRING
   [junit4]   2> 9649288 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.ConfigSetProperties Did not find ConfigSet properties, assuming default properties: Can't find resource 'configsetprops.json' in classpath or '/configs/conf1', cwd=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0
   [junit4]   2> 9649289 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection   x:collection1] o.a.s.c.CoreContainer Creating SolrCore 'collection1' using configuration from collection control_collection
   [junit4]   2> 9649289 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore org.apache.solr.core.HdfsDirectoryFactory
   [junit4]   2> 9649289 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost:43763/solr_hdfs_home
   [junit4]   2> 9649289 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 9649289 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore [[collection1] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/control-001/cores/collection1/], dataDir=[null]
   [junit4]   2> 9649290 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.JmxMonitoredMap JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@2c44579
   [junit4]   2> 9649290 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data
   [junit4]   2> 9649345 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.s.h.HdfsLocalityReporter Registering direcotry hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data for locality metrics.
   [junit4]   2> 9649346 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.CachingDirectoryFactory return new directory for hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data
   [junit4]   2> 9649347 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore New index directory detected: old=null new=hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data/index/
   [junit4]   2> 9649365 WARN  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore [collection1] Solr index directory 'hdfs:/localhost:43763/solr_hdfs_home/control_collection/core_node1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 9649365 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data/index
   [junit4]   2> 9649388 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory Number of slabs of block cache [1] with direct memory allocation set to [true]
   [junit4]   2> 9649388 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory Block cache target memory usage, slab size of [8388608] will allocate [1] slabs and use ~[8388608] bytes
   [junit4]   2> 9649388 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.HdfsDirectoryFactory Creating new single instance HDFS BlockCache
   [junit4]   2> 9649476 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.s.b.BlockDirectory Block cache on write is disabled
   [junit4]   2> 9649476 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.s.h.HdfsLocalityReporter Registering direcotry hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data/index for locality metrics.
   [junit4]   2> 9649476 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.CachingDirectoryFactory return new directory for hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data/index
   [junit4]   2> 9649477 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=43, maxMergedSegmentMB=12.1083984375, floorSegmentMB=1.21484375, forceMergeDeletesPctAllowed=6.952317279418578, segmentsPerTier=26.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
   [junit4]   2> 9649566 INFO  (IPC Server handler 4 on 43763) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:59143 is added to blk_1073741825_1001{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-e5a97b2e-64d8-470d-a759-43bbbfc49217:NORMAL:127.0.0.1:49824|RBW], ReplicaUnderConstruction[[DISK]DS-923ed43a-9fca-4afa-bb50-775d2d9b1f76:NORMAL:127.0.0.1:59143|FINALIZED]]} size 0
   [junit4]   2> 9649578 INFO  (IPC Server handler 5 on 43763) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:49824 is added to blk_1073741825_1001{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-e5a97b2e-64d8-470d-a759-43bbbfc49217:NORMAL:127.0.0.1:49824|RBW], ReplicaUnderConstruction[[DISK]DS-923ed43a-9fca-4afa-bb50-775d2d9b1f76:NORMAL:127.0.0.1:59143|FINALIZED]]} size 0
   [junit4]   2> 9649582 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 	commit{dir=NRTCachingDirectory(BlockDirectory(HdfsDirectory@b73ecc7d lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@3c1284ec); maxCacheMB=192.0 maxMergeSizeMB=16.0),segFN=segments_1,generation=1}
   [junit4]   2> 9649582 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore newest commit generation = 1
   [junit4]   2> 9649598 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain creating updateRequestProcessorChain "nodistrib"
   [junit4]   2> 9649598 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain creating updateRequestProcessorChain "dedupe"
   [junit4]   2> 9649598 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 9649598 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain creating updateRequestProcessorChain "stored_sig"
   [junit4]   2> 9649599 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 9649599 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
   [junit4]   2> 9649599 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 9649599 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.p.UpdateRequestProcessorChain inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 9649600 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore no updateRequestProcessorChain defined as default, creating implicit default
   [junit4]   2> 9649602 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.h.l.XMLLoader xsltCacheLifetimeSeconds=60
   [junit4]   2> 9649603 INFO  (OldIndexDirectoryCleanupThreadForCore-collection1) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore Looking for old index directories to cleanup for core collection1 in hdfs://localhost:43763/solr_hdfs_home/control_collection/core_node1/data/
   [junit4]   2> 9649605 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.h.l.XMLLoader xsltCacheLifetimeSeconds=60
   [junit4]   2> 9649606 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.h.l.XMLLoader xsltCacheLifetimeSeconds=60
   [junit4]   2> 9649608 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.h.l.XMLLoader xsltCacheLifetimeSeconds=60
   [junit4]   2> 9649651 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.RequestHandlers Registered paths: /admin/plugins,/admin/mbeans,/admin/segments,/admin/threads,/admin/luke,/get,/update/json,/update/csv,/replication,/admin/properties,/admin/ping,/admin/file,/schema,standard,/admin/system,/update,/admin/logging,/update/json/docs,/config
   [junit4]   2> 9649653 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
   [junit4]   2> 9649659 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 9649660 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.UpdateLog Initializing HdfsUpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 tlogDfsReplication=2
   [junit4]   2> 9649716 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore Hard AutoCommit: disabled
   [junit4]   2> 9649717 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore Soft AutoCommit: disabled
   [junit4]   2> 9649718 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=38, maxMergeAtOnceExplicit=34, maxMergedSegmentMB=97.5302734375, floorSegmentMB=0.2021484375, forceMergeDeletesPctAllowed=29.43237093662322, segmentsPerTier=19.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.7948610523962821
   [junit4]   2> 9649742 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 	commit{dir=NRTCachingDirectory(BlockDirectory(HdfsDirectory@b73ecc7d lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@3c1284ec); maxCacheMB=192.0 maxMergeSizeMB=16.0),segFN=segments_1,generation=1}
   [junit4]   2> 9649742 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore newest commit generation = 1
   [junit4]   2> 9649743 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.s.SolrIndexSearcher Opening Searcher@41a99670[collection1] main
   [junit4]   2> 9649743 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.c.ZkStateReader Load collection config from:/collections/control_collection
   [junit4]   2> 9649744 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.c.ZkStateReader path=/collections/control_collection configName=conf1 specified config exists in ZooKeeper
   [junit4]   2> 9649744 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.ManagedResourceStorage Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
   [junit4]   2> 9649744 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 9649744 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.RestManager Initializing RestManager with initArgs: {}
   [junit4]   2> 9649745 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.ManagedResourceStorage Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 9649745 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.ManagedResourceStorage No data found for znode /configs/conf1/_rest_managed.json
   [junit4]   2> 9649745 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 9649745 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.r.RestManager Initializing 0 registered ManagedResources
   [junit4]   2> 9649745 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.h.ReplicationHandler Commits will be reserved for  10000
   [junit4]   2> 9649747 INFO  (searcherExecutor-10815-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SolrCore [collection1] Registered new searcher Searcher@41a99670[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 9649748 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.UpdateLog Looking up max value of version field to seed version buckets
   [junit4]   2> 9649749 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.VersionInfo Refreshing highest value of _version_ for 256 version buckets from index
   [junit4]   2> 9649749 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.VersionInfo No terms found for _version_, cannot seed version bucket highest value from index
   [junit4]   2> 9649749 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1524086801641242624
   [junit4]   2> 9649749 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.u.UpdateLog Took 0 ms to seed version buckets with highest version 1524086801641242624
   [junit4]   2> 9649749 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController watch zkdir /configs/conf1
   [junit4]   2> 9649750 INFO  (coreLoadExecutor-10814-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.CoreContainer registering core: collection1
   [junit4]   2> 9649852 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController Register replica - core:collection1 address:http://127.0.0.1:37899/ab_m/wo collection:control_collection shard:shard1
   [junit4]   2> 9649859 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.c.SolrZkClient makePath: /collections/control_collection/leader_elect/shard1/election
   [junit4]   2> 9649863 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.LeaderElector Joined leadership election with path: /collections/control_collection/leader_elect/shard1/election/95255424840040452-core_node1-n_0000000000
   [junit4]   2> 9649864 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Running the leader process for shard shard1
   [junit4]   2> 9649865 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.DistributedQueue NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 9649866 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 9649866 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 9649866 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:37899/ab_m/wo/collection1/
   [junit4]   2> 9649866 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 9649866 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.SyncStrategy http://127.0.0.1:37899/ab_m/wo/collection1/ has no replicas
   [junit4]   2> 9649867 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:37899/ab_m/wo/collection1/ shard1
   [junit4]   2> 9649867 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.c.SolrZkClient makePath: /collections/control_collection/leaders/shard1
   [junit4]   2> 9649867 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "operation":"leader",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection"} current state version: 1
   [junit4]   2> 9649868 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 9649870 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updated cluster state version to 2
   [junit4]   2> 9649872 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "operation":"leader",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "base_url":"http://127.0.0.1:37899/ab_m/wo",
   [junit4]   2>   "core":"collection1",
   [junit4]   2>   "state":"active"} current state version: 2
   [junit4]   2> 9649873 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.DistributedQueue NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 9649874 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 9649875 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updated cluster state version to 3
   [junit4]   2> 9649922 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController We are http://127.0.0.1:37899/ab_m/wo/collection1/ and leader is http://127.0.0.1:37899/ab_m/wo/collection1/
   [junit4]   2> 9649922 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:37899/ab_m/wo
   [junit4]   2> 9649923 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 9649923 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController publishing state=active
   [junit4]   2> 9649923 INFO  (coreZkRegister-10808-thread-1-processing-s:shard1 x:collection1 c:control_collection r:core_node1 n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo c:control_collection s:shard1 r:core_node1 x:collection1] o.a.s.c.ZkController numShards not found on descriptor - reading it from system property
   [junit4]   2> 9649924 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.DistributedQueue NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 9649925 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core_node_name":"core_node1",
   [junit4]   2>   "operation":"state",
   [junit4]   2>   "core":"collection1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "node_name":"127.0.0.1:37899_ab_m%2Fwo",
   [junit4]   2>   "base_url":"http://127.0.0.1:37899/ab_m/wo"} current state version: 3
   [junit4]   2> 9649926 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.o.ReplicaMutator Update state numShards=1 message={
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core_node_name":"core_node1",
   [junit4]   2>   "operation":"state",
   [junit4]   2>   "core":"collection1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "node_name":"127.0.0.1:37899_ab_m%2Fwo",
   [junit4]   2>   "base_url":"http://127.0.0.1:37899/ab_m/wo"}
   [junit4]   2> 9649952 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9649983 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9649999 INFO  (zkCallback-2855-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@10468d4d name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9650003 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9650004 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9650004 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ZkStateReader Updating cluster state from ZooKeeper... 
   [junit4]   2> 9650006 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false
   [junit4]   2> 9650028 INFO  (zkCallback-2855-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 9650028 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 9650030 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updated cluster state version to 4
   [junit4]   2> 9650037 INFO  (zkCallback-2855-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated cluster state version to 4
   [junit4]   2> 9650365 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.SolrTestCaseJ4 Writing core.properties file to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores/collection1
   [junit4]   2> 9650376 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001
   [junit4]   2> 9650387 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.Server jetty-9.2.11.v20150529
   [junit4]   2> 9650464 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@b0176c6{/ab_m/wo,null,AVAILABLE}
   [junit4]   2> 9650464 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.ServerConnector Started ServerConnector@d1cc6cf{HTTP/1.1}{127.0.0.1:48419}
   [junit4]   2> 9650464 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.e.j.s.Server Started @9653621ms
   [junit4]   2> 9650465 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=hdfs://localhost:43763/hdfs__localhost_43763__x1_jenkins_jenkins-slave_workspace_Lucene-Solr-NightlyTests-5.3_solr_build_solr-core_test_J0_temp_solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001_tempDir-002_jetty1, hostContext=/ab_m/wo, solrconfig=solrconfig.xml, hostPort=48419, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores}
   [junit4]   2> 9650465 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.s.SolrDispatchFilter SolrDispatchFilter.init(): sun.misc.Launcher$AppClassLoader@7b3cb2c6
   [junit4]   2> 9650465 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrResourceLoader new SolrResourceLoader for directory: '/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/'
   [junit4]   2> 9650501 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9650524 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9650542 INFO  (zkCallback-2856-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@730c916c name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9650542 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9650543 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9650544 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 9650544 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/solr.xml
   [junit4]   2> 9650558 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoresLocator Config-defined core root directory: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores
   [junit4]   2> 9650559 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer New CoreContainer 693354047
   [junit4]   2> 9650559 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer Loading cores into CoreContainer [instanceDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/]
   [junit4]   2> 9650559 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.CoreContainer loading shared library: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/lib
   [junit4]   2> 9650559 WARN  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.SolrResourceLoader Can't find (or read) directory to add to classloader: lib (resolved as: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/lib).
   [junit4]   2> 9650579 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.h.c.HttpShardHandlerFactory created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 9650580 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.u.UpdateShardHandler Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 9650581 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.l.LogWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 9650581 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.l.LogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 9650581 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:60619/solr
   [junit4]   2> 9650582 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.ZkController zkHost includes chroot
   [junit4]   2> 9650582 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkCredentialsProvider
   [junit4]   2> 9650595 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9650635 INFO  (zkCallback-2858-thread-1) [    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@c61d75f name:ZooKeeperConnection Watcher:127.0.0.1:60619 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9650636 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9650637 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [    ] o.a.s.c.c.SolrZkClient Using default ZkACLProvider
   [junit4]   2> 9650667 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Waiting for client to connect to ZooKeeper
   [junit4]   2> 9650683 INFO  (zkCallback-2859-thread-1-processing-n:127.0.0.1:48419_ab_m%2Fwo) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Watcher org.apache.solr.common.cloud.ConnectionManager@4a36ba70 name:ZooKeeperConnection Watcher:127.0.0.1:60619/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 9650684 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.c.ConnectionManager Client is connected to ZooKeeper
   [junit4]   2> 9650687 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.c.ZkStateReader Updating cluster state from ZooKeeper... 
   [junit4]   2> 9651691 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:48419_ab_m%2Fwo
   [junit4]   2> 9651691 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.c.SolrZkClient makePath: /live_nodes/127.0.0.1:48419_ab_m%2Fwo
   [junit4]   2> 9651695 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 9651695 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.LeaderElector Joined leadership election with path: /overseer_elect/election/95255424840040456-127.0.0.1:48419_ab_m%2Fwo-n_0000000001
   [junit4]   2> 9651696 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.LeaderElector Watching path /overseer_elect/election/95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000 to know if I could be the leader
   [junit4]   2> 9651733 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.CoreContainer Security conf doesn't exist. Skipping setup for authorization module.
   [junit4]   2> 9651733 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.CoreContainer No authentication plugin used.
   [junit4]   2> 9651739 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.CoresLocator Looking for core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores
   [junit4]   2> 9651739 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.SolrCore Created CoreDescriptor: {shard=, config=solrconfig.xml, schema=schema.xml, transient=false, absoluteInstDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores/collection1/, loadOnStartup=true, coreNodeName=, dataDir=data/, configSetProperties=configsetprops.json, instanceDir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores/collection1, name=collection1, collection=collection1}
   [junit4]   2> 9651740 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.CoresLocator Found core collection1 in /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0/temp/solr.cloud.hdfs.StressHdfsTest_7141961A233E85D7-001/shard-1-001/cores/collection1/
   [junit4]   2> 9651740 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.c.CoresLocator Found 1 core definitions
   [junit4]   2> 9651741 INFO  (coreLoadExecutor-10825-thread-1-processing-n:127.0.0.1:48419_ab_m%2Fwo) [n:127.0.0.1:48419_ab_m%2Fwo c:collection1   x:collection1] o.a.s.c.ZkController publishing state=down
   [junit4]   2> 9651741 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.s.SolrDispatchFilter user.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/J0
   [junit4]   2> 9651741 INFO  (TEST-StressHdfsTest.test-seed#[7141961A233E85D7]) [n:127.0.0.1:48419_ab_m%2Fwo    ] o.a.s.s.SolrDispatchFilter SolrDispatchFilter.init() done
   [junit4]   2> 9651741 INFO  (coreLoadExecutor-10825-thread-1-processing-n:127.0.0.1:48419_ab_m%2Fwo) [n:127.0.0.1:48419_ab_m%2Fwo c:collection1   x:collection1] o.a.s.c.ZkController numShards not found on descriptor - reading it from system property
   [junit4]   2> 9651742 INFO  (zkCallback-2853-thread-1-processing-n:127.0.0.1:37899_ab_m%2Fwo) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.DistributedQueue NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 9651742 INFO  (coreLoadExecutor-10825-thread-1-processing-n:127.0.0.1:48419_ab_m%2Fwo) [n:127.0.0.1:48419_ab_m%2Fwo c:collection1   x:collection1] o.a.s.c.ZkController look for our core node name
   [junit4]   2> 9651743 INFO  (OverseerStateUpdate-95255424840040452-127.0.0.1:37899_ab_m%2Fwo-n_0000000000) [n:127.0.0.1:37899_ab_m%2Fwo    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "state":"down",
   [junit4]   2>  

[...truncated too long message...]

ar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/commons-math3-3.4.1.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/easymock-3.0.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/ehcache-core-2.4.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/hadoop-common-2.6.0-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/hadoop-hdfs-2.6.0-tests.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/hadoop-minikdc-2.6.0.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jackson-annotations-2.5.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jackson-databind-2.5.4.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jcl-over-slf4j-1.7.7.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jersey-core-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jersey-server-1.9.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jetty-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/jetty-util-6.1.26.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/mina-core-2.0.0-M5.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/core/test-lib/objenesis-1.2.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/lucene/build/analysis/icu/lucene-analyzers-icu-5.3.2-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/contrib/solr-analysis-extras/classes/java:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/contrib/analysis-extras/lib/icu4j-54.1.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-launcher.jar:/x1/jenkins/.ant/lib/ivy-2.3.0.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-antlr.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-regexp.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-resolver.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-commons-net.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-jmf.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-oro.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-xalan2.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-commons-logging.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-jsch.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-junit4.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-javamail.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-jai.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-jdepend.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-junit.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-bsf.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-swing.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-netrexx.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-bcel.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-testutil.jar:/home/jenkins/jenkins-slave/tools/hudson.tasks.Ant_AntInstallation/ant-1.8.2/lib/ant-apache-log4j.jar:/x1/jenkins/jenkins-slave/tools/hudson.model.JDK/latest1.7/lib/tools.jar:/x1/jenkins/.ivy2/cache/com.carrotsearch.randomizedtesting/junit4-ant/jars/junit4-ant-2.1.13.jar com.carrotsearch.ant.tasks.junit4.slave.SlaveMainSafe -eventsfile /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/temp/junit4-J2-20160122_142625_686.events @/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build/solr-core/test/temp/junit4-J2-20160122_142625_686.suites
   [junit4] ERROR: JVM J2 ended with an exception: Quit event not received from the forked process? This may indicate JVM crash or runner bugs.
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.JUnit4.executeSlave(JUnit4.java:1504)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.JUnit4.access$000(JUnit4.java:133)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:964)
   [junit4] 	at com.carrotsearch.ant.tasks.junit4.JUnit4$2.call(JUnit4.java:961)
   [junit4] 	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
   [junit4] 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
   [junit4] 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
   [junit4] 	at java.lang.Thread.run(Thread.java:745)

BUILD FAILED
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/build.xml:733: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/build.xml:670: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/build.xml:59: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/build.xml:230: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/solr/common-build.xml:524: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/lucene/common-build.xml:1449: The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-NightlyTests-5.3/lucene/common-build.xml:1003: At least one slave process threw an exception, first: Quit event not received from the forked process? This may indicate JVM crash or runner bugs.

Total time: 376 minutes 18 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
No prior successful build to compare, so performing full copy of artifacts
ERROR: Failed to archive {README.txt=README.txt, solr/build/solr-core/test/temp/junit4-J2-20160122_142625_686.events=solr/build/solr-core/test/temp/junit4-J2-20160122_142625_686.events, heapdumps/java_pid2447.hprof=heapdumps/java_pid2447.hprof, solr/build/solr-core/test/temp/junit4-J1-20160122_142625_681.events=solr/build/solr-core/test/temp/junit4-J1-20160122_142625_681.events, solr/build/solr-core/test/temp/junit4-J0-20160122_142625_681.events=solr/build/solr-core/test/temp/junit4-J0-20160122_142625_681.events} due to internal error; falling back to full archiving
java.lang.IllegalArgumentException: Negative time
	at java.io.File.setLastModified(File.java:1421)
	at hudson.FilePath.readFromTar(FilePath.java:2289)
	at hudson.FilePath.copyRecursiveTo(FilePath.java:2208)
	at jenkins.model.StandardArtifactManager.archive(StandardArtifactManager.java:61)
	at com.cloudbees.jenkins.plugins.jsync.archiver.JSyncArtifactManager.remoteSync(JSyncArtifactManager.java:111)
	at com.cloudbees.jenkins.plugins.jsync.archiver.JSyncArtifactManager.archive(JSyncArtifactManager.java:72)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:219)
	at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:74)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:776)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:723)
	at hudson.model.Build$BuildExecution.post2(Build.java:183)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:670)
	at hudson.model.Run.execute(Run.java:1763)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:98)
	at hudson.model.Executor.run(Executor.java:381)
ERROR: Build step failed with exception
java.lang.IllegalArgumentException: Negative time
	at java.io.File.setLastModified(File.java:1421)
	at hudson.FilePath.readFromTar(FilePath.java:2289)
	at hudson.FilePath.copyRecursiveTo(FilePath.java:2208)
	at jenkins.model.StandardArtifactManager.archive(StandardArtifactManager.java:61)
	at com.cloudbees.jenkins.plugins.jsync.archiver.JSyncArtifactManager.archive(JSyncArtifactManager.java:76)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:219)
	at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:74)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:776)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:723)
	at hudson.model.Build$BuildExecution.post2(Build.java:183)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:670)
	at hudson.model.Run.execute(Run.java:1763)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:98)
	at hudson.model.Executor.run(Executor.java:381)
Build step 'Archive the artifacts' marked build as failure
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any



Mime
View raw message