hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Tony Reix (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-11617) Several tests are not stable (on OpenJDK / x86_64 / RHEL 7.1)
Date Wed, 25 Feb 2015 16:54:06 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-11617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14336746#comment-14336746
] 

Tony Reix commented on HADOOP-11617:
------------------------------------

Running org.apache.hadoop.fs.TestLocalDirAllocator
Tests run: 30, Failures: 9, Errors: 0, Skipped: 0, Time elapsed: 0.675 sec <<< FAILURE!
- in org.apache.hadoop.fs.TestLocalDirAllocator
test0[0](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.047 sec  <<<
FAILURE!
java.lang.AssertionError: Checking for build/test/temp/RELATIVE1 in build/test/temp/RELATIVE0/block6983547632959748375.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.test0(TestLocalDirAllocator.java:142)

testROBufferDirAndRWBufferDir[0](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed:
0.007 sec  <<< FAILURE!
java.lang.AssertionError: Checking for build/test/temp/RELATIVE2 in build/test/temp/RELATIVE1/block8298588481450601389.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testROBufferDirAndRWBufferDir(TestLocalDirAllocator.java:164)

testRWBufferDirBecomesRO[0](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.005
sec  <<< FAILURE!
java.lang.AssertionError: Checking for build/test/temp/RELATIVE3 in build/test/temp/RELATIVE4/block2843993964655020837.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testRWBufferDirBecomesRO(TestLocalDirAllocator.java:219)

test0[1](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.004 sec  <<<
FAILURE!
java.lang.AssertionError: Checking for /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE1
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE0/block4794871297104004564.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.test0(TestLocalDirAllocator.java:141)

testROBufferDirAndRWBufferDir[1](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed:
0.004 sec  <<< FAILURE!
java.lang.AssertionError: Checking for /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE2
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE1/block7909203284714183270.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testROBufferDirAndRWBufferDir(TestLocalDirAllocator.java:163)

testRWBufferDirBecomesRO[1](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.004
sec  <<< FAILURE!
java.lang.AssertionError: Checking for /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE3
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/ABSOLUTE4/block5800284409701937229.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testRWBufferDirBecomesRO(TestLocalDirAllocator.java:219)

test0[2](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.005 sec  <<<
FAILURE!
java.lang.AssertionError: Checking for file:/home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED1
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED0/block3321839206317898374.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.test0(TestLocalDirAllocator.java:141)

testROBufferDirAndRWBufferDir[2](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed:
0.004 sec  <<< FAILURE!
java.lang.AssertionError: Checking for file:/home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED2
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED1/block7755055899410246198.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testROBufferDirAndRWBufferDir(TestLocalDirAllocator.java:163)

testRWBufferDirBecomesRO[2](org.apache.hadoop.fs.TestLocalDirAllocator)  Time elapsed: 0.004
sec  <<< FAILURE!
java.lang.AssertionError: Checking for file:/home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED3
in /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-common-project/hadoop-common/build/test/temp/QUALIFIED4/block2032950190531603316.tmp
- FAILED!
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.fs.TestLocalDirAllocator.validateTempDirCreation(TestLocalDirAllocator.java:110)
        at org.apache.hadoop.fs.TestLocalDirAllocator.testRWBufferDirBecomesRO(TestLocalDirAllocator.java:220)


Running org.apache.hadoop.hdfs.qjournal.client.TestQuorumJournalManager
Tests run: 21, Failures: 1, Errors: 10, Skipped: 0, Time elapsed: 44.297 sec <<<
FAILURE! - in org.apache.hadoop.hdfs.qjournal.client.TestQuorumJournalManager
testCrashBetweenSyncLogAndPersistPaxosData(org.apache.hadoop.hdfs.qjournal.client.TestQuorumJournalManager)
 Time elapsed: 20.674 sec  <<< ERROR!
java.lang.Exception: test timed out after 20000 milliseconds
        at java.lang.Object.wait(Native Method)
        at java.lang.Object.wait(Object.java:503)

testCrashAtBeginningOfSegment(org.apache.hadoop.hdfs.qjournal.client.TestQuorumJournalManager)
 Time elapsed: 0.959 sec  <<< ERROR!
org.apache.hadoop.hdfs.qjournal.client.QuorumException: Got too many exceptions to achieve
quorum size 2/3. 3 exceptions thrown:
127.0.0.1:44453: Injected
        at org.apache.hadoop.hdfs.qjournal.server.Journal.acceptRecovery(Journal.java:840)
        at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.acceptRecovery(JournalNodeRpcServer.java:205)

testOutOfSyncAtBeginningOfSegment0(org.apache.hadoop.hdfs.qjournal.client.TestQuorumJournalManager)
 Time elapsed: 1.125 sec  <<< ERROR!
org.apache.hadoop.hdfs.qjournal.client.QuorumException: Got too many exceptions to achieve
quorum size 2/3. 3 exceptions thrown:
127.0.0.1:47793: Injected
        at org.apache.hadoop.hdfs.qjournal.server.Journal.acceptRecovery(Journal.java:840)
        at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.acceptRecovery(JournalNodeRpcServer.java:205)

127.0.0.1:54901: Call From soe01x/9.3.126.7 to localhost:54901 failed on connection exception:
java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused
        at org.apache.hadoop.hdfs.qjournal.client.QuorumException.create(QuorumException.java:81)
        at org.apache.hadoop.hdfs.qjournal.client.QuorumCall.rethrowException(QuorumCall.java:223)

Running org.apache.hadoop.hdfs.server.datanode.TestBPOfferService
Tests run: 5, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 2.526 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.server.datanode.TestBPOfferService
testBPInitErrorHandling(org.apache.hadoop.hdfs.server.datanode.TestBPOfferService)  Time elapsed:
0.148 sec  <<< FAILURE!
java.lang.AssertionError: expected:<2> but was:<1>
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.failNotEquals(Assert.java:647)
        at org.junit.Assert.assertEquals(Assert.java:128)
        at org.junit.Assert.assertEquals(Assert.java:472)
        at org.junit.Assert.assertEquals(Assert.java:456)
        at org.apache.hadoop.hdfs.server.datanode.TestBPOfferService.testBPInitErrorHandling(TestBPOfferService.java:334)


Running org.apache.hadoop.hdfs.server.datanode.TestDeleteBlockPool
Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 10.783 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.server.datanode.TestDeleteBlockPool
testDeleteBlockPool(org.apache.hadoop.hdfs.server.datanode.TestDeleteBlockPool)  Time elapsed:
7.704 sec  <<< ERROR!
java.io.IOException: All datanodes 127.0.0.1:36456 are bad. Aborting...
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.setupPipelineForAppendOrRecovery(DFSOutputStream.java:1128)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.processDatanodeError(DFSOutputStream.java:924)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:486)


Running org.apache.hadoop.hdfs.server.datanode.TestDnRespectsBlockReportSplitThreshold
Tests run: 3, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 3.934 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.server.datanode.TestDnRespectsBlockReportSplitThreshold
testAlwaysSplit(org.apache.hadoop.hdfs.server.datanode.TestDnRespectsBlockReportSplitThreshold)
 Time elapsed: 1.818 sec  <<< ERROR!
java.io.IOException: Cannot remove data directory: /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/datapath
'/home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data':
        absolute:/home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data
        permissions: drwx

Running org.apache.hadoop.hdfs.server.namenode.TestCheckpoint
Tests run: 37, Failures: 3, Errors: 0, Skipped: 0, Time elapsed: 77.452 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.server.namenode.TestCheckpoint
testCheckpoint(org.apache.hadoop.hdfs.server.namenode.TestCheckpoint)  Time elapsed: 1.984
sec  <<< FAILURE!
java.lang.AssertionError: Bad value for metric GetImageNumOps
Expected: gt(0)
     got: <0L>

        at org.junit.Assert.assertThat(Assert.java:780)
        at org.apache.hadoop.test.MetricsAsserts.assertCounterGt(MetricsAsserts.java:318)
        at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testCheckpoint(TestCheckpoint.java:1054)

testNameDirError(org.apache.hadoop.hdfs.server.namenode.TestCheckpoint)  Time elapsed: 0.708
sec  <<< FAILURE!
java.lang.AssertionError: NN should have failed to start with /home/tony/HADOOP/hadoop-common-FromIBMSOE-20150210/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/name1
set unreadable
        at org.junit.Assert.fail(Assert.java:93)
        at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testNameDirError(TestCheckpoint.java:194)

testCheckpointWithSeparateDirsAfterNameFails(org.apache.hadoop.hdfs.server.namenode.TestCheckpoint)
 Time elapsed: 0.897 sec  <<< FAILURE!
java.lang.AssertionError: Did not fail to checkpoint when there are no valid storage dirs
        at org.junit.Assert.fail(Assert.java:93)
        at org.apache.hadoop.hdfs.server.namenode.TestCheckpoint.testCheckpointWithSeparateDirsAfterNameFails(TestCheckpoint.java:2119)


Running org.apache.hadoop.hdfs.TestSafeMode
Tests run: 7, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 42.989 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.TestSafeMode
testInitializeReplQueuesEarly(org.apache.hadoop.hdfs.TestSafeMode)  Time elapsed: 7.322 sec
 <<< FAILURE!
java.lang.AssertionError: expected:<13> but was:<1>
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.failNotEquals(Assert.java:647)
        at org.junit.Assert.assertEquals(Assert.java:128)
        at org.junit.Assert.assertEquals(Assert.java:472)
        at org.junit.Assert.assertEquals(Assert.java:456)
        at org.apache.hadoop.hdfs.TestSafeMode.testInitializeReplQueuesEarly(TestSafeMode.java:212)

Running org.apache.hadoop.hdfs.TestSafeMode
Tests run: 7, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 47.591 sec <<< FAILURE!
- in org.apache.hadoop.hdfs.TestSafeMode
testInitializeReplQueuesEarly(org.apache.hadoop.hdfs.TestSafeMode)  Time elapsed: 12.261 sec
 <<< ERROR!
java.util.concurrent.TimeoutException: Timed out waiting for condition. Thread diagnostics:
Timestamp: 2015-02-21 06:37:17,092


http.TestHttpServerLifecycle  : Results were not captured by Maven 20 times.



Running org.apache.hadoop.ipc.TestRPC
====> TEST TIMED OUT. PRINTING THREAD DUMP. <====
Timestamp: 2015-02-22 07:01:59,579
"IPC Server handler 1 on 47469" daemon prio=5 tid=28 timed_waiting
java.lang.Thread.State: TIMED_WAITING
        at sun.misc.Unsafe.park(Native Method)
        at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:226)
        at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2082)
        at java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:467)
        at org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1985)


Running org.apache.hadoop.mapred.gridmix.TestGridMixClasses
Tests run: 12, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 7.916 sec <<< FAILURE!
- in org.apache.hadoop.mapred.gridmix.TestGridMixClasses
testSleepReducer(org.apache.hadoop.mapred.gridmix.TestGridMixClasses)  Time elapsed: 1.006
sec  <<< ERROR!
java.lang.Exception: test timed out after 1000 milliseconds


mapred.TestJobName  : Maven did not captured results once.


Running org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt
Tests run: 13, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 6.925 sec <<< FAILURE!
- in org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt
testTooManyFetchFailureAfterKill(org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt)
 Time elapsed: 0.326 sec  <<< ERROR!
java.lang.NullPointerException: null
        at org.apache.hadoop.security.token.Token.write(Token.java:221)
        at org.apache.hadoop.mapred.ShuffleHandler.serializeServiceData(ShuffleHandler.java:272)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.createCommonContainerLaunchContext(TaskAttemptImpl.java:715)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.createContainerLaunchContext(TaskAttemptImpl.java:801)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl$ContainerAssignedTransition.transition(TaskAttemptImpl.java:1516)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl$ContainerAssignedTransition.transition(TaskAttemptImpl.java:1493)
        at org.apache.hadoop.yarn.state.StateMachineFactory$SingleInternalArc.doTransition(StateMachineFactory.java:362)
        at org.apache.hadoop.yarn.state.StateMachineFactory.doTransition(StateMachineFactory.java:302)
        at org.apache.hadoop.yarn.state.StateMachineFactory.access$300(StateMachineFactory.java:46)
        at org.apache.hadoop.yarn.state.StateMachineFactory$InternalStateMachine.doTransition(StateMachineFactory.java:448)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.handle(TaskAttemptImpl.java:1058)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt.testTooManyFetchFailureAfterKill(TestTaskAttempt.java:660)

testFetchFailureAttemptFinishTime(org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt)
 Time elapsed: 0.022 sec  <<< ERROR!
java.lang.NullPointerException: null
        at org.apache.hadoop.security.token.Token.write(Token.java:221)
        at org.apache.hadoop.mapred.ShuffleHandler.serializeServiceData(ShuffleHandler.java:272)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.createCommonContainerLaunchContext(TaskAttemptImpl.java:715)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.createContainerLaunchContext(TaskAttemptImpl.java:801)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl$ContainerAssignedTransition.transition(TaskAttemptImpl.java:1516)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl$ContainerAssignedTransition.transition(TaskAttemptImpl.java:1493)
        at org.apache.hadoop.yarn.state.StateMachineFactory$SingleInternalArc.doTransition(StateMachineFactory.java:362)
        at org.apache.hadoop.yarn.state.StateMachineFactory.doTransition(StateMachineFactory.java:302)
        at org.apache.hadoop.yarn.state.StateMachineFactory.access$300(StateMachineFactory.java:46)
        at org.apache.hadoop.yarn.state.StateMachineFactory$InternalStateMachine.doTransition(StateMachineFactory.java:448)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TaskAttemptImpl.handle(TaskAttemptImpl.java:1058)
        at org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt.testFetchFailureAttemptFinishTime(TestTaskAttempt.java:771)


mapreduce.v2.app.TestRMContainerAllocator : Maven did not captured the results several times.


Running org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl
Tests run: 6, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 1.468 sec <<< FAILURE!
- in org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl
testMultiThreadedPublish(org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl)  Time elapsed:
0.037 sec  <<< FAILURE!
java.lang.AssertionError: Passed
Passed
Metric not collected!
Metric not collected!
Metric not collected!
Metric not collected!
Metric not collected!
Metric not collected!
Metric not collected!
Passed
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl.testMultiThreadedPublish(TestMetricsSystemImpl.java:233)


Running org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA
Tests run: 17, Failures: 1, Errors: 16, Skipped: 0, Time elapsed: 83.665 sec <<<
FAILURE! - in org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA
testGetApplicationsOnHA(org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA) 
Time elapsed: 5.247 sec  <<< ERROR!
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.io.IOException: NodeManager 0
failed to start
        at org.apache.hadoop.yarn.server.MiniYARNCluster$NodeManagerWrapper.serviceStart(MiniYARNCluster.java:572)
        at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
        at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at org.apache.hadoop.yarn.client.ProtocolHATestBase.startHACluster(ProtocolHATestBase.java:277)
        at org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA.initiate(TestApplicationClientProtocolOnHA.java:54)

testGetClusterNodesOnHA(org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA) 
Time elapsed: 13.848 sec  <<< FAILURE!
java.lang.AssertionError: Client couldn't connect to the Active RM
        at org.junit.Assert.fail(Assert.java:93)
        at org.apache.hadoop.yarn.client.ProtocolHATestBase.verifyClientConnection(ProtocolHATestBase.java:236)
        at org.apache.hadoop.yarn.client.ProtocolHATestBase.verifyConnections(ProtocolHATestBase.java:218)
        at org.apache.hadoop.yarn.client.ProtocolHATestBase.startHACluster(ProtocolHATestBase.java:280)
        at org.apache.hadoop.yarn.client.TestApplicationClientProtocolOnHA.initiate(TestApplicationClientProtocolOnHA.java:54)


Running org.apache.hadoop.yarn.client.TestRMFailover
Tests run: 5, Failures: 1, Errors: 4, Skipped: 0, Time elapsed: 41.336 sec <<< FAILURE!
- in org.apache.hadoop.yarn.client.TestRMFailover
testExplicitFailover(org.apache.hadoop.yarn.client.TestRMFailover)  Time elapsed: 5.268 sec
 <<< ERROR!
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.io.IOException: NodeManager 0
failed to start
        at org.apache.hadoop.yarn.server.MiniYARNCluster$NodeManagerWrapper.serviceStart(MiniYARNCluster.java:572)
        at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
        at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at org.apache.hadoop.yarn.client.TestRMFailover.testExplicitFailover(TestRMFailover.java:164)

testEmbeddedWebAppProxy(org.apache.hadoop.yarn.client.TestRMFailover)  Time elapsed: 23.594
sec  <<< FAILURE!
java.lang.AssertionError: NMs failed to connect to the RM
        at org.junit.Assert.fail(Assert.java:93)
        at org.junit.Assert.assertTrue(Assert.java:43)
        at org.apache.hadoop.yarn.client.TestRMFailover.verifyConnections(TestRMFailover.java:133)
        at org.apache.hadoop.yarn.client.TestRMFailover.testEmbeddedWebAppProxy(TestRMFailover.java:260)



Running org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler
Tests run: 58, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 3.944 sec <<< FAILURE!
- in org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler
testContinuousScheduling(org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler)
 Time elapsed: 0.538 sec  <<< FAILURE!
junit.framework.AssertionFailedError: expected:<2> but was:<1>
        at junit.framework.Assert.fail(Assert.java:50)
        at junit.framework.Assert.failNotEquals(Assert.java:287)
        at junit.framework.Assert.assertEquals(Assert.java:67)
        at junit.framework.Assert.assertEquals(Assert.java:199)
        at junit.framework.Assert.assertEquals(Assert.java:205)
        at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler.testContinuousScheduling(TestFairScheduler.java:2429)


> Several tests are not stable (on OpenJDK / x86_64 / RHEL 7.1)
> -------------------------------------------------------------
>
>                 Key: HADOOP-11617
>                 URL: https://issues.apache.org/jira/browse/HADOOP-11617
>             Project: Hadoop Common
>          Issue Type: Bug
>         Environment: RHEL 7.1 / x86_64
>            Reporter: Tony Reix
>
> I'm now porting Hadoop 2.4.1 on PPC64 HW.
> In order to have an idea about the stability of the tests, I've run several times all
Hadoop tests in a RHEL 7.1 / x86_64 / OpenJDK environment.
> I've run it 12 times, for now, on the EXACT same environment.
> Using some tool, I have extracted the list of tests that do not always produce the same
results, listed below, showing some test unstability.
> I 'm now doing the same for version 2.6.0 on OpenJDK/Ubuntu/x86_64.
> Report gives:
> test-name N: T F E S |
> where:
> - N is the number of the test log file
> - and:
> T: Tests run
> F: Failure
> E: Error
> S: Skipped
> As an example (see below for full data):
> hdfs.server.datanode.TestDeleteBlockPool                      0:    2   0   0   0 |
> ........
> hdfs.server.datanode.TestDeleteBlockPool                      4:    2   0   1   0 |
> ........
> hdfs.server.datanode.TestDeleteBlockPool                     11:    2   0   0   0 |
> means that this test, out of 12 runs, had an error once, with 0 Failure and 0 Skipped
test.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message