[ 
https://issues.apache.org/jira/browse/HADOOP-14201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15931828#comment-15931828
 ] 

Steve Loughran commented on HADOOP-14201:
-----------------------------------------

{code}
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.555 sec <<< 
FAILURE! - in org.apache.hadoop.fs.TestFsS
hellList
testList(org.apache.hadoop.fs.TestFsShellList)  Time elapsed: 0.095 sec  <<< 
ERROR!
org.apache.hadoop.io.nativeio.NativeIOException: The filename, directory name, 
or volume label syntax is incorrect.

        at 
org.apache.hadoop.io.nativeio.NativeIO$Windows.createFileWithMode0(Native 
Method)
        at 
org.apache.hadoop.io.nativeio.NativeIO$Windows.createFileOutputStreamWithMode(NativeIO.java:556)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:231)
        at 
org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.<init>(RawLocalFileSystem.java:221)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.createOutputStreamWithMode(RawLocalFileSystem.java:319)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:308)
        at 
org.apache.hadoop.fs.RawLocalFileSystem.create(RawLocalFileSystem.java:339)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.<init>(ChecksumFileSystem.java:399)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:462)
        at 
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:441)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:928)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:806)
        at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:795)
        at 
org.apache.hadoop.fs.TestFsShellList.createFile(TestFsShellList.java:57)
        at 
org.apache.hadoop.fs.TestFsShellList.testList(TestFsShellList.java:69)

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.fs.TestDU
Tests run: 3, Failures: 2, Errors: 0, Skipped: 0, Time elapsed: 17.349 sec <<< 
FAILURE! - in org.apache.hadoop.fs.TestDU

testDU(org.apache.hadoop.fs.TestDU)  Time elapsed: 11.058 sec  <<< FAILURE!
junit.framework.AssertionFailedError: Invalid on-disk size
        at junit.framework.Assert.fail(Assert.java:57)
        at junit.framework.Assert.assertTrue(Assert.java:22)
        at junit.framework.TestCase.assertTrue(TestCase.java:192)
        at org.apache.hadoop.fs.TestDU.testDU(TestDU.java:87)

testDUSetInitialValue(org.apache.hadoop.fs.TestDU)  Time elapsed: 6.084 sec  
<<< FAILURE!
junit.framework.AssertionFailedError: Usage didn't get updated
        at junit.framework.Assert.fail(Assert.java:57)
        at junit.framework.Assert.assertTrue(Assert.java:22)
        at junit.framework.TestCase.assertTrue(TestCase.java:192)
        at org.apache.hadoop.fs.TestDU.testDUSetInitialValue(TestDU.java:133)
        
        
        testAppendBlockCompression(org.apache.hadoop.io.TestSequenceFileAppend) 
 Time elapsed: 0.028 sec  <<< ERROR!
java.io.IOException: not a gzip file
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.processBasicHeader(BuiltInGzipDecompressor.java:49
6)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.executeHeaderState(BuiltInGzipDecompressor.java:25
7)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.decompress(BuiltInGzipDecompressor.java:186)
        at 
org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:111)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:105)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:92)
        at java.io.DataInputStream.readByte(DataInputStream.java:265)
        at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:308)
        at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:329)
        at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:2612)
        at 
org.apache.hadoop.io.TestSequenceFileAppend.verify2Values(TestSequenceFileAppend.java:362)
        at 
org.apache.hadoop.io.TestSequenceFileAppend.testAppendBlockCompression(TestSequenceFileAppend.java:194)

testAppendSort(org.apache.hadoop.io.TestSequenceFileAppend)  Time elapsed: 
0.024 sec  <<< ERROR!
java.io.IOException: not a gzip file
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.processBasicHeader(BuiltInGzipDecompressor.java:49
6)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.executeHeaderState(BuiltInGzipDecompressor.java:25
7)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.decompress(BuiltInGzipDecompressor.java:186)
        at 
org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:111)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:105)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:92)
        at java.io.DataInputStream.readByte(DataInputStream.java:265)
        at org.apache.hadoop.io.WritableUtils.readVLong(WritableUtils.java:308)
        at org.apache.hadoop.io.WritableUtils.readVInt(WritableUtils.java:329)
        at 
org.apache.hadoop.io.SequenceFile$Reader.nextRaw(SequenceFile.java:2516)
        at 
org.apache.hadoop.io.SequenceFile$Sorter$SortPass.run(SequenceFile.java:2947)
        at 
org.apache.hadoop.io.SequenceFile$Sorter.sortPass(SequenceFile.java:2885)
        at org.apache.hadoop.io.SequenceFile$Sorter.sort(SequenceFile.java:2833)
        at org.apache.hadoop.io.SequenceFile$Sorter.sort(SequenceFile.java:2874)
        at 
org.apache.hadoop.io.TestSequenceFileAppend.testAppendSort(TestSequenceFileAppend.java:353)

testAppendRecordCompression(org.apache.hadoop.io.TestSequenceFileAppend)  Time 
elapsed: 0.017 sec  <<< ERROR!
java.io.IOException: not a gzip file
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.processBasicHeader(BuiltInGzipDecompressor.java:49
6)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.executeHeaderState(BuiltInGzipDecompressor.java:25
7)
        at 
org.apache.hadoop.io.compress.zlib.BuiltInGzipDecompressor.decompress(BuiltInGzipDecompressor.java:186)
        at 
org.apache.hadoop.io.compress.DecompressorStream.decompress(DecompressorStream.java:111)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:105)
        at 
org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:92)
        at java.io.FilterInputStream.read(FilterInputStream.java:83)
        at 
java.io.ObjectInputStream$PeekInputStream.peek(ObjectInputStream.java:2303)
        at 
java.io.ObjectInputStream$BlockDataInputStream.peek(ObjectInputStream.java:2596)
        at 
java.io.ObjectInputStream$BlockDataInputStream.peekByte(ObjectInputStream.java:2606)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1319)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371)
        at 
org.apache.hadoop.io.serializer.JavaSerialization$JavaSerializationDeserializer.deserialize(JavaSerialization
.java:59)
        at 
org.apache.hadoop.io.serializer.JavaSerialization$JavaSerializationDeserializer.deserialize(JavaSerialization
.java:40)
        at 
org.apache.hadoop.io.SequenceFile$Reader.deserializeValue(SequenceFile.java:2343)
        at 
org.apache.hadoop.io.SequenceFile$Reader.getCurrentValue(SequenceFile.java:2316)
        at 
org.apache.hadoop.io.TestSequenceFileAppend.verify2Values(TestSequenceFileAppend.java:363)
        at 
org.apache.hadoop.io.TestSequenceFileAppend.testAppendRecordCompression(TestSequenceFileAppend.java:160)


Tests run: 37, Failures: 1, Errors: 1, Skipped: 1, Time elapsed: 149.693 sec 
<<< FAILURE! - in org.apache.hadoop.ipc.Tes
tIPC
testInsecureVersionMismatch(org.apache.hadoop.ipc.TestIPC)  Time elapsed: 0.02 
sec  <<< ERROR!
java.io.IOException: Failed on local exception: java.io.IOException: An 
established connection was aborted by the softwa
re in your host machine; Host Details : local host is: "morzine/192.168.1.24"; 
destination host is: "0.0.0.0":53790;
        at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:785)
        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1485)
        at org.apache.hadoop.ipc.Client.call(Client.java:1427)
        at org.apache.hadoop.ipc.TestIPC.call(TestIPC.java:155)
        at org.apache.hadoop.ipc.TestIPC.call(TestIPC.java:148)
        at org.apache.hadoop.ipc.TestIPC.checkVersionMismatch(TestIPC.java:1450)
        at 
org.apache.hadoop.ipc.TestIPC.testInsecureVersionMismatch(TestIPC.java:1415)
Caused by: java.io.IOException: An established connection was aborted by the 
software in your host machine
        at sun.nio.ch.SocketDispatcher.read0(Native Method)
        at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:43)
        at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
        at sun.nio.ch.IOUtil.read(IOUtil.java:197)
        at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)
        at 
org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57)
        at 
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
        at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
        at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131)
        at java.io.FilterInputStream.read(FilterInputStream.java:133)
        at java.io.BufferedInputStream.fill(BufferedInputStream.java:246)
        at java.io.BufferedInputStream.read(BufferedInputStream.java:265)
        at java.io.FilterInputStream.read(FilterInputStream.java:83)
        at java.io.FilterInputStream.read(FilterInputStream.java:83)
        at 
org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:553)
        at java.io.DataInputStream.readInt(DataInputStream.java:387)
        at 
org.apache.hadoop.ipc.Client$IpcStreams.readResponse(Client.java:1786)
        at 
org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1155)
        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1052)

testConnectionIdleTimeouts(org.apache.hadoop.ipc.TestIPC)  Time elapsed: 4.181 
sec  <<< FAILURE!
java.lang.AssertionError: expected:<7> but was:<4>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:555)
        at org.junit.Assert.assertEquals(Assert.java:542)
        at 
org.apache.hadoop.ipc.TestIPC.testConnectionIdleTimeouts(TestIPC.java:949)

        Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.726 
sec - in org.apache.hadoop.metrics2.impl.Te
        ueue
        Java HotSpot(TM) 64-Bit Server VM warning: ignoring option 
MaxPermSize=768m; support was removed in 8.0
          
Running org.apache.hadoop.metrics2.impl.TestStatsDMetrics
Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 0.791 sec <<< 
FAILURE! - in org.apache.hadoop.met
mpl.TestStatsDMetrics
testPutMetrics2(org.apache.hadoop.metrics2.impl.TestStatsDMetrics)  Time 
elapsed: 0.552 sec  <<< ERROR!
org.apache.hadoop.metrics2.MetricsException: Error writing metric to StatsD
        at java.net.TwoStacksPlainDatagramSocketImpl.send(Native Method)
        at java.net.DatagramSocket.send(DatagramSocket.java:693)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink$StatsD.write(StatsDSink.java:203)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink.writeMetric(StatsDSink.java:151)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink.putMetrics(StatsDSink.java:144)
        at 
org.apache.hadoop.metrics2.impl.TestStatsDMetrics.testPutMetrics2(TestStatsDMetrics.java:109)

testPutMetrics(org.apache.hadoop.metrics2.impl.TestStatsDMetrics)  Time 
elapsed: 0.005 sec  <<< ERROR!
org.apache.hadoop.metrics2.MetricsException: Error writing metric to StatsD
        at java.net.TwoStacksPlainDatagramSocketImpl.send(Native Method)
        at java.net.DatagramSocket.send(DatagramSocket.java:693)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink$StatsD.write(StatsDSink.java:203)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink.writeMetric(StatsDSink.java:151)
        at 
org.apache.hadoop.metrics2.sink.StatsDSink.putMetrics(StatsDSink.java:144)
        at 
org.apache.hadoop.metrics2.impl.TestStatsDMetrics.testPutMetrics(TestStatsDMetrics.java:74)


testProxyUserFromEnvironment(org.apache.hadoop.security.TestProxyUserFromEnv)  
Time elapsed: 0.58 sec  <<< FAILURE!
org.junit.ComparisonFailure: expected:<[a]dministrator> but 
was:<[A]dministrator>
        at org.junit.Assert.assertEquals(Assert.java:115)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.security.TestProxyUserFromEnv.testProxyUserFromEnvironment(TestProxyUserFromEnv.java:54)

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0

  
Running org.apache.hadoop.security.TestShellBasedUnixGroupsMapping
Tests run: 4, Failures: 3, Errors: 0, Skipped: 0, Time elapsed: 0.367 sec <<< 
FAILURE! - in org.apache.hadoop.security.T
estShellBasedUnixGroupsMapping
testGetNumericGroupsResolvable(org.apache.hadoop.security.TestShellBasedUnixGroupsMapping)
  Time elapsed: 0.206 sec  <<<
 FAILURE!
java.lang.AssertionError: null
        at org.junit.Assert.fail(Assert.java:86)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at org.junit.Assert.assertTrue(Assert.java:52)
        at 
org.apache.hadoop.security.TestShellBasedUnixGroupsMapping.testGetNumericGroupsResolvable(TestShellBasedUnixG
roupsMapping.java:160)

testGetGroupsNotResolvable(org.apache.hadoop.security.TestShellBasedUnixGroupsMapping)
  Time elapsed: 0.003 sec  <<< FAI
LURE!
java.lang.AssertionError: null
        at org.junit.Assert.fail(Assert.java:86)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at org.junit.Assert.assertTrue(Assert.java:52)
        at 
org.apache.hadoop.security.TestShellBasedUnixGroupsMapping.testGetGroupsNotResolvable(TestShellBasedUnixGroup
sMapping.java:112)

testGetGroupsResolvable(org.apache.hadoop.security.TestShellBasedUnixGroupsMapping)
  Time elapsed: 0.002 sec  <<< FAILUR
E!
java.lang.AssertionError: null
        at org.junit.Assert.fail(Assert.java:86)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at org.junit.Assert.assertTrue(Assert.java:52)
        at 
org.apache.hadoop.security.TestShellBasedUnixGroupsMapping.testGetGroupsResolvable(TestShellBasedUnixGroupsMa
pping.java:206)


Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.506 sec - in 
org.apache.hadoop.test.TestJUnitSetup
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.test.TestLambdaTestUtils
Tests run: 11, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.959 sec <<< 
FAILURE! - in org.apache.hadoop.test.Test
LambdaTestUtils
testAwaitAlwaysFalse(org.apache.hadoop.test.TestLambdaTestUtils)  Time elapsed: 
0.076 sec  <<< FAILURE!
java.lang.AssertionError: null
        at org.junit.Assert.fail(Assert.java:86)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at org.junit.Assert.assertTrue(Assert.java:52)
        at 
org.apache.hadoop.test.TestLambdaTestUtils.testAwaitAlwaysFalse(TestLambdaTestUtils.java:141)

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.test.TestMultithreadedTestUtil
Tests run: 4, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 5.197 sec <<< 
FAILURE! - in org.apache.hadoop.test.TestM
ultithreadedTestUtil
testRepeatingThread(org.apache.hadoop.test.TestMultithreadedTestUtil)  Time 
elapsed: 4.522 sec  <<< FAILURE!
java.lang.AssertionError: Test took 4500ms
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at 
org.apache.hadoop.test.TestMultithreadedTestUtil.testRepeatingThread(TestMultithreadedTestUtil.java:132)

Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.test.TestTimedOutTestsListener
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.686 sec - in 
org.apache.hadoop.test.TestTimedOutTestsL
  
  
Running org.apache.hadoop.util.TestWinUtils
Tests run: 11, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.502 sec <<< 
FAILURE! - in org.apache.hadoop.util.Test
WinUtils
testTaskCreateWithLimits(org.apache.hadoop.util.TestWinUtils)  Time elapsed: 
0.814 sec  <<< FAILURE!
java.lang.AssertionError: Failed to get Shell.ExitCodeException with 
insufficient memory
  at org.junit.Assert.fail(Assert.java:88)
  at 
org.apache.hadoop.util.TestWinUtils.testTaskCreateWithLimits(TestWinUtils.java:605)

testNodeHealthScript(org.apache.hadoop.util.TestNodeHealthScriptRunner)  Time 
elapsed: 0.292 sec  <<< ERROR!
java.lang.NullPointerException: null
        at 
org.apache.hadoop.util.TestNodeHealthScriptRunner.writeNodeHealthScriptFile(TestNodeHealthScriptRunner.java:6
8)
        at 
org.apache.hadoop.util.TestNodeHealthScriptRunner.testNodeHealthScript(TestNodeHealthScriptRunner.java:112)

 {code}

> Fix some failing tests on windows
> ---------------------------------
>
>                 Key: HADOOP-14201
>                 URL: https://issues.apache.org/jira/browse/HADOOP-14201
>             Project: Hadoop Common
>          Issue Type: Task
>          Components: test
>    Affects Versions: 2.8.0
>         Environment: Windows Server 2012.
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>            Priority: Minor
>
> Some of the 2.8.0 tests are failing locally, without much in the way of 
> diagnostics. They may be false alarms related to system, VM setup, 
> performance, or they may be a sign of a problem.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to