steveloughran commented on PR #8007:
URL: https://github.com/apache/hadoop/pull/8007#issuecomment-3554136829

   checksyles were about some spaces, an unused import and line length. Can you 
do the line length for the code lines, but don't worry about the comments?
   ```
   
./hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/statistics/StreamStatisticNames.java:500:
  public static final String STREAM_READ_PARQUET_FOOTER_PARSING_FAILED = 
"stream_read_parquet_footer_parsing_failed";: Line is longer than 100 
characters (found 117). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:31:import
 org.apache.hadoop.fs.s3a.S3AUtils;:8: Unused import - 
org.apache.hadoop.fs.s3a.S3AUtils. [UnusedImports]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:74:
  private static final String REQUEST_COALESCE_TOLERANCE_KEY = 
ANALYTICS_ACCELERATOR_CONFIGURATION_PREFIX + ".": Line is longer than 100 
characters (found 111). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:76:
  private static final String READ_BUFFER_SIZE_KEY = 
ANALYTICS_ACCELERATOR_CONFIGURATION_PREFIX + ".": Line is longer than 100 
characters (found 101). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:78:
  private static final String SMALL_OBJECT_PREFETCH_ENABLED_KEY = 
ANALYTICS_ACCELERATOR_CONFIGURATION_PREFIX + ".": Line is longer than 100 
characters (found 114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:102:
    // Set the minimum block size to 32KB. AAL uses a default block size of 
128KB, which means the minimum size a S3: Line is longer than 100 characters 
(found 116). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:103:
    // request will be is 128KB. Since the file being read is 128KB, we need to 
 use this here to demonstrate that: Line is longer than 100 characters (found 
114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:153:
    fileRanges.add(FileRange.createFileRange(4 * S_1K , 4 * S_1K));:55: ',' is 
preceded with whitespace. [NoWhitespaceBefore]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:154:
    fileRanges.add(FileRange.createFileRange(80 * S_1K , 4 * S_1K));:56: ',' is 
preceded with whitespace. [NoWhitespaceBefore]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:169:
      // Verify ranges are coalesced, we are using a coalescing tolerance of 
16KB, so [0-100, 800-200, 4KB-8KB] will: Line is longer than 100 characters 
(found 116). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:176:
      // read the same ranges again to demonstrate that the data is cached, and 
no new GETs are made.: Line is longer than 100 characters (found 101). 
[LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:180:
      // Because of how AAL is currently written, it is not possible to track 
cache hits that originate from a: Line is longer than 100 characters (found 
110). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/contract/s3a/ITestS3AContractAnalyticsStreamVectoredRead.java:181:
      // readVectored() accurately. For this reason, cache hits from 
readVectored are currently not tracked, for more: Line is longer than 100 
characters (found 117). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:108:
   long fileLength = fs.getFileStatus(externalTestFile).getLen();: 'method def' 
child has incorrect indentation level 3, expected level should be 4. 
[Indentation]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:111:
    verifyStatisticCounterValue(fs.getIOStatistics(), AUDIT_REQUEST_EXECUTION, 
initialAuditCount + 1);: Line is longer than 100 characters (found 102). 
[LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:141:
    // Since policy is WHOLE_FILE, the whole file starts getting prefetched as 
soon as the stream to it is opened.: Line is longer than 100 characters (found 
114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:152:
    verifyStatisticCounterValue(fs.getIOStatistics(), AUDIT_REQUEST_EXECUTION, 
initialAuditCount + 1 + 4);: Line is longer than 100 characters (found 106). 
[LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:160:
    // AAL uses a caffeine cache, and expires any prefetched data for a key 1s 
after it was last accessed by default.: Line is longer than 100 characters 
(found 117). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:161:
    // While this works well when running on EC2, for local testing, it can 
take more than 1s to download large chunks: Line is longer than 100 characters 
(found 118). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:173:
    // Here we read through the 21MB external test file, but do not pass in the 
WHOLE_FILE policy. Instead, we rely: Line is longer than 100 characters (found 
115). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:174:
    // on AAL detecting a sequential pattern being read, and then prefetching 
bytes in a geometrical progression.: Line is longer than 100 characters (found 
113). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:175:
    // AAL's sequential prefetching starts prefetching in increments 4MB, 8MB, 
16MB etc. depending on how many: Line is longer than 100 characters (found 
110). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:187:
      // These next two reads are within the last prefetched bytes, so no 
further bytes are prefetched.: Line is longer than 100 characters (found 103). 
[LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:196:
      // Cache hit is still 2, as the previous read required a new GET request 
as it was outside the previously fetched: Line is longer than 100 characters 
(found 119). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:205:
      // Though the next GP should prefetch 16MB, since the file is ~23MB, only 
the bytes till EoF are prefetched.: Line is longer than 100 characters (found 
114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:206:
      verifyStatisticCounterValue(ioStats, STREAM_READ_PREFETCHED_BYTES, 10 * 
ONE_MB + bytesRemainingForPrefetch);: Line is longer than 100 characters (found 
114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:253:
    // This file has a content length of 451. Since it's a parquet file, AAL 
will prefetch the footer bytes (last 32KB),: Line is longer than 100 characters 
(found 120). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:260:
    // Open a stream to the object twice, verifying that data is cached, and 
streams to the same object, do not: Line is longer than 100 characters (found 
111). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:314:
    // S3A makes a HEAD request on the stream open(), and then AAL makes a GET 
request to get the object, total audit: Line is longer than 100 characters 
(found 117). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:317:
    verifyStatisticCounterValue(getFileSystem().getIOStatistics(), 
AUDIT_REQUEST_EXECUTION, currentAuditCount);: Line is longer than 100 
characters (found 111). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:399:
      stream1.read(buffer, 0 , 10 * ONE_KB);:30: ',' is preceded with 
whitespace. [NoWhitespaceBefore]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:407:
      // Since it's a small file (ALL will prefetch the whole file for size < 
8MB), the whole file is prefetched: Line is longer than 100 characters (found 
112). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:411:
      // The second stream will not prefetch any bytes, as they have already 
been prefetched by stream 1.: Line is longer than 100 characters (found 105). 
[LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:416:
    verifyStatisticCounterValue(getFileSystem().getIOStatistics(), 
STREAM_READ_PREFETCHED_BYTES, fileLen);: Line is longer than 100 characters 
(found 106). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/ITestS3AAnalyticsAcceleratorStreamReading.java:418:
    // We did 3 reads, all of them were served from the small object cache. In 
this case, the whole object was: Line is longer than 100 characters (found 
110). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/performance/ITestS3AOpenCost.java:187:
    // If AAL is enabled, skip this test. AAL uses S3A's default S3 client, and 
if checksumming is disabled on the: Line is longer than 100 characters (found 
114). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/performance/ITestS3AOpenCost.java:459:
        // For AAL, if there is no eTag, the provided length will not be passed 
in, and a HEAD request will be made.: Line is longer than 100 characters (found 
116). [LineLength]
   
./hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/performance/ITestS3AOpenCost.java:460:
        // AAL requires the etag to detect changes in the object and then do 
cache eviction if required.: Line is longer than 100 characters (found 104). 
[LineLength]
   ```
   
   the javadoc comments may be existing, but just close to changed code...so 
add something to shut it up
   ```
   
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/impl/StoreContext.java:184:
 warning: no comment
   
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/impl/StoreContext.java:213:
 warning: no comment
   
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/impl/StoreContext.java:217:
 warning: no comment
   ```
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to