ajantha-bhat commented on code in PR #11216:
URL: https://github.com/apache/iceberg/pull/11216#discussion_r1818891380


##########
data/src/test/java/org/apache/iceberg/data/TestPartitionStatsHandler.java:
##########
@@ -0,0 +1,569 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg.data;
+
+import static org.apache.iceberg.data.PartitionStatsHandler.Column;
+import static org.apache.iceberg.types.Types.NestedField.optional;
+import static org.apache.iceberg.types.Types.NestedField.required;
+import static org.assertj.core.api.Assertions.assertThat;
+import static org.assertj.core.api.Assertions.assertThatThrownBy;
+
+import java.io.File;
+import java.io.IOException;
+import java.util.Arrays;
+import java.util.Collections;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Random;
+import java.util.UUID;
+import java.util.concurrent.ThreadLocalRandom;
+import org.apache.iceberg.DataFile;
+import org.apache.iceberg.DeleteFile;
+import org.apache.iceberg.FileFormat;
+import org.apache.iceberg.Files;
+import org.apache.iceberg.Parameter;
+import org.apache.iceberg.ParameterizedTestExtension;
+import org.apache.iceberg.Parameters;
+import org.apache.iceberg.PartitionSpec;
+import org.apache.iceberg.PartitionStatisticsFile;
+import org.apache.iceberg.PartitionStats;
+import org.apache.iceberg.Partitioning;
+import org.apache.iceberg.Schema;
+import org.apache.iceberg.Snapshot;
+import org.apache.iceberg.SortOrder;
+import org.apache.iceberg.Table;
+import org.apache.iceberg.TableProperties;
+import org.apache.iceberg.TestHelpers;
+import org.apache.iceberg.TestTables;
+import org.apache.iceberg.deletes.PositionDelete;
+import org.apache.iceberg.expressions.Literal;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.io.OutputFile;
+import org.apache.iceberg.relocated.com.google.common.collect.ImmutableList;
+import org.apache.iceberg.relocated.com.google.common.collect.ImmutableMap;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.types.Types;
+import org.apache.iceberg.util.UUIDUtil;
+import org.assertj.core.groups.Tuple;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.TestTemplate;
+import org.junit.jupiter.api.extension.ExtendWith;
+import org.junit.jupiter.api.io.TempDir;
+
+@ExtendWith(ParameterizedTestExtension.class)
+public class TestPartitionStatsHandler {
+  private static final Schema SCHEMA =
+      new Schema(
+          optional(1, "c1", Types.IntegerType.get()),
+          optional(2, "c2", Types.StringType.get()),
+          optional(3, "c3", Types.StringType.get()));
+
+  protected static final PartitionSpec SPEC =
+      PartitionSpec.builderFor(SCHEMA).identity("c2").identity("c3").build();
+
+  @TempDir public File temp;
+
+  private static final Random RANDOM = ThreadLocalRandom.current();
+
+  @Parameters(name = "fileFormat = {0}")
+  public static List<Object> parameters() {
+    return Arrays.asList(FileFormat.PARQUET, FileFormat.ORC, FileFormat.AVRO);
+  }
+
+  @Parameter private FileFormat format;
+
+  @Test
+  public void testPartitionStatsOnEmptyTable() throws Exception {
+    Table testTable = TestTables.create(tempDir("empty_table"), "empty_table", 
SCHEMA, SPEC, 2);
+    
assertThat(PartitionStatsHandler.computeAndWriteStatsFile(testTable)).isNull();
+  }
+
+  @Test
+  public void testPartitionStatsOnEmptyBranch() throws Exception {
+    Table testTable = TestTables.create(tempDir("empty_branch"), 
"empty_branch", SCHEMA, SPEC, 2);
+    testTable.manageSnapshots().createBranch("b1").commit();
+    PartitionStatisticsFile partitionStatisticsFile =
+        PartitionStatsHandler.computeAndWriteStatsFile(testTable, "b1");
+    // creates an empty stats file since the dummy snapshot exist
+    assertThat(partitionStatisticsFile.fileSizeInBytes()).isEqualTo(0L);
+    assertThat(partitionStatisticsFile.snapshotId())
+        .isEqualTo(testTable.refs().get("b1").snapshotId());
+  }
+
+  @Test
+  public void testPartitionStatsOnInvalidSnapshot() throws Exception {
+    Table testTable =
+        TestTables.create(tempDir("invalid_snapshot"), "invalid_snapshot", 
SCHEMA, SPEC, 2);
+    assertThatThrownBy(
+            () -> PartitionStatsHandler.computeAndWriteStatsFile(testTable, 
"INVALID_BRANCH"))
+        .isInstanceOf(IllegalArgumentException.class)
+        .hasMessage("Couldn't find the snapshot for the branch 
INVALID_BRANCH");
+  }
+
+  @Test
+  public void testPartitionStatsOnUnPartitionedTable() throws Exception {
+    Table testTable =
+        TestTables.create(
+            tempDir("unpartitioned_table"),
+            "unpartitioned_table",
+            SCHEMA,
+            PartitionSpec.unpartitioned(),
+            2);
+
+    List<Record> records = prepareRecords(testTable.schema());
+    DataFile dataFile = FileHelpers.writeDataFile(testTable, outputFile(), 
records);
+    testTable.newAppend().appendFile(dataFile).commit();
+
+    assertThatThrownBy(() -> 
PartitionStatsHandler.computeAndWriteStatsFile(testTable))
+        .isInstanceOf(IllegalStateException.class)
+        .hasMessage("table must be partitioned");
+  }
+
+  @Test
+  public void testAllDatatypePartitionWriting() throws Exception {
+    Schema schema =
+        new Schema(
+            required(100, "id", Types.LongType.get()),
+            optional(101, "data", Types.StringType.get()),
+            required(102, "b", Types.BooleanType.get()),
+            optional(103, "i", Types.IntegerType.get()),
+            required(104, "l", Types.LongType.get()),
+            optional(105, "f", Types.FloatType.get()),
+            required(106, "d", Types.DoubleType.get()),
+            optional(107, "date", Types.DateType.get()),
+            required(108, "ts", Types.TimestampType.withoutZone()),
+            required(110, "s", Types.StringType.get()),
+            required(111, "uuid", Types.UUIDType.get()),
+            required(112, "fixed", Types.FixedType.ofLength(7)),
+            optional(113, "bytes", Types.BinaryType.get()),
+            required(114, "dec_9_0", Types.DecimalType.of(9, 0)),
+            required(115, "dec_11_2", Types.DecimalType.of(11, 2)),
+            required(116, "dec_38_10", Types.DecimalType.of(38, 10)), // 
maximum precision
+            required(117, "time", Types.TimeType.get()));
+
+    PartitionSpec spec =
+        PartitionSpec.builderFor(schema)
+            .identity("b")
+            .identity("i")
+            .identity("l")
+            .identity("f")
+            .identity("d")
+            .identity("date")
+            .identity("ts")
+            .identity("s")
+            .identity("uuid")
+            .identity("fixed")
+            .identity("bytes")
+            .identity("dec_9_0")
+            .identity("dec_11_2")
+            .identity("dec_38_10")
+            .identity("time")
+            .build();
+
+    Table testTable =
+        TestTables.create(
+            tempDir("test_all_type"), "test_all_type", schema, spec, 
SortOrder.unsorted(), 2);
+
+    Types.StructType partitionSchema = Partitioning.partitionType(testTable);
+    Schema dataSchema = PartitionStatsHandler.schema(partitionSchema);
+
+    Record partitionData =
+        
GenericRecord.create(dataSchema.findField(Column.PARTITION.name()).type().asStructType());
+    partitionData.set(0, true);
+    partitionData.set(1, 42);
+    partitionData.set(2, 42L);
+    partitionData.set(3, 3.14f);
+    partitionData.set(4, 3.141592653589793);
+    partitionData.set(5, 
Literal.of("2022-01-01").to(Types.DateType.get()).value());
+    partitionData.set(
+        6, 
Literal.of("2017-12-01T10:12:55.038194").to(Types.TimestampType.withoutZone()).value());
+    partitionData.set(7, "string");
+    partitionData.set(8, UUIDUtil.convertToByteBuffer(UUID.randomUUID()));
+    partitionData.set(9, new byte[] {0, 1, 2, 3, 4, 5, 6});
+    partitionData.set(10, new byte[] {1, 2, 3});
+    partitionData.set(11, Literal.of("123456789").to(Types.DecimalType.of(9, 
0)).value());
+    partitionData.set(12, Literal.of("1234567.89").to(Types.DecimalType.of(11, 
2)).value());
+    partitionData.set(
+        13, 
Literal.of("12345678901234567890.1234567890").to(Types.DecimalType.of(38, 
10)).value());
+    partitionData.set(14, 
Literal.of("10:10:10").to(Types.TimeType.get()).value());
+
+    PartitionStats partitionStats = new PartitionStats(partitionData, 
RANDOM.nextInt(10));
+    partitionStats.set(Column.DATA_RECORD_COUNT.id(), RANDOM.nextLong());
+    partitionStats.set(Column.DATA_FILE_COUNT.id(), RANDOM.nextInt());
+    partitionStats.set(Column.TOTAL_DATA_FILE_SIZE_IN_BYTES.id(), 1024L * 
RANDOM.nextInt(20));
+
+    Iterator<PartitionStatsRecord> convertedRecords =
+        
PartitionStatsHandler.statsToRecords(Collections.singletonList(partitionStats), 
dataSchema);

Review Comment:
   There are already other tests below (for example `testPartitionStats`) that 
checks multiple rows/records. 
   
   This testcase is to check one row of "all possible data types" in partition 
values. I don't see the benefit of having more rows here.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to