binayakd commented on code in PR #1354: URL: https://github.com/apache/iceberg-python/pull/1354#discussion_r1853355886
########## tests/io/test_pyarrow_stats.py: ########## @@ -681,6 +681,73 @@ def test_stats_types(table_schema_nested: Schema) -> None: ] +def construct_test_table_without_stats() -> Tuple[pq.FileMetaData, Union[TableMetadataV1, TableMetadataV2]]: + table_metadata = { + "format-version": 2, + "location": "s3://bucket/test/location", + "last-column-id": 7, + "current-schema-id": 0, + "schemas": [ + { + "type": "struct", + "schema-id": 0, + "fields": [ + {"id": 1, "name": "strings", "required": False, "type": "string"}, + {"id": 2, "name": "floats", "required": False, "type": "float"} + ] + } + ], + "default-spec-id": 0, + "partition-specs": [{"spec-id": 0, "fields": []}], + "properties": {}, + } + + table_metadata = TableMetadataUtil.parse_obj(table_metadata) + arrow_schema = schema_to_pyarrow(table_metadata.schemas[0]) + _strings = ["zzzzzzzzzzzzzzzzzzzz", "rrrrrrrrrrrrrrrrrrrr", None, "aaaaaaaaaaaaaaaaaaaa"] + _floats = [3.14, math.nan, 1.69, 100] + + table = pa.Table.from_pydict( + { + "strings": _strings, + "floats": _floats + }, + schema=arrow_schema, + ) + + metadata_collector: List[Any] = [] + + with pa.BufferOutputStream() as f: + with pq.ParquetWriter(f, table.schema, metadata_collector=metadata_collector, write_statistics=False) as writer: + writer.write_table(table) + + return metadata_collector[0], table_metadata + + +def test_is_stats_set_false() -> None: + metadata, table_metadata = construct_test_table_without_stats() + schema = get_current_schema(table_metadata) + statistics = data_file_statistics_from_parquet_metadata( + parquet_metadata=metadata, + stats_columns=compute_statistics_plan(schema, table_metadata.properties), + parquet_column_mapping=parquet_path_to_id_mapping(schema), + ) + datafile = DataFile(**statistics.to_serialized_dict()) + + # assert attributes except for column_aggregates and null_value_counts are present Review Comment: Rewrote the test to use the test table but only with the "strings" column having statistics. ```python # write statistics for only for "strings" column metadata, table_metadata = construct_test_table(write_statistics=["strings"]) ``` Added asserts to make sure the input metadata only has stats for the first (strings) column, and the rest don't have have any stats, specially the floats column (since the non-primitive columns get skipped in the iteration) ```python # expect only "strings" column to have statistics in metadata assert metadata.row_group(0).column(0).is_stats_set is True assert metadata.row_group(0).column(0).statistics is not None # expect all other columns to have no statistics for r in range(metadata.num_row_groups): for pos in range(1, metadata.num_columns): assert metadata.row_group(r).column(pos).is_stats_set is False assert metadata.row_group(r).column(pos).statistics is None ``` From what I understand `col_aggs` is used to compute the `upper_bound`, `lower_bound` if the `datafile`, so then we assert that the `upper_bound`, `lower_bound` and `null_value_counts` props of the `datafile` reflect only the values from the "strings" column, and no error is thrown: ```python # expect only "strings" column values to be reflected in the # upper_bound, lower_bound and null_value_counts props of datafile assert len(datafile.lower_bounds) == 1 assert datafile.lower_bounds[1].decode() == "aaaaaaaaaaaaaaaa" assert len(datafile.upper_bounds) == 1 assert datafile.upper_bounds[1].decode() == "zzzzzzzzzzzzzzz{" assert len(datafile.null_value_counts) == 1 assert datafile.null_value_counts[1] == 1 ``` This should cover the case of some columns having stats, and some not? Not sure if its a valid case. Hopefully this makes sense? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org