lakshmanan-v commented on a change in pull request #7226: URL: https://github.com/apache/pinot/pull/7226#discussion_r679301301
########## File path: pinot-core/src/main/java/org/apache/pinot/core/query/aggregation/function/PercentileRawEstAggregationFunction.java ########## @@ -0,0 +1,137 @@ +/** + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ +package org.apache.pinot.core.query.aggregation.function; + +import java.util.Map; +import org.apache.pinot.common.request.context.ExpressionContext; +import org.apache.pinot.common.utils.DataSchema.ColumnDataType; +import org.apache.pinot.core.common.BlockValSet; +import org.apache.pinot.core.query.aggregation.AggregationResultHolder; +import org.apache.pinot.core.query.aggregation.groupby.GroupByResultHolder; +import org.apache.pinot.segment.local.customobject.QuantileDigest; +import org.apache.pinot.segment.local.utils.CustomSerDeUtils; +import org.apache.pinot.segment.spi.AggregationFunctionType; +import org.apache.pinot.spi.utils.BytesUtils; + + +public class PercentileRawEstAggregationFunction extends BaseSingleInputAggregationFunction<QuantileDigest, String> { + private final PercentileEstAggregationFunction _percentileRawEstAggregationFunction; + + public PercentileRawEstAggregationFunction(ExpressionContext expressionContext, double percentile) { + this(expressionContext, new PercentileEstAggregationFunction(expressionContext, percentile)); + } + + public PercentileRawEstAggregationFunction(ExpressionContext expressionContext, int percentile) { + this(expressionContext, new PercentileEstAggregationFunction(expressionContext, percentile)); + } + + PercentileRawEstAggregationFunction(ExpressionContext expression, + PercentileEstAggregationFunction percentileRawEstAggregationFunction) { + super(expression); + _percentileRawEstAggregationFunction = percentileRawEstAggregationFunction; + } + + @Override + public AggregationFunctionType getType() { + return AggregationFunctionType.PERCENTILERAWEST; + } + + @Override + public String getColumnName() { + final double percentile = _percentileRawEstAggregationFunction._percentile; + final int version = _percentileRawEstAggregationFunction._version; + final String type = getType().getName(); + + return version == 0 ? type + (int) percentile + "_" + _expression : type + percentile + "_" + _expression; + } + + @Override + public String getResultColumnName() { + final double percentile = _percentileRawEstAggregationFunction._percentile; + final int version = _percentileRawEstAggregationFunction._version; + final String type = getType().getName().toLowerCase(); + + return version == 0 ? type + (int) percentile + "(" + _expression + ")" + : type + "(" + _expression + ", " + percentile + ")"; + } + + @Override + public AggregationResultHolder createAggregationResultHolder() { + return _percentileRawEstAggregationFunction.createAggregationResultHolder(); + } + + @Override + public GroupByResultHolder createGroupByResultHolder(int initialCapacity, int maxCapacity) { + return _percentileRawEstAggregationFunction.createGroupByResultHolder(initialCapacity, maxCapacity); + } + + @Override + public void aggregate(int length, AggregationResultHolder aggregationResultHolder, + Map<ExpressionContext, BlockValSet> blockValSetMap) { + _percentileRawEstAggregationFunction.aggregate(length, aggregationResultHolder, blockValSetMap); + } + + @Override + public void aggregateGroupBySV(int length, int[] groupKeyArray, GroupByResultHolder groupByResultHolder, + Map<ExpressionContext, BlockValSet> blockValSetMap) { + _percentileRawEstAggregationFunction.aggregateGroupBySV(length, groupKeyArray, groupByResultHolder, blockValSetMap); + } + + @Override + public void aggregateGroupByMV(int length, int[][] groupKeysArray, GroupByResultHolder groupByResultHolder, + Map<ExpressionContext, BlockValSet> blockValSetMap) { + _percentileRawEstAggregationFunction + .aggregateGroupByMV(length, groupKeysArray, groupByResultHolder, blockValSetMap); + } + + @Override + public QuantileDigest extractAggregationResult(AggregationResultHolder aggregationResultHolder) { + return _percentileRawEstAggregationFunction.extractAggregationResult(aggregationResultHolder); + } + + @Override + public QuantileDigest extractGroupByResult(GroupByResultHolder groupByResultHolder, int groupKey) { + return _percentileRawEstAggregationFunction.extractGroupByResult(groupByResultHolder, groupKey); + } + + @Override + public QuantileDigest merge(QuantileDigest intermediateResult1, QuantileDigest intermediateResult2) { + return _percentileRawEstAggregationFunction.merge(intermediateResult1, intermediateResult2); + } + + @Override + public boolean isIntermediateResultComparable() { + return _percentileRawEstAggregationFunction.isIntermediateResultComparable(); + } + + @Override + public ColumnDataType getIntermediateResultColumnType() { + return _percentileRawEstAggregationFunction.getIntermediateResultColumnType(); + } + + @Override + public ColumnDataType getFinalResultColumnType() { + return ColumnDataType.STRING; + } + + @Override + public String extractFinalResult(QuantileDigest intermediateResult) { + return BytesUtils.toHexString(CustomSerDeUtils.QUANTILE_DIGEST_SER_DE.serialize(intermediateResult)); Review comment: @mayankshriv Some of these serialized hex strings are little longer (about 3K bytes). Thinking of compressing this using [Snappy](https://github.com/xerial/snappy-java) or [LZ4](https://github.com/lz4/lz4-java). Did a quick test using LZ4 and got the following results. Pinot seems to [use](https://github.com/apache/pinot/blob/b826f2f588bc8d0d05ef466468c2317119ce46cc/pinot-segment-local/src/main/java/org/apache/pinot/segment/local/io/compression/LZ4Compressor.java) LZ4 to compress data in segments already. Original Size = 3904; Compressed size = 1599; Compression rate = 59.04200819672132 ` LZ4Factory factory = LZ4Factory.fastestInstance(); LZ4Compressor compressor = factory.highCompressor(17); int maxCompressedLength = compressor.maxCompressedLength(decompressedLength); byte[] compressed = new byte[maxCompressedLength]; int compressedSize = compressor.compress(data, 0, decompressedLength, compressed, 0, maxCompressedLength); double ratio = (1 - compressedSize / (double) originalSize) * 100.0;` Was wondering if we should convert it back to base64 string or just return byte array (lz4 produces byte array). Let me know if you have any other ideas around compressing the results. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@pinot.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@pinot.apache.org For additional commands, e-mail: commits-h...@pinot.apache.org