This is an automated email from the ASF dual-hosted git repository. jackie pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/pinot.git
The following commit(s) were added to refs/heads/master by this push: new b15155460a Do not hard code resource class in BaseClusterIntegrationTest (#13400) b15155460a is described below commit b15155460a98b921db6f8ee844b460ea267fcd7e Author: Xiaotian (Jackie) Jiang <17555551+jackie-ji...@users.noreply.github.com> AuthorDate: Fri Jun 14 19:04:43 2024 -0700 Do not hard code resource class in BaseClusterIntegrationTest (#13400) --- .../core/minion/PinotTaskManagerStatelessTest.java | 12 ++++---- .../tests/BaseClusterIntegrationTest.java | 35 ++++++++++------------ 2 files changed, 22 insertions(+), 25 deletions(-) diff --git a/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java b/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java index 85e4976639..175d11129c 100644 --- a/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java +++ b/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java @@ -152,8 +152,8 @@ public class PinotTaskManagerStatelessTest extends ControllerTest { "MergeRollupTask", ImmutableMap.of("schedule", "0 */10 * ? * * *")))); updateTableConfig(tableConfig); waitForJobGroupNames(_controllerStarter.getTaskManager(), - jgn -> jgn.size() == 2 && jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn - .contains(MinionConstants.MergeRollupTask.TASK_TYPE), + jgn -> jgn.size() == 2 && jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn.contains( + MinionConstants.MergeRollupTask.TASK_TYPE), "JobGroupNames should have SegmentGenerationAndPushTask and MergeRollupTask"); validateJob(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE, "0 */30 * ? * * *"); validateJob(MinionConstants.MergeRollupTask.TASK_TYPE, "0 */10 * ? * * *"); @@ -203,8 +203,10 @@ public class PinotTaskManagerStatelessTest extends ControllerTest { "JobGroupNames should have SegmentGenerationAndPushTask only"); validateJob(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE, "0 */10 * ? * * *"); - // Restart controller + // Restart controller. We need to set the port after stopping the controller because we are reusing the same config. + int controllerPort = _controllerPort; stopController(); + _controllerPort = controllerPort; startController(properties); // wait for controller to start correctly. TestUtils.waitForCondition((aVoid) -> { @@ -232,8 +234,8 @@ public class PinotTaskManagerStatelessTest extends ControllerTest { // The new MergeRollup task wouldn't be scheduled if not eagerly checking table configs // after setting up subscriber on ChildChanges zk event when controller gets restarted. waitForJobGroupNames(_controllerStarter.getTaskManager(), - jgn -> jgn.size() == 2 && jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn - .contains(MinionConstants.MergeRollupTask.TASK_TYPE), + jgn -> jgn.size() == 2 && jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn.contains( + MinionConstants.MergeRollupTask.TASK_TYPE), "JobGroupNames should have SegmentGenerationAndPushTask and MergeRollupTask"); dropOfflineTable(RAW_TABLE_NAME); diff --git a/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java b/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java index 0a127071da..33bd9b008e 100644 --- a/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java +++ b/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java @@ -270,7 +270,7 @@ public abstract class BaseClusterIntegrationTest extends ClusterTest { protected Schema createSchema(String schemaFileName) throws IOException { - InputStream inputStream = BaseClusterIntegrationTest.class.getClassLoader().getResourceAsStream(schemaFileName); + InputStream inputStream = getClass().getClassLoader().getResourceAsStream(schemaFileName); Assert.assertNotNull(inputStream); return Schema.fromInputStream(inputStream); } @@ -291,13 +291,13 @@ public abstract class BaseClusterIntegrationTest extends ClusterTest { * Creates a new OFFLINE table config. */ protected TableConfig createOfflineTableConfig() { - return new TableConfigBuilder(TableType.OFFLINE).setTableName(getTableName()) - .setTimeColumnName(getTimeColumnName()).setSortedColumn(getSortedColumn()) - .setInvertedIndexColumns(getInvertedIndexColumns()).setNoDictionaryColumns(getNoDictionaryColumns()) - .setRangeIndexColumns(getRangeIndexColumns()).setBloomFilterColumns(getBloomFilterColumns()) - .setFieldConfigList(getFieldConfigs()).setNumReplicas(getNumReplicas()).setSegmentVersion(getSegmentVersion()) - .setLoadMode(getLoadMode()).setTaskConfig(getTaskConfig()).setBrokerTenant(getBrokerTenant()) - .setServerTenant(getServerTenant()).setIngestionConfig(getIngestionConfig()).setQueryConfig(getQueryConfig()) + return new TableConfigBuilder(TableType.OFFLINE).setTableName(getTableName()).setTimeColumnName(getTimeColumnName()) + .setSortedColumn(getSortedColumn()).setInvertedIndexColumns(getInvertedIndexColumns()) + .setNoDictionaryColumns(getNoDictionaryColumns()).setRangeIndexColumns(getRangeIndexColumns()) + .setBloomFilterColumns(getBloomFilterColumns()).setFieldConfigList(getFieldConfigs()) + .setNumReplicas(getNumReplicas()).setSegmentVersion(getSegmentVersion()).setLoadMode(getLoadMode()) + .setTaskConfig(getTaskConfig()).setBrokerTenant(getBrokerTenant()).setServerTenant(getServerTenant()) + .setIngestionConfig(getIngestionConfig()).setQueryConfig(getQueryConfig()) .setNullHandlingEnabled(getNullHandlingEnabled()).setSegmentPartitionConfig(getSegmentPartitionConfig()) .build(); } @@ -399,9 +399,8 @@ public abstract class BaseClusterIntegrationTest extends ClusterTest { /** * Creates a new Upsert enabled table config. */ - protected TableConfig createCSVUpsertTableConfig(String tableName, @Nullable String kafkaTopicName, - int numPartitions, Map<String, String> streamDecoderProperties, UpsertConfig upsertConfig, - String primaryKeyColumn) { + protected TableConfig createCSVUpsertTableConfig(String tableName, @Nullable String kafkaTopicName, int numPartitions, + Map<String, String> streamDecoderProperties, UpsertConfig upsertConfig, String primaryKeyColumn) { Map<String, ColumnPartitionConfig> columnPartitionConfigMap = new HashMap<>(); columnPartitionConfigMap.put(primaryKeyColumn, new ColumnPartitionConfig("Murmur", numPartitions)); @@ -469,18 +468,14 @@ public abstract class BaseClusterIntegrationTest extends ClusterTest { Properties properties = getPinotConnectionProperties(); properties.put("useMultistageEngine", "true"); _pinotConnectionV2 = ConnectionFactory.fromZookeeper(getZkUrl() + "/" + getHelixClusterName(), - new JsonAsyncHttpPinotClientTransportFactory() - .withConnectionProperties(properties) - .buildTransport()); + new JsonAsyncHttpPinotClientTransportFactory().withConnectionProperties(properties).buildTransport()); } return _pinotConnectionV2; } if (_pinotConnection == null) { - _pinotConnection = - ConnectionFactory.fromZookeeper(getZkUrl() + "/" + getHelixClusterName(), - new JsonAsyncHttpPinotClientTransportFactory() - .withConnectionProperties(getPinotConnectionProperties()) - .buildTransport()); + _pinotConnection = ConnectionFactory.fromZookeeper(getZkUrl() + "/" + getHelixClusterName(), + new JsonAsyncHttpPinotClientTransportFactory().withConnectionProperties(getPinotConnectionProperties()) + .buildTransport()); } return _pinotConnection; } @@ -554,7 +549,7 @@ public abstract class BaseClusterIntegrationTest extends ClusterTest { */ protected List<File> unpackTarData(String tarFileName, File outputDir) throws Exception { - InputStream inputStream = BaseClusterIntegrationTest.class.getClassLoader().getResourceAsStream(tarFileName); + InputStream inputStream = getClass().getClassLoader().getResourceAsStream(tarFileName); Assert.assertNotNull(inputStream); return TarGzCompressionUtils.untar(inputStream, outputDir); } --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@pinot.apache.org For additional commands, e-mail: commits-h...@pinot.apache.org