This is an automated email from the ASF dual-hosted git repository.

jackie pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/pinot.git


The following commit(s) were added to refs/heads/master by this push:
     new b15155460a Do not hard code resource class in 
BaseClusterIntegrationTest (#13400)
b15155460a is described below

commit b15155460a98b921db6f8ee844b460ea267fcd7e
Author: Xiaotian (Jackie) Jiang <17555551+jackie-ji...@users.noreply.github.com>
AuthorDate: Fri Jun 14 19:04:43 2024 -0700

    Do not hard code resource class in BaseClusterIntegrationTest (#13400)
---
 .../core/minion/PinotTaskManagerStatelessTest.java | 12 ++++----
 .../tests/BaseClusterIntegrationTest.java          | 35 ++++++++++------------
 2 files changed, 22 insertions(+), 25 deletions(-)

diff --git 
a/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java
 
b/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java
index 85e4976639..175d11129c 100644
--- 
a/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java
+++ 
b/pinot-controller/src/test/java/org/apache/pinot/controller/helix/core/minion/PinotTaskManagerStatelessTest.java
@@ -152,8 +152,8 @@ public class PinotTaskManagerStatelessTest extends 
ControllerTest {
             "MergeRollupTask", ImmutableMap.of("schedule", "0 */10 * ? * * 
*"))));
     updateTableConfig(tableConfig);
     waitForJobGroupNames(_controllerStarter.getTaskManager(),
-        jgn -> jgn.size() == 2 && 
jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn
-            .contains(MinionConstants.MergeRollupTask.TASK_TYPE),
+        jgn -> jgn.size() == 2 && 
jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && 
jgn.contains(
+            MinionConstants.MergeRollupTask.TASK_TYPE),
         "JobGroupNames should have SegmentGenerationAndPushTask and 
MergeRollupTask");
     validateJob(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE, "0 
*/30 * ? * * *");
     validateJob(MinionConstants.MergeRollupTask.TASK_TYPE, "0 */10 * ? * * *");
@@ -203,8 +203,10 @@ public class PinotTaskManagerStatelessTest extends 
ControllerTest {
         "JobGroupNames should have SegmentGenerationAndPushTask only");
     validateJob(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE, "0 
*/10 * ? * * *");
 
-    // Restart controller
+    // Restart controller. We need to set the port after stopping the 
controller because we are reusing the same config.
+    int controllerPort = _controllerPort;
     stopController();
+    _controllerPort = controllerPort;
     startController(properties);
     // wait for controller to start correctly.
     TestUtils.waitForCondition((aVoid) -> {
@@ -232,8 +234,8 @@ public class PinotTaskManagerStatelessTest extends 
ControllerTest {
     // The new MergeRollup task wouldn't be scheduled if not eagerly checking 
table configs
     // after setting up subscriber on ChildChanges zk event when controller 
gets restarted.
     waitForJobGroupNames(_controllerStarter.getTaskManager(),
-        jgn -> jgn.size() == 2 && 
jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && jgn
-            .contains(MinionConstants.MergeRollupTask.TASK_TYPE),
+        jgn -> jgn.size() == 2 && 
jgn.contains(MinionConstants.SegmentGenerationAndPushTask.TASK_TYPE) && 
jgn.contains(
+            MinionConstants.MergeRollupTask.TASK_TYPE),
         "JobGroupNames should have SegmentGenerationAndPushTask and 
MergeRollupTask");
 
     dropOfflineTable(RAW_TABLE_NAME);
diff --git 
a/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java
 
b/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java
index 0a127071da..33bd9b008e 100644
--- 
a/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java
+++ 
b/pinot-integration-test-base/src/test/java/org/apache/pinot/integration/tests/BaseClusterIntegrationTest.java
@@ -270,7 +270,7 @@ public abstract class BaseClusterIntegrationTest extends 
ClusterTest {
 
   protected Schema createSchema(String schemaFileName)
       throws IOException {
-    InputStream inputStream = 
BaseClusterIntegrationTest.class.getClassLoader().getResourceAsStream(schemaFileName);
+    InputStream inputStream = 
getClass().getClassLoader().getResourceAsStream(schemaFileName);
     Assert.assertNotNull(inputStream);
     return Schema.fromInputStream(inputStream);
   }
@@ -291,13 +291,13 @@ public abstract class BaseClusterIntegrationTest extends 
ClusterTest {
    * Creates a new OFFLINE table config.
    */
   protected TableConfig createOfflineTableConfig() {
-    return new 
TableConfigBuilder(TableType.OFFLINE).setTableName(getTableName())
-        
.setTimeColumnName(getTimeColumnName()).setSortedColumn(getSortedColumn())
-        
.setInvertedIndexColumns(getInvertedIndexColumns()).setNoDictionaryColumns(getNoDictionaryColumns())
-        
.setRangeIndexColumns(getRangeIndexColumns()).setBloomFilterColumns(getBloomFilterColumns())
-        
.setFieldConfigList(getFieldConfigs()).setNumReplicas(getNumReplicas()).setSegmentVersion(getSegmentVersion())
-        
.setLoadMode(getLoadMode()).setTaskConfig(getTaskConfig()).setBrokerTenant(getBrokerTenant())
-        
.setServerTenant(getServerTenant()).setIngestionConfig(getIngestionConfig()).setQueryConfig(getQueryConfig())
+    return new 
TableConfigBuilder(TableType.OFFLINE).setTableName(getTableName()).setTimeColumnName(getTimeColumnName())
+        
.setSortedColumn(getSortedColumn()).setInvertedIndexColumns(getInvertedIndexColumns())
+        
.setNoDictionaryColumns(getNoDictionaryColumns()).setRangeIndexColumns(getRangeIndexColumns())
+        
.setBloomFilterColumns(getBloomFilterColumns()).setFieldConfigList(getFieldConfigs())
+        
.setNumReplicas(getNumReplicas()).setSegmentVersion(getSegmentVersion()).setLoadMode(getLoadMode())
+        
.setTaskConfig(getTaskConfig()).setBrokerTenant(getBrokerTenant()).setServerTenant(getServerTenant())
+        
.setIngestionConfig(getIngestionConfig()).setQueryConfig(getQueryConfig())
         
.setNullHandlingEnabled(getNullHandlingEnabled()).setSegmentPartitionConfig(getSegmentPartitionConfig())
         .build();
   }
@@ -399,9 +399,8 @@ public abstract class BaseClusterIntegrationTest extends 
ClusterTest {
   /**
    * Creates a new Upsert enabled table config.
    */
-  protected TableConfig createCSVUpsertTableConfig(String tableName, @Nullable 
String kafkaTopicName,
-      int numPartitions, Map<String, String> streamDecoderProperties, 
UpsertConfig upsertConfig,
-      String primaryKeyColumn) {
+  protected TableConfig createCSVUpsertTableConfig(String tableName, @Nullable 
String kafkaTopicName, int numPartitions,
+      Map<String, String> streamDecoderProperties, UpsertConfig upsertConfig, 
String primaryKeyColumn) {
     Map<String, ColumnPartitionConfig> columnPartitionConfigMap = new 
HashMap<>();
     columnPartitionConfigMap.put(primaryKeyColumn, new 
ColumnPartitionConfig("Murmur", numPartitions));
 
@@ -469,18 +468,14 @@ public abstract class BaseClusterIntegrationTest extends 
ClusterTest {
         Properties properties = getPinotConnectionProperties();
         properties.put("useMultistageEngine", "true");
         _pinotConnectionV2 = ConnectionFactory.fromZookeeper(getZkUrl() + "/" 
+ getHelixClusterName(),
-            new JsonAsyncHttpPinotClientTransportFactory()
-                .withConnectionProperties(properties)
-                .buildTransport());
+            new 
JsonAsyncHttpPinotClientTransportFactory().withConnectionProperties(properties).buildTransport());
       }
       return _pinotConnectionV2;
     }
     if (_pinotConnection == null) {
-      _pinotConnection =
-          ConnectionFactory.fromZookeeper(getZkUrl() + "/" + 
getHelixClusterName(),
-              new JsonAsyncHttpPinotClientTransportFactory()
-                  .withConnectionProperties(getPinotConnectionProperties())
-                  .buildTransport());
+      _pinotConnection = ConnectionFactory.fromZookeeper(getZkUrl() + "/" + 
getHelixClusterName(),
+          new 
JsonAsyncHttpPinotClientTransportFactory().withConnectionProperties(getPinotConnectionProperties())
+              .buildTransport());
     }
     return _pinotConnection;
   }
@@ -554,7 +549,7 @@ public abstract class BaseClusterIntegrationTest extends 
ClusterTest {
    */
   protected List<File> unpackTarData(String tarFileName, File outputDir)
       throws Exception {
-    InputStream inputStream = 
BaseClusterIntegrationTest.class.getClassLoader().getResourceAsStream(tarFileName);
+    InputStream inputStream = 
getClass().getClassLoader().getResourceAsStream(tarFileName);
     Assert.assertNotNull(inputStream);
     return TarGzCompressionUtils.untar(inputStream, outputDir);
   }


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@pinot.apache.org
For additional commands, e-mail: commits-h...@pinot.apache.org

Reply via email to