This is an automated email from the ASF dual-hosted git repository. xxyu pushed a commit to branch kylin5 in repository https://gitbox.apache.org/repos/asf/kylin.git
The following commit(s) were added to refs/heads/kylin5 by this push: new cfe3b8e06b KYLIN-5217, fix ut in kylin-engine-spark module cfe3b8e06b is described below commit cfe3b8e06b4d3b65ed1436cb24cb21b130023bae Author: Mukvin <boyboys...@163.com> AuthorDate: Wed Aug 3 16:31:02 2022 +0800 KYLIN-5217, fix ut in kylin-engine-spark module --- .../spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java | 4 ++-- .../org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java b/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java index 16da811964..0e3ce0e763 100644 --- a/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java +++ b/src/spark-project/engine-spark/src/test/java/org/apache/kylin/engine/spark/source/NSparkCubingSourceInputBySparkDataSourceTest.java @@ -60,7 +60,7 @@ public class NSparkCubingSourceInputBySparkDataSourceTest extends NLocalWithSpar KylinBuildEnv.clean(); KylinBuildEnv kylinBuildEnv = KylinBuildEnv.getOrCreate(getTestConfig()); getTestConfig().setProperty("kylin.source.provider.9", - "NSparkDataSource"); + "org.apache.kylin.engine.spark.source.NSparkDataSource"); getTestConfig().setProperty("kylin.build.resource.read-transactional-table-enabled", "true"); NTableMetadataManager tableMgr = NTableMetadataManager.getInstance(getTestConfig(), "ssb"); TableDesc fact = tableMgr.getTableDesc("SSB.P_LINEORDER"); @@ -80,7 +80,7 @@ public class NSparkCubingSourceInputBySparkDataSourceTest extends NLocalWithSpar KylinBuildEnv.clean(); KylinBuildEnv kylinBuildEnv = KylinBuildEnv.getOrCreate(getTestConfig()); getTestConfig().setProperty("kylin.source.provider.9", - "NSparkDataSource"); + "org.apache.kylin.engine.spark.source.NSparkDataSource"); getTestConfig().setProperty("kylin.build.resource.read-transactional-table-enabled", "true"); NTableMetadataManager tableMgr = NTableMetadataManager.getInstance(getTestConfig(), "ssb"); TableDesc fact = tableMgr.getTableDesc("SSB.P_LINEORDER"); diff --git a/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala b/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala index 41706bdefd..84ed8635a2 100644 --- a/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala +++ b/src/spark-project/engine-spark/src/test/scala/org/apache/spark/sql/newSession/WithKylinExternalCatalog.scala @@ -64,7 +64,7 @@ trait WithKylinExternalCatalog extends SparkFunSuite with BeforeAndAfterAll { metaStore.createTestMetadata(metadata: _*) metaStore.overwriteSystemProp("kylin.use.external.catalog", externalCatalog) metaStore.overwriteSystemProp("kylin.NSparkDataSource.data.dir", s"${kylinConf.getMetadataUrlPrefix}/../data") - metaStore.overwriteSystemProp("kylin.source.provider.9", "NSparkDataSource") + metaStore.overwriteSystemProp("kylin.source.provider.9", "org.apache.kylin.engine.spark.source.NSparkDataSource") metaStore.overwriteSystemProp("kylin.query.engine.sparder-additional-files", fitPathForUT(additional)) metaStore.overwriteSystemProp("kylin.source.jdbc.adaptor", "Set By WithKylinExternalCatalog") metaStore.overwriteSystemProp("kylin.source.jdbc.driver", "Set By WithKylinExternalCatalog")