This is an automated email from the ASF dual-hosted git repository. xxyu pushed a commit to branch kylin-on-parquet-v2 in repository https://gitbox.apache.org/repos/asf/kylin.git
The following commit(s) were added to refs/heads/kylin-on-parquet-v2 by this push: new 270b332 KYLIN-4811 Support cube level configuration for BuildingJob 270b332 is described below commit 270b3326dcb0a6f0f7342323cb214bb1a9ece032 Author: XiaoxiangYu <x...@apache.org> AuthorDate: Sun Nov 8 21:12:46 2020 +0800 KYLIN-4811 Support cube level configuration for BuildingJob --- .../apache/kylin/engine/spark/job/NSparkExecutable.java | 14 +++++++++++--- 1 file changed, 11 insertions(+), 3 deletions(-) diff --git a/kylin-spark-project/kylin-spark-engine/src/main/java/org/apache/kylin/engine/spark/job/NSparkExecutable.java b/kylin-spark-project/kylin-spark-engine/src/main/java/org/apache/kylin/engine/spark/job/NSparkExecutable.java index dd306d4..98f63a1 100644 --- a/kylin-spark-project/kylin-spark-engine/src/main/java/org/apache/kylin/engine/spark/job/NSparkExecutable.java +++ b/kylin-spark-project/kylin-spark-engine/src/main/java/org/apache/kylin/engine/spark/job/NSparkExecutable.java @@ -36,6 +36,8 @@ import java.util.Map.Entry; import java.util.Set; +import org.apache.kylin.cube.CubeInstance; +import org.apache.kylin.cube.CubeManager; import org.apache.kylin.engine.spark.utils.MetaDumpUtil; import org.apache.commons.collections.CollectionUtils; import org.apache.commons.io.FileUtils; @@ -103,7 +105,10 @@ public class NSparkExecutable extends AbstractExecutable { @Override protected ExecuteResult doWork(ExecutableContext context) throws ExecuteException { //context.setLogPath(getSparkDriverLogHdfsPath(context.getConfig())); - final KylinConfig config = wrapConfig(context); + CubeManager cubeMgr = CubeManager.getInstance(KylinConfig.getInstanceFromEnv()); + CubeInstance cube = cubeMgr.getCube(this.getCubeName()); + KylinConfig config = cube.getConfig(); + config = wrapConfig(config); String sparkHome = KylinConfig.getSparkHome(); if (StringUtils.isEmpty(sparkHome) && !config.isUTEnv()) { @@ -192,9 +197,12 @@ public class NSparkExecutable extends AbstractExecutable { return String.format("%s.%s.log", config.getJobTmpOutputStorePath(getProject(), getId()), System.currentTimeMillis()); }*/ - + protected KylinConfig wrapConfig(ExecutableContext context) { - KylinConfig originalConfig = context.getConfig(); + return wrapConfig(context.getConfig()); + } + + protected KylinConfig wrapConfig(KylinConfig originalConfig) { String project = getParam(MetadataConstants.P_PROJECT_NAME); Preconditions.checkState(StringUtils.isNotBlank(project), "job " + getId() + " project info is empty");