This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch branch-4.0 in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-4.0 by this push: new 117452a866b8 [SPARK-50605][CONNECT][TESTS][FOLLOW-UP] Remove unused YarnConnectTest 117452a866b8 is described below commit 117452a866b8e40ccce103717465922de916f536 Author: Hyukjin Kwon <gurwls...@apache.org> AuthorDate: Wed Apr 30 17:46:41 2025 +0900 [SPARK-50605][CONNECT][TESTS][FOLLOW-UP] Remove unused YarnConnectTest ### What changes were proposed in this pull request? This PR is a followup of https://github.com/apache/spark/pull/49107 that removes the unused `YarnConnectTest`. ### Why are the changes needed? It was mistakenly added, and forgotten to be removed. ### Does this PR introduce _any_ user-facing change? No, test-only. ### How was this patch tested? CI in this PR should verify the change. ### Was this patch authored or co-authored using generative AI tooling? No Closes #50762 from HyukjinKwon/SPARK-50605-followup. Authored-by: Hyukjin Kwon <gurwls...@apache.org> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> (cherry picked from commit 1fe7f634240b2d3d62e2ddb326eabef58aa5f863) Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- .../spark/deploy/yarn/YarnClusterSuite.scala | 28 ---------------------- 1 file changed, 28 deletions(-) diff --git a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnClusterSuite.scala b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnClusterSuite.scala index 4408817b0426..0092a0494c2a 100644 --- a/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnClusterSuite.scala +++ b/resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/YarnClusterSuite.scala @@ -767,34 +767,6 @@ private object YarnClasspathTest extends Logging { } -private object YarnConnectTest extends Logging { - def main(args: Array[String]): Unit = { - val output = new java.io.PrintStream(new File(args(0))) - val clz = Utils.classForName("org.apache.spark.sql.SparkSession$") - val moduleField = clz.getDeclaredField("MODULE$") - val obj = moduleField.get(null) - var builder = clz.getMethod("builder").invoke(obj) - builder = builder.getClass().getMethod( - "config", classOf[String], classOf[String]).invoke(builder, SPARK_API_MODE.key, "connect") - builder = builder.getClass().getMethod("master", classOf[String]).invoke(builder, "yarn") - val session = builder.getClass().getMethod("getOrCreate").invoke(builder) - - try { - // Check if the current session is a Spark Connect session. - session.getClass().getDeclaredField("client") - val df = session.getClass().getMethod("range", classOf[Long]).invoke(session, 10) - assert(df.getClass().getMethod("count").invoke(df) == 10) - } catch { - case e: Throwable => - e.printStackTrace(new java.io.PrintStream(output)) - throw e - } finally { - session.getClass().getMethod("stop").invoke(session) - output.close() - } - } -} - private object YarnAddJarTest extends Logging { def main(args: Array[String]): Unit = { if (args.length != 3) { --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org