jurossiar commented on issue #9960: URL: https://github.com/apache/iceberg/issues/9960#issuecomment-2507769859
@Syed-bug The issue is with spark-version: 3.5. I though you had the issue and wanted to be able to update tables. :) Refers to my comments: - https://github.com/apache/iceberg/issues/9960#issuecomment-2196827806 - https://github.com/apache/iceberg/issues/9960#issuecomment-2197375635 (We are using a REST catalog) I've just tested it again running something simple like [this](https://github.com/apache/iceberg/issues/9960#issuecomment-2200042116) ``` :: modules in use: org.apache.iceberg#iceberg-spark-runtime-3.5_2.12;1.6.1 from local-m2-cache in [default] org.reactivestreams#reactive-streams;1.0.3 from local-m2-cache in [default] org.slf4j#slf4j-api;1.7.30 from central in [default] software.amazon.awssdk#annotations;2.20.131 from central in [default] software.amazon.awssdk#bundle;2.20.131 from central in [default] software.amazon.awssdk#http-client-spi;2.20.131 from central in [default] software.amazon.awssdk#metrics-spi;2.20.131 from central in [default] software.amazon.awssdk#url-connection-client;2.20.131 from central in [default] software.amazon.awssdk#utils;2.20.131 from central in [default] software.amazon.eventstream#eventstream;1.0.1 from local-m2-cache in [default] ``` ``` Py4JJavaError: An error occurred while calling o52.sql. : org.apache.spark.SparkUnsupportedOperationException: UPDATE TABLE is not supported temporarily. at org.apache.spark.sql.errors.QueryExecutionErrors$.ddlUnsupportedTemporarilyError(QueryExecutionErrors.scala:1109) at org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:896) at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$1(QueryPlanner.scala:63) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:491) at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) at org.apache.spark.sql.execution.SparkStrategies.plan(SparkStrategies.scala:70) at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$3(QueryPlanner.scala:78) at scala.collection.TraversableOnce$folder$1.apply(TraversableOnce.scala:196) at scala.collection.TraversableOnce$folder$1.apply(TraversableOnce.scala:194) at scala.collection.Iterator.foreach(Iterator.scala:943) at scala.collection.Iterator.foreach$(Iterator.scala:943) at scala.collection.AbstractIterator.foreach(Iterator.scala:1431) at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:199) at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:192) at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1431) at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$2(QueryPlanner.scala:75) at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:486) at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:492) at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) at org.apache.spark.sql.execution.SparkStrategies.plan(SparkStrategies.scala:70) at org.apache.spark.sql.execution.QueryExecution$.createSparkPlan(QueryExecution.scala:476) at org.apache.spark.sql.execution.QueryExecution.$anonfun$sparkPlan$1(QueryExecution.scala:162) at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111) at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:202) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:526) at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:202) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827) at org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:201) at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:162) at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:155) at org.apache.spark.sql.execution.QueryExecution.$anonfun$executedPlan$1(QueryExecution.scala:175) at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111) at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:202) at org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:526) at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:202) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827) .... ``` And the issue is still there. But using `iceberg-spark-runtime-3.4_2.12;1.6.1` it works. Hope this help to reproduce it! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org