AwasthiSomesh commented on issue #11297:
URL: https://github.com/apache/iceberg/issues/11297#issuecomment-2405237112

   @nastra 
   
   val spark = SparkSession
         .builder()
         .master("local[*]")
         .appName("Test")
         //.config("spark.sql.defaultCatalog", "spark_catalog")
         
.config("fs.azure.account.auth.type.someshadlsgen2.dfs.core.windows.net", 
"OAuth")
         
.config("fs.azure.account.oauth.provider.type.someshadlsgen2.dfs.core.windows.net",
 "org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider") 
         
.config("fs.azure.account.oauth2.client.id.someshadlsgen2.dfs.core.windows.net",
 "XXXXXXX") 
         
.config("fs.azure.account.oauth2.client.secret.someshadlsgen2.dfs.core.windows.net",
 "XXXXXX")  
         
.config("fs.azure.account.oauth2.client.endpoint.someshadlsgen2.dfs.core.windows.net",
 "https://login.microsoftonline.com/XXXXXXXXXXXXXXXXx/oauth2/token";)
         .config("spark.sql.catalog.spark_catalog", 
"org.apache.iceberg.spark.SparkCatalog")
         .config("spark.sql.catalog.spark_catalog.uri", 
"thrift://hostname:9083") 
         .config("spark.sql.catalog.spark_catalog.warehouse.dir", 
"/opt/hive/data/warehouse/") 
         .config("spark.sql.extensions", 
"org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions")  
         .config("spark.sql.catalog.spark_catalog", 
"org.apache.iceberg.spark.SparkSessionCatalog")
         .config("spark.sql.spark_catalog.io-impl", 
"org.apache.iceberg.azure.adlsv2.ADLSFileIO")
         
.config("spark.sql.hive.hiveserver2.jdbc.url","jdbc:hive2://hostname:10000/default")
         .config("spark.sql.spark_catalog.include-credentials", "true")
         .config("spark.sql.catalog.spark_catalog.type", "hive")
          .config("iceberg.engine.hive.enabled",true).enableHiveSupport()
         .getOrCreate();
   
   Error:
   
   24/10/02 23:29:11 INFO HiveMetaStoreClient: HMS client filtering is enabled.
   24/10/02 23:29:11 INFO HiveMetaStoreClient: Resolved metastore uris: 
[thrift://invsh005vm0101.informatica.com:9083]
   24/10/02 23:29:11 INFO HiveMetaStoreClient: Trying to connect to metastore 
with URI (thrift://invsh005vm0101.informatica.com:9083) in binary transport mode
   24/10/02 23:29:11 INFO HiveMetaStoreClient: Opened a connection to 
metastore, URI (thrift://invsh005vm0101.informatica.com:9083) current 
connections: 1
   24/10/02 23:29:11 INFO RetryingMetaStoreClient: RetryingMetaStoreClient 
proxy=class org.apache.hadoop.hive.metastore.HiveMetaStoreClient ugi=soawasth 
(auth:SIMPLE) retries=1 delay=1 lifetime=0
   Exception in thread "main" org.apache.spark.sql.AnalysisException: 
[TABLE_OR_VIEW_NOT_FOUND] The table or view `iceberg1table2` cannot be found. 
Verify the spelling and correctness of the schema and catalog.
   If you did not qualify the name with a schema, verify the current_schema() 
output, or qualify the name with the correct schema and catalog.
   To tolerate the error on drop use DROP VIEW IF EXISTS or DROP TABLE IF 
EXISTS.; line 1 pos 14;
   'Project [*]
   +- 'UnresolvedRelation [iceberg1table2], [], false
    
        at 
org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.tableNotFound(package.scala:87)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis0$2(CheckAnalysis.scala:202)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis0$2$adapted(CheckAnalysis.scala:182)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:244)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$foreachUp$1(TreeNode.scala:243)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$foreachUp$1$adapted(TreeNode.scala:243)
        at scala.collection.Iterator.foreach(Iterator.scala:929)
        at scala.collection.Iterator.foreach$(Iterator.scala:929)
        at scala.collection.AbstractIterator.foreach(Iterator.scala:1417)
        at scala.collection.IterableLike.foreach(IterableLike.scala:71)
        at scala.collection.IterableLike.foreach$(IterableLike.scala:70)
        at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:243)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis0(CheckAnalysis.scala:182)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis0$(CheckAnalysis.scala:164)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis0(Analyzer.scala:188)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis(CheckAnalysis.scala:160)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis$(CheckAnalysis.scala:150)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:188)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.$anonfun$executeAndCheck$1(Analyzer.scala:211)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:330)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:208)
        at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:77)
        at 
org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:138)
        at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:219)
        at 
org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:546)
        at 
org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:219)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
        at 
org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:218)
        at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:77)
        at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74)
        at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66)
        at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:99)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
        at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97)
        at 
org.apache.spark.sql.SparkSession.$anonfun$sql$4(SparkSession.scala:691)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:900)
        at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:682)
        at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:713)
        at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:744)
        at 
Hive4StandAloneAzureAdlsgen2$.main(Hive4StandAloneAzureAdlsgen2.scala:36)
        at Hive4StandAloneAzureAdlsgen2.main(Hive4StandAloneAzureAdlsgen2.scala)
   24/10/02 23:29:12 INFO SparkContext: Invoking stop() from shutdown hook
   24/10/02 23:29:12 INFO SparkContext: SparkContext is stopping with exitCode 
0.
   24/10/02 23:29:12 INFO SparkUI: Stopped Spark web UI at 
http://invw16con46.informatica.com:4040/
   24/10/02 23:29:12 INFO MapOutputTrackerMasterEndpoint: 
MapOutputTrackerMasterEndpoint stopped!
   24/10/02 23:29:12 INFO MemoryStore: MemoryStore cleared
   24/10/02 23:29:12 INFO BlockManager: BlockManager stopped
   24/10/02 23:29:12 INFO BlockManagerMaster: BlockManagerMaster stopped
   24/10/02 23:29:12 INFO 
OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
OutputCommitCoordinator stopped!
   24/10/02 23:29:12 INFO SparkContext: Successfully stopped SparkContext
   24/10/02 23:29:12 INFO ShutdownHookManager: Shutdown hook called
   24/10/02 23:29:12 INFO ShutdownHookManager: Deleting directory 
C:\Users\soawasth\AppData\Local\Temp\spark-274
   
   
   Please let me know what we are missing
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to