lelandroling opened a new issue, #1571:
URL: https://github.com/apache/iceberg-python/issues/1571

   ### Question
   
   Checking through the GitHub issues, I noticed very few examples and I did 
see the open requests for improved documentation. Understandably, I understand 
that I can use MERGE INTO using Pyspark. My specific example is attempting to 
avoid the large overhead of Pyspark, but if that's the solution... ok. But 
before I walk down that path, I'm trying to understand how the use case looks 
for the ```.overwrite()``` and ```overwrite_filter```.
   
   ```
   conditions = []
   for row in values:
   
   row_condition = And(*[EqualTo(k, v) for k, v in zip(newKeys, row)])
   conditions.append(row_condition)
   
   filter_condition = Or(*conditions)
   ```
   
   I'm using this code to build out the filter_condition, then assigning that 
to overwrite_filter. What I've noticed is that if I have 1000 records, I'm 
hitting a maximum recursion error. My assumption is that I'm not understanding 
how to structure the filter_condition. Or the process can't handle this right 
now and I should move to MERGE INTO and Pyspark.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to