alvarowolfx opened a new issue, #425: URL: https://github.com/apache/arrow-go/issues/425
### Describe the enhancement requested Right now the `ipc.NewWriter` writes the schema data to the output buffer, which is not ideal for some scenarios where the target expect only RecordBatch messages. The kind of equivalent to that in the Python land with pyarrow is that `pyarrow.Table` has a `to_batches` method, that outputs only RecordBatch messages. We found this issue by trying to use `arrow-go` with the BigQuery Storage Write API, which now supports data in Arrow format. Using an `ipc.Writer` makes BigQuery reject the output, as the first message is one with the schema, not a `RecordBatch`. Draft PR on #421 References * Arrow support on BigQuery Storage Write API - https://cloud.google.com/bigquery/docs/supported-data-types#supported-apache-arrow-data-types * PyArrow and BigQuery Storage Write API example - https://cloud.google.com/bigquery/docs/write-api-streaming#arrow-format * Issue with BigQuery Storage Write API and `arrow-go` - https://github.com/googleapis/google-cloud-go/issues/12478 ### Component(s) Other -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@arrow.apache.org.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org