On Tue, 15 Oct 2024 at 06:59, Wong, Kam Fook (TR Technology)
wrote:
> I am trying to copy a table (Postgres) that is close to 1 billion rows into a
> Partition table (Postgres) within the same DB. What is the fastest way to
> copy the data? This table has 37 columns where some of which are te
Hi Wong
On one occasion I had to upload 600 million records, and the most viable
and safest option was to generate plans and upload them through a massively
parallelized process (because for each process we audited that everything
was correct)
Atte.
JRBM
El lun, 14 oct 2024 a las 14:59, Wong, Ka
Hi,
There are many methods to achieve this and one of them is pg_bulkload
utility as described in previous email but I always preferred using python
multiprocessing which I think is more efficient. Below is the code which
you can modify as per your requirement:
import multiprocessing
import psycop
On Mon, 14 Oct, 2024, 23:29 Wong, Kam Fook (TR Technology), <
kamfook.w...@thomsonreuters.com> wrote:
> I am trying to copy a table (Postgres) that is close to 1 billion rows
> into a Partition table (Postgres) within the same DB. What is the fastest
> way to copy the data? This table has 37 co