On 3/10/19 5:53 PM, Julie Nishimura wrote:
Hello friends, I will need to migrate 500+ tables  from one server (8.3) to another (9.3). I cannot dump and load the entire database due to storage limitations (because the source is > 20 TB, and the target is about 1.5 TB).

I was thinking about using pg_dump with customized -t flag, then use restore. The table names will be in the list, or I could dump their names in a table.  What would be your suggestions on how to do it more efficiently?

The sizes you mention above, are they for the uncompressed raw data?

Are the tables all in one schema or multiple?

Where I am going with this is pg_dump -Fc --schema.

See:
https://www.postgresql.org/docs/10/app-pgrestore.html

The pg_restore -l to get a TOC(Table of Contents).

Comment out the items you do not want in the TOC.

Then pg_restore  --use-list.

See:

https://www.postgresql.org/docs/10/app-pgrestore.html


Thank you for your ideas, this is great to have you around, guys!




--
Adrian Klaver
adrian.kla...@aklaver.com

Reply via email to