Re: pg_dump's over 2GB
От | Jeff Hoffmann |
---|---|
Тема | Re: pg_dump's over 2GB |
Дата | |
Msg-id | 39D4C64F.378F09BB@propertykey.com обсуждение исходный текст |
Ответ на | pg_dump's over 2GB ("Bryan White" <bryan@arcamax.com>) |
Ответы |
Re: pg_dump's over 2GB
|
Список | pgsql-general |
Bryan White wrote: > > I am thinking that > instead I will need to pipe pg_dumps output into gzip thus avoiding the > creation of a file of that size. > sure, i do it all the time. unfortunately, i've had it happen a few times where even gzipping a database dump goes over 2GB, which is a real PITA since i have to dump some tables individually. generally, i do something like pg_dump database | gzip > database.pgz to dump the database and gzip -dc database.pgz | psql database to restore it. i've always thought that compress should be an option for pg_dump, but it's really not that much more work to just pipe the input and output through gzip. -- Jeff Hoffmann PropertyKey.com
В списке pgsql-general по дате отправления: