Re: pg_dump's over 2GB
| От | Jeff Hoffmann | 
|---|---|
| Тема | Re: pg_dump's over 2GB | 
| Дата | |
| Msg-id | 39D4C64F.378F09BB@propertykey.com обсуждение исходный текст | 
| Ответ на | pg_dump's over 2GB ("Bryan White" <bryan@arcamax.com>) | 
| Ответы | Re: pg_dump's over 2GB | 
| Список | pgsql-general | 
Bryan White wrote:
>
> I am thinking that
> instead I will need to pipe pg_dumps output into gzip thus avoiding the
> creation of a file of that size.
>
sure, i do it all the time.  unfortunately, i've had it happen a few
times where even gzipping a database dump goes over 2GB, which is a real
PITA since i have to dump some tables individually.  generally, i do
something like
    pg_dump database | gzip > database.pgz
to dump the database and
    gzip -dc database.pgz | psql database
to restore it.  i've always thought that compress should be an option
for pg_dump, but it's really not that much more work to just pipe the
input and output through gzip.
--
Jeff Hoffmann
PropertyKey.com
		
	В списке pgsql-general по дате отправления: