Fast Backups With pg_dump
От | Wayne Piekarski |
---|---|
Тема | Fast Backups With pg_dump |
Дата | |
Msg-id | 199909110513.OAA15799@helpdesk.senet.com.au обсуждение исходный текст |
Список | pgsql-sql |
Hi, Currently, I am running the pg_dump program every hour to produce backups of the database, and then storing them in case something fails or I need recover something. However, the backups are quite large, and the CPU gets really thrashed a lot doing the COPY as it has to make lots of function calls (as mentioned in a previous email of mine to hackers) and I wanted to avoid this if possible. What would be really neat is if you could do something like, give me a dump of all rows changed since the last backup, sortof like a diff file, so I make a full backup at the start of the day, and then just store deltas for the rest of the day. I know that postgres can keep track of this kind of stuff because the MVCC allows it to make a consistent snapshot of the dbms even while updates are occuring, so is this kind of idea possible (or does it already exist?) This would mean that the disks and CPU don't have to spend all their time backing up data which never changes and just record down the little differences. Then you could go backups like once every minute or something, and even synchronise them against another database if you wanted to? I would like to hear what others have to say about this. thanks, Wayne ------------------------------------------------------------------------------ Wayne Piekarski Tel: (08) 8221 5221 Research & Development Manager Fax: (08) 8221 5220 SE Network Access Pty Ltd Mob: 0407 395 889 222 Grote Street Email: wayne@senet.com.au Adelaide SA 5000 WWW: http://www.senet.com.au
В списке pgsql-sql по дате отправления: