Re: Patch: dumping tables data in multiple chunks in pg_dump
| От | Dilip Kumar |
|---|---|
| Тема | Re: Patch: dumping tables data in multiple chunks in pg_dump |
| Дата | |
| Msg-id | CAFiTN-tV4jWKN75E5YLB-jSqb8j0E1PctiDjztv=ccfbe3YPmg@mail.gmail.com обсуждение исходный текст |
| Ответ на | Patch: dumping tables data in multiple chunks in pg_dump (Hannu Krosing <hannuk@google.com>) |
| Список | pgsql-hackers |
On Tue, Nov 11, 2025 at 9:00 PM Hannu Krosing <hannuk@google.com> wrote: > > Attached is a patch that adds the ability to dump table data in multiple chunks. > > Looking for feedback at this point: > 1) what have I missed > 2) should I implement something to avoid single-page chunks > > The flag --huge-table-chunk-pages which tells the directory format > dump to dump tables where the main fork has more pages than this in > multiple chunks of given number of pages, > > The main use case is speeding up parallel dumps in case of one or a > small number of HUGE tables so parts of these can be dumped in > parallel. > +1 for the idea, I haven't done the detailed review but I was just going through the patch, I noticed that we use pg_class->relpages to identify whether to chunk the table or not, which should be fine but don't you think if we use direct size calculation function like pg_relation_size() we might get better idea and not dependent upon whether the stats are updated or not? This will make chunking behavior more deterministic. -- Regards, Dilip Kumar Google
В списке pgsql-hackers по дате отправления: