Re: parallel dump fails to dump large tables
От | Shanker Singh |
---|---|
Тема | Re: parallel dump fails to dump large tables |
Дата | |
Msg-id | 961471F4049EF94EAD4D0165318BD8816259064A@Corp-MBXE3.iii.com обсуждение исходный текст |
Ответ на | Re: parallel dump fails to dump large tables (Tom Lane <tgl@sss.pgh.pa.us>) |
Ответы |
Re: parallel dump fails to dump large tables
|
Список | pgsql-general |
I tried turning off ssl renegotiation by setting "ssl_renegotiation_limit = 0" in postgresql.conf but it had no effect. Theparallel dump still fails on large tables consistently. Thanks Shanker -----Original Message----- From: Tom Lane [mailto:tgl@sss.pgh.pa.us] Sent: Saturday, February 14, 2015 9:00 AM To: rod@iol.ie Cc: Shanker Singh; pgsql-general@postgresql.org Subject: Re: [GENERAL] parallel dump fails to dump large tables "Raymond O'Donnell" <rod@iol.ie> writes: > On 14/02/2015 15:42, Shanker Singh wrote: >> Hi, >> I am having problem using parallel pg_dump feature in postgres >> release 9.4. The size of the table is large(54GB). The dump fails >> with the >> error: "pg_dump: [parallel archiver] a worker process died >> unexpectedly". After this error the pg_dump aborts. The error log >> file gets the following message: >> >> 2015-02-09 15:22:04 PST [8636]: [2-1] >> user=pdroot,db=iii,appname=pg_dump >> STATEMENT: COPY iiirecord.varfield (id, field_type_tag, marc_tag, >> marc_ind1, marc_ind2, field_content, field_group_id, occ_num, >> record_id) TO stdout; >> 2015-02-09 15:22:04 PST [8636]: [3-1] >> user=pdroot,db=iii,appname=pg_dump >> FATAL: connection to client lost > There's your problem - something went wrong with the network. I'm wondering about SSL renegotiation failures as a possible cause of the disconnect --- that would explain why it only happenson large tables. regards, tom lane
В списке pgsql-general по дате отправления: