Re: An idea for parallelizing COPY within one backend
От | Dimitri Fontaine |
---|---|
Тема | Re: An idea for parallelizing COPY within one backend |
Дата | |
Msg-id | 200802270909.15336.dfontaine@hi-media.com обсуждение исходный текст |
Ответ на | An idea for parallelizing COPY within one backend ("Florian G. Pflug" <fgp@phlo.org>) |
Ответы |
Re: An idea for parallelizing COPY within one backend
Re: An idea for parallelizing COPY within one backend |
Список | pgsql-hackers |
Hi, Le mercredi 27 février 2008, Florian G. Pflug a écrit : > Upon reception of a COPY INTO command, a backend would > .) Fork off a "dealer" and N "worker" processes that take over the > client connection. The "dealer" distributes lines received from the > client to the N workes, while the original backend receives them > as tuples back from the workers. This looks so much like what pgloader does now (version 2.3.0~dev2, release candidate) at the client side, when configured for it, that I can't help answering the mail :)http://pgloader.projects.postgresql.org/dev/pgloader.1.html#_parallel_loading section_threads = N split_file_reading= False Of course, the backends still have to parse the input given by pgloader, which only pre-processes data. I'm not sure having the client prepare the data some more (binary format or whatever) is a wise idea, as you mentionned and wrt Tom's follow-up. But maybe I'm all wrong, so I'm all ears! Regards, -- dim
В списке pgsql-hackers по дате отправления: