Re: Load distributed checkpoint
От | Greg Smith |
---|---|
Тема | Re: Load distributed checkpoint |
Дата | |
Msg-id | Pine.GSO.4.64.0612221612050.24353@westnet.com обсуждение исходный текст |
Ответ на | Re: Load distributed checkpoint ("Simon Riggs" <simon@2ndquadrant.com>) |
Список | pgsql-hackers |
On Fri, 22 Dec 2006, Simon Riggs wrote: > I have also seen cases where the WAL drive, even when separated, appears > to spike upwards during a checkpoint. My best current theory, so far > untested, is that the WAL and data drives are using the same CFQ > scheduler and that the scheduler actively slows down WAL requests when > it need not. Mounting the drives as separate block drives with separate > schedulers, CFQ for data and Deadline for WAL should help. The situation I've been seeing is that the database needs a new block to complete a query and issues a read request to get it, but that read is behind the big checkpoint fsync. Client sits there for quite some time waiting for the fsync to finish before it gets the data it needs, and now your trivial select took seconds to complete. It's fairly easy to replicate this problem using pgbench on Linux--I've seen a query sit there for 15 seconds when going out of my way to aggrevate the behavior. One of Takayuki's posts here mentioned a worst-case delay of 13 seconds, that's the problem rearing its ugly head. You may be right that what you're seeing would be solved with a more complicated tuning on a per-device basis (which, by the way, isn't available unless you're running a more recent Linux kernel than most many distributions have available). You can tune the schedulers all day and not make a lick of difference to what I've been running into; I know, I tried. -- * Greg Smith gsmith@gregsmith.com http://www.gregsmith.com Baltimore, MD
В списке pgsql-hackers по дате отправления: