Large Text Search Help
От | psql-mail@freeuk.com |
---|---|
Тема | Large Text Search Help |
Дата | |
Msg-id | E1A7GXx-0007J5-PM@buckaroo.freeuk.net обсуждение исходный текст |
Ответы |
Re: Large Text Search Help
|
Список | pgsql-performance |
Hi, I am trying to design a large text search database. It will have upwards of 6 million documents, along with meta data on each. I am currently looking at tsearch2 to provide fast text searching and also playing around with different hardware configurations. 1. With tsearch2 I get very good query times up until I insert more records. For example with 100,000 records tsearch2 returns in around 6 seconds, with 200,000 records tsearch2 returns in just under a minute. Is this due to the indices fitting entirely in memory with 100,000 records? 2. As well as whole word matching i also need to be able to do substring matching. Is the FTI module the way to approach this? 3. I have just begun to look into distibuted queries. Is there an existing solution for distibuting a postgresql database amongst multiple servers, so each has the same schema but only a subset of the total data? Any other helpful comments or sugestions on how to improve query times using different hardware or software techniques would be appreciated. Thanks, Mat
В списке pgsql-performance по дате отправления: