Search Postgresql Archives

Re: Query results caching?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> -----Original Message-----
> From: Ben-Nes Yonatan [mailto:da@xxxxxxxxxxxx]
> Sent: Monday, August 22, 2005 3:28 PM
> To: Jim C. Nasby; Sean Davis; Dann Corbit
> Cc: pgsql-general@xxxxxxxxxxxxxx
> Subject: Re:  Query results caching?
> 
> > On Mon, Aug 22, 2005 at 10:13:49PM +0200, Ben-Nes Yonatan wrote:
> >
> >>I think that I was misunderstood, Ill make an example:
> >>Lets say that im making the following query for the first time on
the
> >>"motorcycles" table which got an index on the "manufacturer" field:
> >>
> >>EXPLAIN ANALYZE SELECT manufacturer FROM motorcycles WHERE
> >>manufacturer='suzuki';
> >>... Total runtime: 3139.587 ms
> >>
> >>Now im doing the same query again and i get a much faster result
(cause
> >>of the "caching"): Total runtime: 332.53 ms
> >>
> >>After both of those queries I drop the index and query the table
again
> >>with the exact same query as before and now I receive: Total
runtime:
> >>216834.871 ms
> >>
> >>And for my last check I run the exact same query again (without
creating
> >>the INDEX back again) and I get quite similar result to my third
query:
> >>Total runtime: 209218.01 ms
> >>
> >>
> >>My problem is that (maybe I just dont understand something basic
> >>here...) the last 2 (also the second query but I dont care about
that)
> >>queries were using the "cache" that was created after the first
query
> >>(which had an INDEX) so none of them actually showed me what will
happen
> >>if a client will do such a search (without an INDEX) for the first
time.
> >>
> >>I want to delete that "caching" after I do the first 2 queries so my
> >>next queries will show me "real life results".
> >
> >
> 
> Ok I tried to handle both of your replies cause I got them at 2
seperate
> emails.
> 
> Dann Corbit wrote:
>  > These results are all what I would expect.  When you delete the
index,
>  > the query will be forced to do a table scan (to examine every
single
>  > record in the table one by one).  If the table is non-trivial it is
>  > unlikely that either the OS or the database will cache the whole
thing
>  > in memory.  However, when you query a small record set, then it is
>  > likely to be retained in RAM which is literally thousands of times
>  > faster than disk.
> 
> Didnt know that, good to know though doesnt assure me...
> What if I drop the INDEX but create a diffrent INDEX which also make
the
> process alot faster then without an INDEX but slower/faster then the
one
> before, will it wont use the former "caching"?

You can add several indexes to a single table.
If you do some statistics on the query patterns, you can find what
indexes are needed to make the queries as fast as possible.

>  > If a search is to be made on a frequent basis, you should create an
>  > index.
>  > The query results above show you why.
> 
> Obvious :)
> 
>  > Think about this for a minute.  The real life results you want are
>  > very fast results.  For that reason, you should try to model the
>  > customer queries as nearly as possible.  If you have a canned
>  > application like order entry, then the real parameterized query set
>  > will probably be quite small in real life.  If you are creating a
>  > server for ad-hoc queries then it will be far more difficult to
model
>  > in real life.
>  >
>  > What is the real purpose of the application that you are writing?
>  >
>  > Will users be using a pre-programmed front end, or will they be
typing
>  > in queries free-form for whatever their heart desires?
> 
> Ok ill try to describe the system as short & precise as possible (its
> also passed midnight here :)).
> Each day I receive about 4 million rows of data (products) which I
> insert into table1 (after I delete all of the previous data it had),
> along it I receive for every row about another 15 keywords which I
> insert into table2 (where as in table1 I delete all of the previous
data
> it had also), this process is a fact that I cant change.

If the data arrives on a daily basis, and is not updated until the next
day, I suggest creating a lot of indexes, and cluster on the index used
most frequently.

What exactly are the 15 keywords in the second table for?
Are they column names?
Are they categories for the first table?
Why is the second table necessary at all?

> Now the users of the site can search for data from table1 by typing
> whichever (and up to 4) words as they want at a text field (search
input
> string) and the server should display the correct results by querying
> table1 & join table2 for its keywords.

Can you give the exact table definitions for the two tables, and also
the most likely queries you are going to receive?

When the users type in keywords -- can these keywords be applied against
any column in the table or only against a single column or against a
small set of columns or something else?

> I succeded to do it quite fast but when I tried to ORDER BY my results
> its times jumped up drastically (2-3 seconds for a query... and thats
> after the caching..).

Order by will complicate quite a bit.  I have not tried it on
PostgreSQL, but if you know the result set is small, a technique is to
select into a temp table and then order by on the temp table.  It works
well on other database systems (caveat: it has been a while since I
worked as a DBA and I have not worked as a DBA on PostgreSQL).

> I can't allow a situation where a user will search with a keyword
which
> wasnt 'cached' before and because of that he will wait 15 seconds for
a
> result.

You might try throwing hardware at it.  A 4 CPU AMD 64 machine with
Ultra 320 striped SCSI disk array and a few gigabytes of ram will
perform admirably.

You might want more than is possible.  If you have 4 million rows, and
each row is 1K, then that is 4 GB.  If your users do a query that has
not been performed yet and you have to do a table scan, then you cannot
expect some kind of sub-second response times because it won't be
physically possible on any system.

If you know what most queries may look like or if you only have a few
character columns so that you can make an index on each of them and if
you can put a unique clustered index on the most important (frequently
used) item, then you can get the majority of your queries to run very
quickly, and only on rare occasions will the query be slow.

If I have a 4 million row table, with long rows and big varchar columns
and I run a query on a column like this:

	SELECT * FROM inventory WHERE product LIKE '%Table%'

It isn't going to be fast on any system with any database.
[snip]

---------------------------(end of broadcast)---------------------------
TIP 1: if posting/reading through Usenet, please send an appropriate
       subscribe-nomail command to majordomo@xxxxxxxxxxxxxx so that your
       message can get through to the mailing list cleanly


[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Postgresql Jobs]     [Postgresql Admin]     [Postgresql Performance]     [Linux Clusters]     [PHP Home]     [PHP on Windows]     [Kernel Newbies]     [PHP Classes]     [PHP Books]     [PHP Databases]     [Postgresql & PHP]     [Yosemite]
  Powered by Linux