Re: Overhauling GUCS

From: "Hakan Kocaman" <hkocam(at)googlemail(dot)com>
To: "Gregory Stark" <stark(at)enterprisedb(dot)com>
Cc: "Josh Berkus" <josh(at)agliodbs(dot)com>, "Joshua D(dot) Drake" <jd(at)commandprompt(dot)com>, "Tom Lane" <tgl(at)sss(dot)pgh(dot)pa(dot)us>, "Robert Treat" <xzilla(at)users(dot)sourceforge(dot)net>, pgsql-hackers(at)postgresql(dot)org, "Peter Eisentraut" <peter_e(at)gmx(dot)net>, "Greg Smith" <gsmith(at)gregsmith(dot)com>, "Andreas Pflug" <pgadmin(at)pse-consulting(dot)de>, Decibel! <decibel(at)decibel(dot)org>
Subject: Re: Overhauling GUCS
Date: 2008-06-09 21:17:08
Message-ID: 48ca23600806091417u546f3b6dm6314bc55aaa147f6@mail.gmail.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-hackers

On 6/9/08, Gregory Stark <stark(at)enterprisedb(dot)com> wrote:
>
> "Josh Berkus" <josh(at)agliodbs(dot)com> writes:
>
> > Where analyze does systematically fall down is with databases over 500GB
> in
> > size, but that's not a function of d_s_t but rather of our tiny sample
> size.
>
>
> n_distinct. For that Josh is right, we *would* need a sample size
> proportional
> to the whole data set which would practically require us to scan the whole
> table (and have a technique for summarizing the results in a nearly
> constant
> sized data structure).
>
>
Hi,
is this (summarizing results in a constant sized data structure) something
which could be achived by Bloom-Filters ?
http://archives.postgresql.org/pgsql-general/2008-06/msg00076.php

Kind regards
Hakan Kocaman

In response to

Responses

Browse pgsql-hackers by date

  From Date Subject
Next Message Gregory Stark 2008-06-09 23:01:45 Re: Overhauling GUCS
Previous Message Gregory Stark 2008-06-09 20:52:20 Re: pg_dump restore time and Foreign Keys