Re: Reading data in bulk - help?

From: Josh Berkus <josh(at)agliodbs(dot)com>
To: Chris Huston <chuston(at)bangjafwac(dot)com>
Cc: pgsql-performance(at)postgresql(dot)org
Subject: Re: Reading data in bulk - help?
Date: 2003-09-10 17:16:14
Message-ID: 200309101016.14652.josh@agliodbs.com
Views: Raw Message | Whole Thread | Download mbox | Resend email
Thread:
Lists: pgsql-performance

Chris,

> The system is currently running on a single processor 500Mhz G4. We're
> likely to move to a two processor 2Ghz G5 in the next few months. Then
> each block may take only a 30-60 milliseconds to complete and their can
> be two concurrent blocks processing at once.

What about explaining your disk setup? Or mentioning postgresql.conf? For
somebody who wants help, you're ignoring a lot of advice and questions.

Personally, I'm not going to be of any further help until you report back on
the other 3 of 4 options.

> RELATED QUESTION: How now do I speed up the following query: "select
> distinct group_id from datatable"? Which results in a sequential scan
> of the db. Why doesn't it use the group_id index? I only do this once
> per run so it's not as critical as the fetch speed which is done 6817
> times.

Because it can't until PostgreSQL 7.4, which has hash aggregates. Up to
7.3, we have to use seq scans for all group bys. I'd suggest that you keep a
table of group_ids, instead.

--
Josh Berkus
Aglio Database Solutions
San Francisco

In response to

Browse pgsql-performance by date

  From Date Subject
Next Message aturner 2003-09-10 17:53:40 Upgrade Woes
Previous Message Bruce Momjian 2003-09-10 16:06:52 Re: Hardware recommendations to scale to silly load