pg_Restore

Started by bhanu udayaabout 13 years ago57 messagesgeneral
Jump to latest
#1bhanu udaya
udayabhanu1984@hotmail.com

Hello, Greetings !I am new to postgres and recently got migrated from oracle.I am trying to restore 9.5G database (1GB dumpfile) which has 500 schemas with 1 lakh rows in each schema. Could take the data dump using pg_dump and it takes around 40 minutes. I tried to use pg_restore to restore this dump, but it takes hours to restore the dump. I have used the configurations parameters as below: shared_buffers = 1024MB work_mem = 512MBmaintenance_work_mem = 512MBwal_buffers = 100MB I have used command to restore as below:pg_Restore -d newdb -j2 e:\file.dmp My machine configurations are as below:Windows Core I5 with 4GB Ram. Thanks for your help. Thanks and RegardsRadha Krishna

#2Guillaume Lelarge
guillaume@lelarge.info
In reply to: bhanu udaya (#1)
Re: pg_Restore

On Mon, 2013-01-21 at 00:03 +0530, bhanu udaya wrote:

Hello, Greetings !I am new to postgres and recently got migrated from oracle.I am trying to restore 9.5G database (1GB dumpfile) which has 500 schemas with 1 lakh rows in each schema. Could take the data dump using pg_dump and it takes around 40 minutes. I tried to use pg_restore to restore this dump, but it takes hours to restore the dump. I have used the configurations parameters as below: shared_buffers = 1024MB work_mem = 512MBmaintenance_work_mem = 512MBwal_buffers = 100MB I have used command to restore as below:pg_Restore -d newdb -j2 e:\file.dmp My machine configurations are as below:Windows Core I5 with 4GB Ram. Thanks for your help. Thanks and RegardsRadha Krishna

You should better look for help on the pgsql-* mailing lists (for
example pgsql-general).

Anyway, from what you say, you should probably take a look at your
checkpoint settings. They probably hurt you more than you think.

--
Guillaume
http://blog.guillaume.lelarge.info
http://www.dalibo.com

--
Sent via pgadmin-support mailing list (pgadmin-support@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgadmin-support

#3bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#1)
Re: pg_Restore

Hello,

Greetings !
I am new to postgres and recently got migrated from oracle. I am using postgresql 9.2 version.
I am trying to restore 9.5G database (1GB dumpfile) which has 500 schemas with 1 lakh rows in each schema. Could take the data dump using pg_dump and it takes around 40 minutes. I tried to use pg_restore to restore this dump, but it takes hours to restore the dump. I have used the configurations parameters as below:

shared_buffers = 1024MB
work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB

I have used command to restore as below:
pg_Restore -d newdb -j2 e:\file.dmp

My machine configurations are as below:
Windows Core I5 with 4GB Ram.

Thanks for your help.

Thanks and Regards
Radha Krishna

#4François Beausoleil
francois@teksol.info
In reply to: bhanu udaya (#3)
Re: pg_Restore

Le 2013-01-20 à 23:10, bhanu udaya a écrit :

I am new to postgres and recently got migrated from oracle. I am using postgresql 9.2 version.

Welcome, and good choice for the version.

I am trying to restore 9.5G database (1GB dumpfile) which has 500 schemas with 1 lakh rows in each schema. Could take the data dump using pg_dump and it takes around 40 minutes. I tried to use pg_restore to restore this dump, but it takes hours to restore the dump. I have used the configurations parameters as below:

shared_buffers = 1024MB
work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB

I have used command to restore as below:
pg_Restore -d newdb -j2 e:\file.dmp

My machine configurations are as below:
Windows Core I5 with 4GB Ram.

Other settings you can change during the initial restore / load phase:

fsync = off # ONLY DURING INITIAL DATA LOAD!

checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per file, check disk space)
checkpoint_timeout = 30min
checkpoint_completion_target = 0.9

wal_level = minimal # You'll need to do a full base backup if you use this

Read this section of the manual: http://www.postgresql.org/docs/current/static/runtime-config-wal.html

Have a great day!
François Beausoleil

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#5bhanu udaya
udayabhanu1984@hotmail.com
In reply to: François Beausoleil (#4)
Re: pg_Restore

Hello,Greetings !Thank you for the prompt reply. I have changed the settings as listed below:> > shared_buffers = 1024MB

work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB> fsync = off # ONLY DURING INITIAL DATA LOAD!

checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per file, check disk space)
checkpoint_timeout = 30min
checkpoint_completion_target = 0.9
wal_level = minimal # You'll need to do a full base backup if you use this

But, have same problem. It is almost 1 hour now, the restoration is still going on. After every test case execution, we would like to refresh the database and expected refresh should be completed in less than 10 minutes. Is this achievable with the kind of configuration I have listed in my earlier email. Kindly help , as how to speed up this restoration process. Thanks and RegardsRadha Krishna > Subject: Re: [GENERAL] pg_Restore

Show quoted text

From: francois@teksol.info
Date: Sun, 20 Jan 2013 23:19:44 -0500
CC: pgsql-general@postgresql.org
To: udayabhanu1984@hotmail.com

Le 2013-01-20 à 23:10, bhanu udaya a écrit :

I am new to postgres and recently got migrated from oracle. I am using postgresql 9.2 version.

Welcome, and good choice for the version.

I am trying to restore 9.5G database (1GB dumpfile) which has 500 schemas with 1 lakh rows in each schema. Could take the data dump using pg_dump and it takes around 40 minutes. I tried to use pg_restore to restore this dump, but it takes hours to restore the dump. I have used the configurations parameters as below:

shared_buffers = 1024MB
work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB

I have used command to restore as below:
pg_Restore -d newdb -j2 e:\file.dmp

My machine configurations are as below:
Windows Core I5 with 4GB Ram.

Other settings you can change during the initial restore / load phase:

fsync = off # ONLY DURING INITIAL DATA LOAD!

checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per file, check disk space)
checkpoint_timeout = 30min
checkpoint_completion_target = 0.9

wal_level = minimal # You'll need to do a full base backup if you use this

Read this section of the manual: http://www.postgresql.org/docs/current/static/runtime-config-wal.html

Have a great day!
François Beausoleil

#6Magnus Hagander
magnus@hagander.net
In reply to: bhanu udaya (#5)
Re: pg_Restore

On Jan 21, 2013 7:17 AM, "bhanu udaya" <udayabhanu1984@hotmail.com> wrote:

Hello,
Greetings !
Thank you for the prompt reply. I have changed the settings as listed

below:

shared_buffers = 1024MB
work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB

fsync = off # ONLY DURING INITIAL DATA LOAD!
checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per

file, check disk space)

checkpoint_timeout = 30min
checkpoint_completion_target = 0.9
wal_level = minimal # You'll need to do a full base backup if you use

this

But, have same problem. It is almost 1 hour now, the restoration is still

going on. After every test case execution, we would like to refresh the
database and expected refresh should be completed in less than 10 minutes.
Is this achievable with the kind of configuration I have listed in my
earlier email.

Kindly help , as how to speed up this restoration process.

Try running pg_restore with the -1 option. If that doesn't help, try -m4 or
something like that (you'll have to remove the first option then, can't use
both at once)

But it's going to be pushing it anyway. Your scenario is going to create
thousands of files (assuming you have multiple tables in each of your
schemas as is normal), and that's just not something ntfs does very fast.
Once the files are there, I bet loading the data is reasonably fast since
it can't be all that big....

/Magnus

#7bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Magnus Hagander (#6)
Re: pg_Restore

Hello,Greetings !I tried with all the below options. It approximatly takes 1 hour 30 minutes for restoring a 9GB database. This much time can not be affordable as the execution of test cases take only 10% of this whole time and waiting 1 hour 30 minutes after every test case execution is alot for the team. Kindly let me know if we can reduce the database restoration time . Thanks and RegardsRadha Krishna
Date: Mon, 21 Jan 2013 08:15:47 +0100
Subject: Re: [GENERAL] pg_Restore
From: magnus@hagander.net
To: udayabhanu1984@hotmail.com
CC: francois@teksol.info; pgsql-general@postgresql.org

On Jan 21, 2013 7:17 AM, "bhanu udaya" <udayabhanu1984@hotmail.com> wrote:

Hello,

Greetings !

Thank you for the prompt reply. I have changed the settings as listed below:

shared_buffers = 1024MB

work_mem = 512MB

maintenance_work_mem = 512MB

wal_buffers = 100MB

fsync = off # ONLY DURING INITIAL DATA LOAD!

checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per file, check disk space)

checkpoint_timeout = 30min

checkpoint_completion_target = 0.9

wal_level = minimal # You'll need to do a full base backup if you use this

But, have same problem. It is almost 1 hour now, the restoration is still going on. After every test case execution, we would like to refresh the database and expected refresh should be completed in less than 10 minutes. Is this achievable with the kind of configuration I have listed in my earlier email.

Kindly help , as how to speed up this restoration process.

Try running pg_restore with the -1 option. If that doesn't help, try -m4 or something like that (you'll have to remove the first option then, can't use both at once)
But it's going to be pushing it anyway. Your scenario is going to create thousands of files (assuming you have multiple tables in each of your schemas as is normal), and that's just not something ntfs does very fast. Once the files are there, I bet loading the data is reasonably fast since it can't be all that big....

/Magnus

#8Raghavendra
raghavendra.rao@enterprisedb.com
In reply to: bhanu udaya (#7)
Re: pg_Restore

On Mon, Jan 21, 2013 at 3:01 PM, bhanu udaya <udayabhanu1984@hotmail.com>wrote:

Hello,
Greetings !
I tried with all the below options. It approximatly takes 1 hour 30
minutes for restoring a 9GB database. This much time can not be affordable
as the execution of test cases take only 10% of this whole time and waiting
1 hour 30 minutes after every test case execution is alot for the
team. Kindly let me know if we can reduce the database restoration time .

On linux, below settings work well using using -j option of pg_restore.
Since its windows, give another try with below option.(as already best
suggested in this email).

shared_buffers= 1024MB
work_mem= 512MB
maintenance_work_mem = 1GB
checkpoint_segments=(in between 128 - 256)
checkpoint_timeout=(default is 15mns make to 1h)
autovacuum=off
track_counts=off
fsync=off
full_page_writes=off
synchronous_commit=off
bgwriter_delay=(default 200ms, change to 50ms)

---
Regards,
Raghavendra
EnterpriseDB Corporation
Blog: http://raghavt.blogspot.com/

Show quoted text

Thanks and Regards
Radha Krishna

------------------------------
Date: Mon, 21 Jan 2013 08:15:47 +0100
Subject: Re: [GENERAL] pg_Restore
From: magnus@hagander.net
To: udayabhanu1984@hotmail.com
CC: francois@teksol.info; pgsql-general@postgresql.org

On Jan 21, 2013 7:17 AM, "bhanu udaya" <udayabhanu1984@hotmail.com> wrote:

Hello,
Greetings !
Thank you for the prompt reply. I have changed the settings as listed

below:

shared_buffers = 1024MB
work_mem = 512MB
maintenance_work_mem = 512MB
wal_buffers = 100MB

fsync = off # ONLY DURING INITIAL DATA LOAD!
checkpoint_segments = 128 # large value, such as 128 or 256 (16MB per

file, check disk space)

checkpoint_timeout = 30min
checkpoint_completion_target = 0.9
wal_level = minimal # You'll need to do a full base backup if you use

this

But, have same problem. It is almost 1 hour now, the restoration is

still going on. After every test case execution, we would like to refresh
the database and expected refresh should be completed in less than 10
minutes. Is this achievable with the kind of configuration I have listed in
my earlier email.

Kindly help , as how to speed up this restoration process.

Try running pg_restore with the -1 option. If that doesn't help, try -m4
or something like that (you'll have to remove the first option then, can't
use both at once)

But it's going to be pushing it anyway. Your scenario is going to create
thousands of files (assuming you have multiple tables in each of your
schemas as is normal), and that's just not something ntfs does very fast.
Once the files are there, I bet loading the data is reasonably fast since
it can't be all that big....

/Magnus

#9Laurenz Albe
laurenz.albe@cybertec.at
In reply to: bhanu udaya (#7)
Re: pg_Restore

bhanu udaya wrote:

I tried with all the below options. It approximatly takes 1 hour 30 minutes for restoring a 9GB
database. This much time can not be affordable as the execution of test cases take only 10% of this
whole time and waiting 1 hour 30 minutes after every test case execution is alot for the team. Kindly
let me know if we can reduce the database restoration time .

I don't know if that helps, but have you tried creating a template database
and doing DROP DATABASE xxx; CREATE DATABASE xxx TEMPLATE mytemplate;
instead of restoring a dump every time?

Maybe that is faster.

Yours,
Laurenz Albe

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#10dinesh kumar
dineshkumar02@gmail.com
In reply to: Laurenz Albe (#9)
Re: pg_Restore

Hi Bhanu,

Yes, below is the faster approach to follow.

I don't know if that helps, but have you tried creating a template database

and doing DROP DATABASE xxx; CREATE DATABASE xxx TEMPLATE mytemplate;
instead of restoring a dump every time?

Maybe that is faster.

If you are trying to take the dump from one cluster and restoring it in
another cluster, then make sure your pg_restore use parallel option "-j"
and also follow the parameters what Raghav said and tune WAL_BUFFERS to
some 32 to 64 MB value. And also if possible, keep your dump file into
another partition than the PGDATA which can improve the I/O balance.

Thanks.

Best Regards,
Dinesh
manojadinesh.blogspot.com

#11Chris Travers
chris.travers@gmail.com
In reply to: Laurenz Albe (#9)
Re: pg_Restore

On Mon, Jan 21, 2013 at 3:39 AM, Albe Laurenz <laurenz.albe@wien.gv.at>wrote:

bhanu udaya wrote:

I tried with all the below options. It approximatly takes 1 hour 30

minutes for restoring a 9GB

database. This much time can not be affordable as the execution of test

cases take only 10% of this

whole time and waiting 1 hour 30 minutes after every test case execution

is alot for the team. Kindly

let me know if we can reduce the database restoration time .

I don't know if that helps, but have you tried creating a template database
and doing DROP DATABASE xxx; CREATE DATABASE xxx TEMPLATE mytemplate;
instead of restoring a dump every time?

Also for test cases, my preferred way is to put every test case in a
transaction that cannot commit. This way the tests are safe to run on
production environments. See pgTAP for one possibility here if you are
testing stored procedures. (Application code we run through wrappers that
filter out commits.)

But also the template approach is a good one fi you cannot guarantee that
the tests always role back.

Best Wishes,
Chris Travers

#12Chris Ernst
cernst@zvelo.com
In reply to: bhanu udaya (#5)
Re: pg_Restore

On 01/20/2013 11:17 PM, bhanu udaya wrote:

I am trying to restore 9.5G database (1GB dumpfile) which has 500
schemas with 1 lakh rows in each schema. Could take the data dump using
pg_dump and it takes around 40 minutes. I tried to use pg_restore to
restore this dump, but it takes hours to restore the dump. I have used
the configurations parameters as

....

But, have same problem. It is almost 1 hour now, the restoration is
still going on. After every test case execution, we would like to
refresh the database and expected refresh should be completed in less
than 10 minutes. Is this achievable with the kind of configuration I
have listed in my earlier email.

Probably not what you want to hear, but I think This is a completely
unrealistic expectation. If it takes 40 minutes for pg_dump, I would
expect pg_restore to take at least as long and likely significantly
longer (assuming both are done on similar hardware).

pg_dump only has to read the schema(s) and data and write them to a
file. pg_restore has to read write the schema and data into a new
database *AND* re-create all of the indexes, analyze, check referential
integrity......

So if the dump is taking 40 minutes, I would expect the restore to take
somewhere in the 60-90 minute range, depending on the number of and
complexity of the indexing.

- Chris

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#13bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Chris Travers (#11)
Re: pg_Restore

Hello,Thanks alot for all your replies. I tried all settings suggested, it did not work. pg_restore is very slow. It does not come out less than 1 1/2 hour. Can you please let me know the procedure for Template. Will it restore the data also . Please update. I need the database (metadata + data) to be restored back after certain Java executions. Ex:- I have 9.5 gb database (with 500 schemas + data). This is treated as base database and it needs to be restored every time after certain transactions. Thanks for your reply. Thanks and REgardsRadha Krishna
Date: Mon, 21 Jan 2013 04:16:19 -0800
Subject: Re: [GENERAL] pg_Restore
From: chris.travers@gmail.com
To: laurenz.albe@wien.gv.at
CC: udayabhanu1984@hotmail.com; magnus@hagander.net; francois@teksol.info; pgsql-general@postgresql.org

On Mon, Jan 21, 2013 at 3:39 AM, Albe Laurenz <laurenz.albe@wien.gv.at> wrote:

bhanu udaya wrote:

I tried with all the below options. It approximatly takes 1 hour 30 minutes for restoring a 9GB

database. This much time can not be affordable as the execution of test cases take only 10% of this

whole time and waiting 1 hour 30 minutes after every test case execution is alot for the team. Kindly

let me know if we can reduce the database restoration time .

I don't know if that helps, but have you tried creating a template database

and doing DROP DATABASE xxx; CREATE DATABASE xxx TEMPLATE mytemplate;

instead of restoring a dump every time?

Also for test cases, my preferred way is to put every test case in a transaction that cannot commit. This way the tests are safe to run on production environments. See pgTAP for one possibility here if you are testing stored procedures. (Application code we run through wrappers that filter out commits.)

But also the template approach is a good one fi you cannot guarantee that the tests always role back.
Best Wishes,Chris Travers

#14Laurenz Albe
laurenz.albe@cybertec.at
In reply to: bhanu udaya (#13)
Re: pg_Restore

bhanu udaya wrote:

Can you please let me know the procedure for Template. Will it restore the data also
.

It will create a complete copy of an existing database

The procedure is
CREATE DATABASE newdb TEMPLATE olddb;

Nobody may be connected to olddb for this to work.

Yours,
Laurenz Albe

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#15Alban Hertroys
haramrae@gmail.com
In reply to: bhanu udaya (#13)
Re: pg_Restore

On 21 January 2013 16:10, bhanu udaya <udayabhanu1984@hotmail.com> wrote:

Can you please let me know the procedure for Template.

As they say, Google is your friend.

The basic principle is this: You create a read-only (template) version of
your sample database and use that as a template for the creation of new
ones. Of course, now you have another copy of the database stored, which
takes up another 9.5GB of disk space, but that's not much on modern systems
(oh wait, Windows & 4GB memory?...)

pg_dump would probably be quite a lot faster on a less limited system -
databases like memory and fast raid arrays for disks.

#16Alan Hodgson
ahodgson@simkin.ca
In reply to: bhanu udaya (#13)
Re: pg_Restore

On Monday, January 21, 2013 08:40:05 PM bhanu udaya wrote:

Hello,Thanks alot for all your replies. I tried all settings suggested, it
did not work. pg_restore is very slow. It does not come out less than 1 1/2
hour. Can you please let me know the procedure for Template. Will it
restore the data also . Please update. I need the database (metadata +
data) to be restored back after certain Java executions. Ex:- I have 9.5
gb database (with 500 schemas + data). This is treated as base database and
it needs to be restored every time after certain transactions.

Don't use pg_restore, do the backups/restorations outside of PostgreSQL:

- run on a server with a snapshot-capable volume manager, use that for quick
restores
- just try using rsync from a backup copy of the base data directory

(either of the above require PostgreSQL to not be running during the
restorations)

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#17bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Laurenz Albe (#14)
Re: pg_Restore

Hello,Thanks all for the great help. Template is very good option so far seen. It comes little quickly than pg_restore. But, this also takes 40 minutes time. I am using Windows with 4GB Ram. Thanks and RegardsRadha Krishna

Show quoted text

From: laurenz.albe@wien.gv.at
To: udayabhanu1984@hotmail.com; chris.travers@gmail.com
CC: magnus@hagander.net; francois@teksol.info; pgsql-general@postgresql.org
Subject: RE: [GENERAL] pg_Restore
Date: Mon, 21 Jan 2013 15:19:45 +0000

bhanu udaya wrote:

Can you please let me know the procedure for Template. Will it restore the data also
.

It will create a complete copy of an existing database

The procedure is
CREATE DATABASE newdb TEMPLATE olddb;

Nobody may be connected to olddb for this to work.

Yours,
Laurenz Albe

#18bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#17)
Re: pg_Restore

Hello All,
Can we achieve this template or pg_Restore in less than 20 minutes time. Any more considerations. Kindly reply. Thanks and RegardsRadha Krishna From: udayabhanu1984@hotmail.com
To: laurenz.albe@wien.gv.at; chris.travers@gmail.com
CC: magnus@hagander.net; francois@teksol.info; pgsql-general@postgresql.org
Subject: RE: [GENERAL] pg_Restore
Date: Mon, 21 Jan 2013 21:45:32 +0530

Hello,
Thanks all for the great help. Template is very good option so far seen. It comes little quickly than pg_restore. But, this also takes 40 minutes time.

I am using Windows with 4GB Ram.

Thanks and Regards
Radha Krishna

Show quoted text

From: laurenz.albe@wien.gv.at
To: udayabhanu1984@hotmail.com; chris.travers@gmail.com
CC: magnus@hagander.net; francois@teksol.info; pgsql-general@postgresql.org
Subject: RE: [GENERAL] pg_Restore
Date: Mon, 21 Jan 2013 15:19:45 +0000

bhanu udaya wrote:

Can you please let me know the procedure for Template. Will it restore the data also
.

It will create a complete copy of an existing database

The procedure is
CREATE DATABASE newdb TEMPLATE olddb;

Nobody may be connected to olddb for this to work.

Yours,
Laurenz Albe

#19Adrian Klaver
adrian.klaver@aklaver.com
In reply to: bhanu udaya (#18)
Re: pg_Restore

On 01/21/2013 08:46 AM, bhanu udaya wrote:

Hello All,

Can we achieve this template or pg_Restore in less than 20 minutes time.
Any more considerations. Kindly reply.

Seems to me this is where Point in Time Recovery(PITR) might be helpful.

http://www.postgresql.org/docs/9.2/static/continuous-archiving.html

Thanks and Regards

--
Adrian Klaver
adrian.klaver@gmail.com

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#20Kevin Grittner
Kevin.Grittner@wicourts.gov
In reply to: Adrian Klaver (#19)
Re: pg_Restore

Adrian Klaver wrote:

On 01/21/2013 08:46 AM, bhanu udaya wrote:

Can we achieve this template or pg_Restore in less than 20
minutes time.

Seems to me this is where Point in Time Recovery(PITR) might be
helpful.

Maybe, if the source is on a different set of drives, to reduce
contention for storage and head movement. Either way it is down to
just a straight file copy, so it is the speed of your disk system
that is the limiting factor, not anything within PostgreSQL.

The more sure way of speeding it up is to add more spindles to your
drive array and make sure you are using a good RAID controller with
battery-backed cache.

-Kevin

--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general

#21Eduardo Morras
emorrasg@yahoo.es
In reply to: bhanu udaya (#18)
#22C. Bensend
benny@bennyvision.com
In reply to: Eduardo Morras (#21)
#23John R Pierce
pierce@hogranch.com
In reply to: C. Bensend (#22)
#24bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Kevin Grittner (#20)
#25bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#24)
#26Craig Ringer
craig@2ndquadrant.com
In reply to: bhanu udaya (#25)
#27Alban Hertroys
haramrae@gmail.com
In reply to: bhanu udaya (#24)
#28bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#24)
#29AI Rumman
rummandba@gmail.com
In reply to: bhanu udaya (#28)
#30John R Pierce
pierce@hogranch.com
In reply to: bhanu udaya (#28)
#31bhanu udaya
udayabhanu1984@hotmail.com
In reply to: AI Rumman (#29)
#32Richard Huxton
dev@archonet.com
In reply to: bhanu udaya (#28)
#33Merlin Moncure
mmoncure@gmail.com
In reply to: Richard Huxton (#32)
#34Alan Hodgson
ahodgson@simkin.ca
In reply to: Merlin Moncure (#33)
#35Tom Lane
tgl@sss.pgh.pa.us
In reply to: Alan Hodgson (#34)
#36bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#28)
#37Laurenz Albe
laurenz.albe@cybertec.at
In reply to: bhanu udaya (#36)
#38bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Laurenz Albe (#37)
#39John R Pierce
pierce@hogranch.com
In reply to: bhanu udaya (#38)
#40Alban Hertroys
haramrae@gmail.com
In reply to: bhanu udaya (#38)
#41bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Alban Hertroys (#40)
#42Alban Hertroys
haramrae@gmail.com
In reply to: bhanu udaya (#41)
#43Lee Hachadoorian
Lee.Hachadoorian+L@gmail.com
In reply to: bhanu udaya (#41)
#44Joshua D. Drake
jd@commandprompt.com
In reply to: bhanu udaya (#36)
#45bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Lee Hachadoorian (#43)
#46bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Joshua D. Drake (#44)
#47John R Pierce
pierce@hogranch.com
In reply to: bhanu udaya (#46)
#48Joshua D. Drake
jd@commandprompt.com
In reply to: bhanu udaya (#46)
#49Neil Tiffin
neilt@neiltiffin.com
In reply to: bhanu udaya (#46)
#50Michael Shapiro
mshapiro51@gmail.com
In reply to: Joshua D. Drake (#48)
#51bhanu udaya
udayabhanu1984@hotmail.com
In reply to: Neil Tiffin (#49)
#52bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#51)
#53bhanu udaya
udayabhanu1984@hotmail.com
In reply to: bhanu udaya (#52)
#54Arjen Nienhuis
a.g.nienhuis@gmail.com
In reply to: bhanu udaya (#52)
#55Laurenz Albe
laurenz.albe@cybertec.at
In reply to: bhanu udaya (#38)
#56Ingmar Brouns
swingi@gmail.com
In reply to: Laurenz Albe (#55)
#57Laurenz Albe
laurenz.albe@cybertec.at
In reply to: Ingmar Brouns (#56)