Scaleable DB structure for counters...

Started by Eci Soujiover 19 years ago9 messagesgeneral
Jump to latest
#1Eci Souji
eci.souji@gmail.com

So we've got a table called "books" and we want to build records of how
often each book is accessed and when. How would you store such
information so that it wouldn't become a huge unmanageable table?
Before I go out trying to plan something like this I figured I'd ask and
see if anyone had any experience with such a beast.

One idea I had was to create a separate DB for these counters and create
a schema for each year. Within each year schema I would create month
tables. Then I'd write a function to hit whatever schema existed like,
ala...

SELECT * FROM public.get_counters(date, hour, book_id);

get_day_counters would break up the date and based on the year do a
select counters from "2006".may WHERE day=12 and book_id=37. If hour
had a value it could do select counters from "2006".may where day=12 and
book_id=37 and hour=18.

Offline scripts would take care of generating and populating these
tables, as they'd be historical and never real-time.

Thoughts? I'm hoping someone has done something similar and can point
me in the right direction.

- E

#2Harald Armin Massa
haraldarminmassa@gmail.com
In reply to: Eci Souji (#1)
Re: Scaleable DB structure for counters...

Eci,

the usual way is:

create table books (id_book serial, author text, title text ...)
create table access (id_access serial, id_book int4, timeofaccess
timestamp,...)

then for every access you write 1 record to access.

A rough estimate: a book may be lent out every hour once, so that is 8544
records per year and book;

IF you expect that table gets "to big", you still can move over to
inheritance:

create table access2006 inherits access
create table access2007 inherits access

and put rules on them to make sure the data goes into the correct table when
you access only the access table. Google up "constraint exclusion" within
the 8.1 release notes / the postgresql documentation.

Harald

On 7/16/06, Eci Souji <eci.souji@gmail.com> wrote:

So we've got a table called "books" and we want to build records of how
often each book is accessed and when. How would you store such
information so that it wouldn't become a huge unmanageable table?
Before I go out trying to plan something like this I figured I'd ask and
see if anyone had any experience with such a beast.

One idea I had was to create a separate DB for these counters and create
a schema for each year. Within each year schema I would create month
tables. Then I'd write a function to hit whatever schema existed like,
ala...

SELECT * FROM public.get_counters(date, hour, book_id);

get_day_counters would break up the date and based on the year do a
select counters from "2006".may WHERE day=12 and book_id=37. If hour
had a value it could do select counters from "2006".may where day=12 and
book_id=37 and hour=18.

Offline scripts would take care of generating and populating these
tables, as they'd be historical and never real-time.

Thoughts? I'm hoping someone has done something similar and can point
me in the right direction.

- E

---------------------------(end of broadcast)---------------------------
TIP 6: explain analyze is your friend

--
GHUM Harald Massa
persuadere et programmare
Harald Armin Massa
Reinsburgstraße 202b
70197 Stuttgart
0173/9409607
-
on different matter:
EuroPython 2006 is over. It was a GREAT conference. If you missed it, now
you can prepare budget for visiting EuroPython 2007.

#3Eci Souji
eci.souji@gmail.com
In reply to: Harald Armin Massa (#2)
Re: Scaleable DB structure for counters...

What if instead of book checkouts we were looking at how often a book
was referenced? In which case we're talking multiple times an hour, and
we could easily have each book requiring hundreds of thousands of rows.
Multiply that by hundreds of thousands of books and a the table seems
to become huge quite quick. Would breaking up the table by year still
make sense? I'm just not familiar with having to deal with a table that
could easily hit millions of records.

Thanks for your reply,

- E

Harald Armin Massa wrote:

Show quoted text

Eci,

the usual way is:

create table books (id_book serial, author text, title text ...)
create table access (id_access serial, id_book int4, timeofaccess
timestamp,...)

then for every access you write 1 record to access.

A rough estimate: a book may be lent out every hour once, so that is
8544 records per year and book;

IF you expect that table gets "to big", you still can move over to
inheritance:

create table access2006 inherits access
create table access2007 inherits access

and put rules on them to make sure the data goes into the correct table
when you access only the access table. Google up "constraint exclusion"
within the 8.1 release notes / the postgresql documentation.

Harald

On 7/16/06, *Eci Souji* <eci.souji@gmail.com
<mailto:eci.souji@gmail.com>> wrote:

So we've got a table called "books" and we want to build records of how
often each book is accessed and when. How would you store such
information so that it wouldn't become a huge unmanageable table?
Before I go out trying to plan something like this I figured I'd ask and
see if anyone had any experience with such a beast.

One idea I had was to create a separate DB for these counters and create
a schema for each year. Within each year schema I would create month
tables. Then I'd write a function to hit whatever schema existed like,
ala...

SELECT * FROM public.get_counters(date, hour, book_id);

get_day_counters would break up the date and based on the year do a
select counters from "2006".may WHERE day=12 and book_id=37. If hour
had a value it could do select counters from "2006".may where day=12 and
book_id=37 and hour=18.

Offline scripts would take care of generating and populating these
tables, as they'd be historical and never real-time.

Thoughts? I'm hoping someone has done something similar and can point
me in the right direction.

- E

---------------------------(end of
broadcast)---------------------------
TIP 6: explain analyze is your friend

--
GHUM Harald Massa
persuadere et programmare
Harald Armin Massa
Reinsburgstra�e 202b
70197 Stuttgart
0173/9409607
-
on different matter:
EuroPython 2006 is over. It was a GREAT conference. If you missed it,
now you can prepare budget for visiting EuroPython 2007.

#4Harald Armin Massa
haraldarminmassa@gmail.com
In reply to: Eci Souji (#3)
Re: Scaleable DB structure for counters...

Eci,

I could not google them up quickly, but there are people dealing with tables
with millions of records in PostgreSQL.

Per technical data the number of rows in a table is unlimited in PostgreSQL:
http://www.postgresql.org/about/

There may be performance-reasons to split up a table of that size, but still
you can trust PostgreSQLs table inheritance together with constraint
exclusion to deal with that: just inherit your tables on a monthly base:

create table access200601 inherits ....

and adjust your rules accordingly.

read up on this documentation for examples of table partitioning,

http://www.postgresql.org/docs/8.1/interactive/ddl-partitioning.html

what is the technical term for the method you are looking for.

Best wishes,

Harald

--
GHUM Harald Massa
persuadere et programmare
Harald Armin Massa
Reinsburgstraße 202b
70197 Stuttgart
0173/9409607
-
on different matter:
EuroPython 2006 is over. It was a GREAT conference. If you missed it, now
you can prepare budget for visiting EuroPython 2007.

In reply to: Eci Souji (#1)
Re: Scaleable DB structure for counters...

On 7/16/06, Eci Souji <eci.souji@gmail.com> wrote:

So we've got a table called "books" and we want to build records of how
often each book is accessed and when. How would you store such
information so that it wouldn't become a huge unmanageable table?
Before I go out trying to plan something like this I figured I'd ask and
see if anyone had any experience with such a beast.

from your other email i understand that you will deal with a milion or so
records in access-list table.
the simplest approach would be not to divide it into multiple tables, but
instead just add triger on access table to increment counters perbook.
simple and very effective.

depesz

#6Christian Kratzer
ck-lists@cksoft.de
In reply to: Eci Souji (#3)
Re: Scaleable DB structure for counters...

Hi,

On Sun, 16 Jul 2006, Eci Souji wrote:

What if instead of book checkouts we were looking at how often a book was
referenced? In which case we're talking multiple times an hour, and we could
easily have each book requiring hundreds of thousands of rows. Multiply that
by hundreds of thousands of books and a the table seems to become huge quite
quick. Would breaking up the table by year still make sense? I'm just not
familiar with having to deal with a table that could easily hit millions of
records.

you might want to keep a separate table with counters per book
and per year or month which you regularly compute from your yearly
or month totals.

something like following untested code:

INSERT INTO access_count
SELECT id_book, date_trunc('day',timeofaccess) AS dayofaccess,count(id_book)
FROM access
WHERE date_trunc('day',timeofaccess) = date_trunc('day',now())
GROUP BY id_book, dayofaccess

That way you do not need to count all the access records.
You just sum up the pre computed counts for each period.

SELECT sum(count) FROM access_count WHERE id_book=?

You also have the option of throwing away the raw access data
for a certain day or month once that period of time is over.

This is more efficient than calling a trigger on each access and
also more scalable as there is no contention over a per book count
record.

Keeping the raw data in per month or year partitions is also propably
a good idea as it allows you to easily drop specific partitions.

Greetings
Christian

--
Christian Kratzer ck@cksoft.de
CK Software GmbH http://www.cksoft.de/
Phone: +49 7452 889 135 Fax: +49 7452 889 136

#7Ron Johnson
ron.l.johnson@cox.net
In reply to: Eci Souji (#3)
Re: Scaleable DB structure for counters...

-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1

Eci Souji wrote:

What if instead of book checkouts we were looking at how often a book
was referenced? In which case we're talking multiple times an hour, and
we could easily have each book requiring hundreds of thousands of rows.
Multiply that by hundreds of thousands of books and a the table seems
to become huge quite quick. Would breaking up the table by year still
make sense? I'm just not familiar with having to deal with a table that
could easily hit millions of records.

Are all 200000 books accessed every hour? What kind of library is
this? Do you have robot librarians moving at hyperspeed? Wouldn't
a more reasonable value be 5000 books per *day*?

It's easy to know when a book is checked out. How do you know when
a book is referenced? Are all books only accessed by the librarians?

- --
Ron Johnson, Jr.
Jefferson LA USA

Is "common sense" really valid?
For example, it is "common sense" to white-power racists that
whites are superior to blacks, and that those with brown skins
are mud people.
However, that "common sense" is obviously wrong.
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.3 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFEumvvS9HxQb37XmcRAg0wAKDOYZWThOFbIK3MWJw9RoD6Ql3zawCfdrBf
k3GCt4HEvMtrxfQMQM2Wv9M=
=fW2P
-----END PGP SIGNATURE-----

#8Eci Souji
eci.souji@gmail.com
In reply to: Ron Johnson (#7)
Re: Scaleable DB structure for counters...

I think "books" may have thrown everyone for a loop. These are not
physical books, but rather complete scanned collections that would be
available for search and reference online. One of the most important
features required would be keeping track of how often each book was
referenced and when. Time of day, days of week, etc. This is why I was
looking into how to construct some form of counter system that would
allow us to keep track of accesses.

Although I would love to see a robot librarian at work. :-)

- E

Ron Johnson wrote:

Show quoted text

-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1

Eci Souji wrote:

What if instead of book checkouts we were looking at how often a book
was referenced? In which case we're talking multiple times an hour, and
we could easily have each book requiring hundreds of thousands of rows.
Multiply that by hundreds of thousands of books and a the table seems
to become huge quite quick. Would breaking up the table by year still
make sense? I'm just not familiar with having to deal with a table that
could easily hit millions of records.

Are all 200000 books accessed every hour? What kind of library is
this? Do you have robot librarians moving at hyperspeed? Wouldn't
a more reasonable value be 5000 books per *day*?

It's easy to know when a book is checked out. How do you know when
a book is referenced? Are all books only accessed by the librarians?

- --
Ron Johnson, Jr.
Jefferson LA USA

Is "common sense" really valid?
For example, it is "common sense" to white-power racists that
whites are superior to blacks, and that those with brown skins
are mud people.
However, that "common sense" is obviously wrong.
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.3 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFEumvvS9HxQb37XmcRAg0wAKDOYZWThOFbIK3MWJw9RoD6Ql3zawCfdrBf
k3GCt4HEvMtrxfQMQM2Wv9M=
=fW2P
-----END PGP SIGNATURE-----

---------------------------(end of broadcast)---------------------------
TIP 3: Have you checked our extensive FAQ?

http://www.postgresql.org/docs/faq

#9Ron Johnson
ron.l.johnson@cox.net
In reply to: Eci Souji (#8)
Re: Scaleable DB structure for counters...

-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1

IOW, files. No problem.

The # of files is known. That's a start. Is there any existing
metric as to how often they are accessed? That's what you need to
know before deciding on a design.

This simple design might be perfectly feasible:
CREATE TABLE T_USAGE_TXN (
BOOK_ID INTEGER,
USER_ID INTEGER,
REFERENCED_DT DATE,
REFERENCED_TM TIME )

*All* the rows (in field order) would be the PK, and I'd then add
secondary indexes on
USER_ID/BOOK_ID
REFERENCED_DT/BOOK_ID
REFERENCED_DT/USER_ID

Lastly, create and algorithmically *pre-populate* this table :
T_CALENDAR (
DATE_ANSI DATE,
YEARNUM SMALLINT,
MONTH_NUM SMALLINT,
DAY_OF_MONTH SMALLINT,
DAY_OF_WEEK SMALLINT,
JULIAN_DAY SMALLINT)

So, if you want a list and count of all books that were referenced
on Sundays in 2006:

SELECT UT.BOOK_ID, COUNT(*)
FROM T_USAGE_COUNT UT,
T_CALENDAR C
WHERE C.YEARNUM = 2006
AND C.DAY_OF_WEEK = 0
AND C.DATE_ANSI = UT.REFERENCED_DT;

Eci Souji wrote:

I think "books" may have thrown everyone for a loop. These are
not physical books, but rather complete scanned collections that
would be available for search and reference online. One of the
most important features required would be keeping track of how
often each book was referenced and when. Time of day, days of
week, etc. This is why I was looking into how to construct some
form of counter system that would allow us to keep track of
accesses.

Although I would love to see a robot librarian at work. :-)

- E

Ron Johnson wrote: Eci Souji wrote:

What if instead of book checkouts we were looking at how
often a book was referenced? In which case we're talking
multiple times an hour, and we could easily have each book
requiring hundreds of thousands of rows. Multiply that by
hundreds of thousands of books and a the table seems to
become huge quite quick. Would breaking up the table by
year still make sense? I'm just not familiar with having
to deal with a table that could easily hit millions of
records.

Are all 200000 books accessed every hour? What kind of library
is this? Do you have robot librarians moving at hyperspeed?
Wouldn't a more reasonable value be 5000 books per *day*?

It's easy to know when a book is checked out. How do you know
when a book is referenced? Are all books only accessed by the
librarians?

- --
Ron Johnson, Jr.
Jefferson LA USA

Is "common sense" really valid?
For example, it is "common sense" to white-power racists that
whites are superior to blacks, and that those with brown skins
are mud people.
However, that "common sense" is obviously wrong.
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.3 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFEuxlMS9HxQb37XmcRAmXCAJ42IBwCvaDMlfMsiJoPsELxL0e1QQCfUBWH
6M7o4n9q2CEKbYn/xgh6OnY=
=iQF3
-----END PGP SIGNATURE-----