Poorly-thought-out handling of double variables in pgbench

Started by Tom Lanealmost 10 years ago7 messageshackers
Jump to latest
#1Tom Lane
tgl@sss.pgh.pa.us

While testing 7a622b273 I happened to notice this:

\set x greatest(3, 2, 4.9)
create table mytab (x numeric);
insert into mytab values(:x);

results in this table:

x
----------------------
4.900000000000000355
(1 row)

The reason for that is that the result of a "double" calculation
is coerced to text like this:

sprintf(res, "%.18e", result.u.dval);

apparently on the theory that this will result in being able to convert it
back to double with no loss of low-order bits. But of course the last two
or three digits of such output are pretty much garbage to the naked eye.
Then what gets interpolated as the variable value is something like
'4.900000000000000355e+00'.

I think this is a very poor decision; it's going to cause surprises and
probably bug reports. Moreover, if we were testing this behavior in the
buildfarm (which we are not, but only because the test coverage for
pgbench is so shoddy), we would be seeing failures, because those
low-order digits are going to be platform dependent.

The only value of doing it like this would be if people chained multiple
floating-point calculations in successive pgbench \set commands and needed
full precision to be retained from one \set to the next ... but how likely
is that?

A minimum-change fix would be to print only DBL_DIG digits here. A better
answer, perhaps, would be to store double-valued variables in double
format to begin with, coercing to text only when and if the value is
interpolated into a string. That's probably a bigger change than we
want to be putting in right now, though I'm a bit tempted to go try it.

Thoughts?

BTW, just to add insult to injury, the debug() function prints double
values with "%.f", which evidently had even less thought put into it.
That one should definitely be %g with DBL_DIG precision.

regards, tom lane

--
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

#2Fabien COELHO
coelho@cri.ensmp.fr
In reply to: Tom Lane (#1)
Re: Poorly-thought-out handling of double variables in pgbench

Hello Tom,

While testing 7a622b273 I happened to notice this:

\set x greatest(3, 2, 4.9)
create table mytab (x numeric);
insert into mytab values(:x);
x
----------------------
4.900000000000000355

The reason for that is that the result of a "double" calculation
is coerced to text like this:

sprintf(res, "%.18e", result.u.dval);

apparently on the theory that this will result in being able to convert it
back to double with no loss of low-order bits.

Yep. I did that for this very reason.

ISTM that the alternative was to reimplement the variable stuff to add &
manage types, but that induces significant changes, and such change are
likely to be rejected, or at least to raise long debates and questions,
make the patch harder to check... so I just avoided them.

But of course the last two or three digits of such output are pretty
much garbage to the naked eye.

Sure, it is binary to decimal conversion noise.

Then what gets interpolated as the variable value is something like
'4.900000000000000355e+00'.

I think this is a very poor decision; it's going to cause surprises and
probably bug reports.

Moreover, if we were testing this behavior in the buildfarm (which we
are not, but only because the test coverage for pgbench is so shoddy),

I think that "none" is the right word?

we would be seeing failures, because those low-order digits are going to
be platform dependent.

Possibly.

The only value of doing it like this would be if people chained multiple
floating-point calculations in successive pgbench \set commands and needed
full precision to be retained from one \set to the next ... but how likely
is that?

The constraint for me is that a stored double should not lose precision.
Any solution which achieves this goal is fine with me.

A minimum-change fix would be to print only DBL_DIG digits here.

Probably 15, but there is some rounding implied I think (?). I'm not that
sure of DBL_DIG+1, so I put 18 to be safer. I did not envision that
someone would store that in a NUMERIC:-) Your example would work fine with
a DOUBLE PRECISION attribute.

A better answer, perhaps, would be to store double-valued variables in
double format to begin with, coercing to text only when and if the value
is interpolated into a string.

Yep, but that was yet more changes for a limited benefit and would have
increase the probability that the patch would have been rejected.

That's probably a bigger change than we want to be putting in right now,
though I'm a bit tempted to go try it.

Thoughts?

My 0.02ᅵ:

I definitely agree that the text variable solution is pretty ugly, but it
was the minimum change solution, and I do not have much time available.

I do not think that rounding values is desirable, on principle.

Now doubles are only really there because random_*() functions need one
double argument, otherwise only integers make much sense as keys to select
tuples in a performance bench. So this is not a key feature, just a side
effect of having more realistic non uniform random generators and
expressions. In an early submission I did not bother to include double
variables because I cannot see much actual use to them, and there were
implementation issues given how integer variables were managed.

Also I think that the above NUMERIC/DOUBLE case is a little contrived, so
I would wait for someone to actually complain with a good use case before
spending any significant time to change how variables are stored in
pgbench.

I would expect a long review process for such a patch if I submitted it,
because there are details and the benefit is quite limited.

BTW, just to add insult to injury, the debug() function prints double
values with "%.f", which evidently had even less thought put into it.

AFAICR the thought I put into it is that it is definitely useful to be
able to get a simple trace for debugging purpose. The precision of this
debug output is not very important, so I probably just put the simplest
possible format.

That one should definitely be %g with DBL_DIG precision.

That would be fine as well.

--
Fabien.
--
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

#3Tom Lane
tgl@sss.pgh.pa.us
In reply to: Fabien COELHO (#2)
Re: Poorly-thought-out handling of double variables in pgbench

Fabien COELHO <coelho@cri.ensmp.fr> writes:

A better answer, perhaps, would be to store double-valued variables in
double format to begin with, coercing to text only when and if the value
is interpolated into a string.

Yep, but that was yet more changes for a limited benefit and would have
increase the probability that the patch would have been rejected.

That's probably a bigger change than we want to be putting in right now,
though I'm a bit tempted to go try it.

I definitely agree that the text variable solution is pretty ugly, but it
was the minimum change solution, and I do not have much time available.

Well, I felt like doing some minor hacking, so I went and adjusted the
code to work this way. I'm pretty happy with the result, what do you
think?

regards, tom lane

Attachments:

pgbench-real-numeric-variables.patchtext/x-diff; charset=us-ascii; name=pgbench-real-numeric-variables.patchDownload+267-231
#4Fabien COELHO
coelho@cri.ensmp.fr
In reply to: Tom Lane (#3)
Re: Poorly-thought-out handling of double variables in pgbench

Hello Tom,

That's probably a bigger change than we want to be putting in right now,
though I'm a bit tempted to go try it.

I definitely agree that the text variable solution is pretty ugly, but it
was the minimum change solution, and I do not have much time available.

Well, I felt like doing some minor hacking, so I went and adjusted the
code to work this way. I'm pretty happy with the result, what do you
think?

This is a definite improvement.

I like the lazyness between string & numeric forms, and for sorting, that
is what was needed doing to have something clean.

Applied on head, it works for me.

While testing the patch I found a minor preexisting (mine...) bug: when
string-scanning doubles, whether the whole string is consumed or not is
not checked. This means that -D x=0one is interpreted as double 0.

I came up with the attached check, but maybe there is a cleaner way to do
that.

--
Fabien.

Attachments:

pgbench-scan-double.patchtext/x-diff; name=pgbench-scan-double.patchDownload+3-1
#5Tom Lane
tgl@sss.pgh.pa.us
In reply to: Fabien COELHO (#4)
Re: Poorly-thought-out handling of double variables in pgbench

Fabien COELHO <coelho@cri.ensmp.fr> writes:

This is a definite improvement.

I like the lazyness between string & numeric forms, and for sorting, that
is what was needed doing to have something clean.

Applied on head, it works for me.

OK, pushed.

While testing the patch I found a minor preexisting (mine...) bug: when
string-scanning doubles, whether the whole string is consumed or not is
not checked. This means that -D x=0one is interpreted as double 0.

I came up with the attached check, but maybe there is a cleaner way to do
that.

I like the way the zic code does it:

else /* type should be double */
{
double dv;

! 		if (sscanf(var->value, "%lf", &dv) != 1)
  		{
  			fprintf(stderr,
  					"malformed variable \"%s\" value: \"%s\"\n",
--- 928,936 ----
  	else /* type should be double */
  	{
  		double dv;
+ 		char xs;

! if (sscanf(var->value, "%lf%c", &dv, &xs) != 1)
{
fprintf(stderr,
"malformed variable \"%s\" value: \"%s\"\n",

This relies on the idea that if there is any trailing junk, one character
will get assigned into "xs" and then sscanf will say it filled two fields.
Otherwise, it'll report filling only one field (or none, if there's
no number either). This requires only a minimal amount of code and no
extra strlen() calculations.

Neither way allows for trailing whitespace, though. I don't see that
that's important here.

regards, tom lane

--
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

#6Fabien COELHO
coelho@cri.ensmp.fr
In reply to: Tom Lane (#5)
Re: Poorly-thought-out handling of double variables in pgbench

While testing the patch I found a minor preexisting (mine...) bug: when
string-scanning doubles, whether the whole string is consumed or not is
not checked. This means that -D x=0one is interpreted as double 0.

I came up with the attached check, but maybe there is a cleaner way to do
that.

I like the way the zic code does it:

+ char xs;
! if (sscanf(var->value, "%lf%c", &dv, &xs) != 1)

Indeed... I cannot say that it is cleaner such, but at least it is short
and it avoids the strlen call.

Neither way allows for trailing whitespace, though. I don't see that
that's important here.

No, this is not an issue here for variables specified from the command
line. Consider pushing this fix?

--
Fabien.

--
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers

#7Tom Lane
tgl@sss.pgh.pa.us
In reply to: Fabien COELHO (#6)
Re: Poorly-thought-out handling of double variables in pgbench

Fabien COELHO <coelho@cri.ensmp.fr> writes:

Neither way allows for trailing whitespace, though. I don't see that
that's important here.

No, this is not an issue here for variables specified from the command
line. Consider pushing this fix?

Done already.

regards, tom lane

--
Sent via pgsql-hackers mailing list (pgsql-hackers@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers