digitalmars.D.learn - 9999999999999999.0 - 9999999999999998.0
- Samir (17/17) Jan 05 2019 I saw the following thread[1] today on Hacker News that discusses
 - Adam D. Ruppe (19/23) Jan 05 2019 That's because it is done at compile time, since both are
 - Jesse Phillips (3/5) Jan 05 2019 Since you got your answer you may also like
 - Samir (6/14) Jan 06 2019 Thanks for that explanation, Adam! Very helpful.
 
I saw the following thread[1] today on Hacker News that discusses 
an article that compares how various languages compute 
9999999999999999.0 - 9999999999999998.0.  A surprisingly large 
number of languages return 2 as the answer.  I ran the following 
which returned 1:
import std.stdio: writeln;
void main(){
     writeln(cast(double)9999999999999999.0-9999999999999998.0);
}
I don't know anything about IEEE 754[2] which, according to the 
HN discussion, is the standard for floating point arthimetic, but 
was pleasantly surprised to see how D handles this.  Does anyone 
know why?
Thanks
Samir
[1] https://news.ycombinator.com/item?id=18832155
[2] https://en.wikipedia.org/wiki/IEEE_754
 Jan 05 2019
On Sunday, 6 January 2019 at 00:20:40 UTC, Samir wrote:
 import std.stdio: writeln;
 void main(){
     writeln(cast(double)9999999999999999.0-9999999999999998.0);
 }
That's because it is done at compile time, since both are 
compile-time constants. The compiler will evaluate it using the 
maximum precision available to the compiler, ignoring your 
request to cast it to double (which annoys some people who value 
predictability over precision btw). At different precisions, you 
get different results.
I suggest breaking it up into a different variable to force a 
runtime evaluation instead of using the compiler's constant 
folding.
import std.stdio: writeln;
void main(){
     double d = 9999999999999999.0;
     writeln(d-9999999999999998.0);
}
This gives 1. Making it float instead of double, you get 
something different. With real (which btw is higher precision, 
but terrible speed), you get 1 - this is what the compiler 
happened to use at compile time.
 Jan 05 2019
On Sunday, 6 January 2019 at 00:20:40 UTC, Samir wrote:[1] https://news.ycombinator.com/item?id=18832155 [2] https://en.wikipedia.org/wiki/IEEE_754Since you got your answer you may also like http://dconf.org/2016/talks/clugston.html
 Jan 05 2019
On Sunday, 6 January 2019 at 01:05:08 UTC, Adam D. Ruppe wrote:That's because it is done at compile time, since both are compile-time constants. The compiler will evaluate it using the maximum precision available to the compiler, ignoring your request to cast it to double (which annoys some people who value predictability over precision btw). At different precisions, you get different results.Thanks for that explanation, Adam! Very helpful. On Sunday, 6 January 2019 at 03:33:45 UTC, Jesse Phillips wrote:Since you got your answer you may also like http://dconf.org/2016/talks/clugston.htmlThank you for pointing out that talk, Jesse. I will set aside some time to go through that! Samir
 Jan 06 2019








 
 
 
 Adam D. Ruppe <destructionator gmail.com> 