This page contains a list of user images about Double which are relevant to the point and besides images, you can also use the tabs in the bottom to browse Double news, videos, wiki information, tweets, documents and weblinks.
A substitute teacher from the inner city refuses to be messed with while taking attendance.
Jimmy Kimmel Live - Celebrities Read Mean Tweets #2 Jimmy Kimmel Live's YouTube channel features clips and recaps of every episode from the late night TV sho...
Download this song: http://bit.ly/EpicRap7 New ERB merch: http://bit.ly/MNwYxq Tweet this Vid-ee-oh: http://clicktotweet.com/TpUg9 Hi. My name is Nice Peter,...
Music video by P!nk performing Try (The Truth About Love - Live From Los Angeles). (C) 2012 RCA Records, a division of Sony Music Entertainment.
"Just One Last Time" feat. Taped Rai. Available to download on iTunes including remixes of : Tiësto, HARD ROCK SOFA & Deniz Koyu http://smarturl.it/DGJustOne...
Macklemore & Ryan Lewis present the official music video for Can't Hold Us feat. Ray Dalton. Can't Hold Us on iTunes: https://itunes.apple.com/us/album/cant-...
This video accidentally turned out kind of sad, ME SO SOWWY IT NOT POSED TO BE SAD WHO WANTS HUGS AND COOKIES? Also, FYI for anyone attempting this, it takes...
Jimmy reveals that he is f*@#ing Ben Affleck.
So i was pretty hesitant to make this video... but after all of your request, here is my Draw My Life video! Check out my 2nd Channel for more vlogs: http://...
Cry ▻ http://www.youtube.com/chaoticmonki Click Here To Subscribe! ▻ http://bit.ly/JoinBroArmy Fanfiction: http://www.fanfiction.net/s/9010921/1/Flowers-For-...
Follow on Twitter! - https://twitter.com/#!/GavinFree Watch this one in HD! The slow mo guys are well aware that water balloons are always good in slow motio...
In computing, double precision is a computer number format that occupies two adjacent storage locations in computer memory. A double-precision number, sometimes simply called a double, may be defined to be an integer, fixed point, or floating point (in which case it is often referred to as FP64).
Modern computers with 32-bit storage locations use two memory locations to store a 64-bit double-precision number (a single storage location can hold a single-precision number). Double-precision floating-point is an IEEE 754 standard for encoding binary or decimal floating-point numbers in 64 bits (8 bytes).
IEEE 754 double-precision binary floating-point format: binary64 
Double-precision binary floating-point is a commonly used format on PCs, due to its wider range over single-precision floating point, in spite of its performance and bandwidth cost. As with single-precision floating-point format, it lacks precision on integer numbers when compared with an integer format of the same size. It is commonly known simply as double. The IEEE 754 standard specifies a binary64 as having:
This gives from 15–17 significant decimal digits precision. If a decimal string with at most 15 significant decimal is converted to IEEE 754 double precision and then converted back to the same number of significant decimal, then the final string should match the original; and if an IEEE 754 double precision is converted to a decimal string with at least 17 significant decimal and then converted back to double, then the final number must match the original.
The format is written with the significand having an implicit integer bit of value 1, unless the written exponent is all zeros. With the 52 bits of the fraction significand appearing in the memory format, the total precision is therefore 53 bits (approximately 16 decimal digits, 53 log10(2) ≈ 15.955). The bits are laid out as follows:
The real value assumed by a given 64-bit double-precision data with a given biased exponent e and a 52-bit fraction is or more precisely:
Between 252=4,503,599,627,370,496 and 253=9,007,199,254,740,992 the representable numbers are exactly the integers. For the next range, from 253 to 254, everything is multiplied by 2, so the representable numbers are the even ones, etc. Conversely, for the previous range from 251 to 252, the spacing is 0.5, etc.
The spacing as a fraction of the numbers in the range from 2n to 2n+1 is 2n−52. The maximum relative rounding error when rounding a number to the nearest representable one (the machine epsilon) is therefore 2−53.
Exponent encoding 
The double-precision binary floating-point exponent is encoded using an offset-binary representation, with the zero offset being 1023; also known as exponent bias in the IEEE 754 standard. Examples of such representations would be:
- Emin (1) = −1022
- E (50) = −973
- Emax (2046) = 1023
Thus, as defined by the offset-binary representation, in order to get the true exponent the exponent bias of 1023 has to be subtracted from the written exponent.
7ff16 have a special meaning:
00016is used to represent zero (if M=0) and subnormals (if M≠0); and
7ff16is used to represent ∞ (if M=0) and NaNs (if M≠0),
where M is the fraction mantissa. All bit patterns are valid encoding.
Except for the above exceptions, the entire double-precision number is described by:
Double-precision examples 
3ff0 0000 0000 000016 = 1 3ff0 0000 0000 000116 ≈ 1.0000000000000002, the smallest number > 1 3ff0 0000 0000 000216 ≈ 1.0000000000000004 4000 0000 0000 000016 = 2 c000 0000 0000 000016 = –2
0000 0000 0000 000116 = 2-1022-52 ≈ 5 x 10−324 (Min subnormal positive double, has a precision of only one bit, i.e. ±2−1075) 000f ffff ffff ffff16 = 2−1022 - 2-1022-52 ≈ 2.2250738585072009 x 10−308 (Max subnormal double) 0010 0000 0000 000016 = 2−1022 ≈ 2.2250738585072014 x 10−308 (Min normal positive double) 7fef ffff ffff ffff16 = (1 + (1 - 2−52)) x 21023 ≈ 1.7976931348623157 x 10308 (Max Double)
0000 0000 0000 000016 = 0 8000 0000 0000 000016 = –0
7ff0 0000 0000 000016 = ∞ fff0 0000 0000 000016 = −∞
3fd5 5555 5555 555516 ≈ 1/3
(1/3 rounds down instead of up like single precision, because of the odd number of bits in the significand.)
In more detail:
Given the hexadecimal representation 3FD5 5555 5555 555516, Sign = 0 Exponent = 3FD16 = 1021 Exponent Bias = 1023 (constant value; see above) Significand = 5 5555 5555 555516 Value = 2(Exponent − Exponent Bias) × 1.Significand – Note the Significand must not be converted to decimal here = 2−2 × (15 5555 5555 555516 × 2−52) = 2−54 × 15 5555 5555 555516 = 0.333333333333333314829616256247390992939472198486328125 ≈ 1/3
Execution speed with double-precision arithmetic 
Using floating-point variables and mathematical functions (sin(), cos(), atan2(), log(), exp(), sqrt() are the most popular ones) of double precision as opposed to single precision comes at execution cost: the operations with double precision are usually slower. On average, on a PC of year 2012 build, calculations with double precision are 1.1–1.6 times slower than with single precision.
See also 
Notes and references 
- William Kahan (1 October 1987). "Lecture Notes on the Status of IEEE Standard 754 for Binary Floating-Point Arithmetic".