decimal128 floating-point format
Floating-point formats |
---|
IEEE 754 |
|
Other |
Alternatives |
Tapered floating point |
This article may require cleanup to meet Wikipedia's quality standards. The specific problem is: This article spreads technical wrong info about e.g the use of bits from combination field which is different between BID and DPD, and the tables are difficult to decode. I'll try to provide step by step improvements in the next weeks and would appreciate help. (December 2024) |
Floating-point formats |
---|
IEEE 754 |
|
Other |
Alternatives |
Tapered floating point |
In computing, decimal128 is a decimal floating-point number format that occupies 16 bytes (128 bits) in memory.
Purpose and use
[edit]Like the binary128 formats, decimal128 takes place where extreme precision or ranges are to be handeled.
In contrast to the binaryxxx data formats the decimalxxx formats provide exact representation of decimal fractions, exact calculations with them and enable human common 'ties away from zero' rounding[1] (in some range, to some precision, to some degree). In a trade-off for reduced performance, which is especially harming decimal128 computations on common 64- or 32-bit hardware. They are intended for applications where it's requested to come near to schoolhouse math, such as financial and tax computations. (In short they avoid plenty of problems like 0.2 + 0.1 -> 0.30000000000000000000000000000000004 which happen with binary128 datatypes.)
Range and precision
[edit]decimal128 supports 'normal' values that can have 34 digit precision from ±1.000000000000000000000000000000000×10 −6143 to ±9.999999999999999999999999999999999×10 +6144, plus 'denormal' values with ramp-down relative precision down to ±1 × 10−6176, signed zeros, signed infinities and NaN (Not a Number).
The binary format of the same bit-size supports a range from denormal-min ±6×10 −4966, over normal-min with full 113-bit precision ±3.3621031431120935062626778173217526×10 −4932 to max ±1.189731495357231765085759326628007×10 +4932.
Representation / encoding of decimal128 values
[edit]decimal128 values are represented in a 'not normalized' near to 'scientific format', with combining some bits of the exponent with the leading bits of the significand in a 'combination field'.
Sign | Combination | Trailing significand bits |
---|---|---|
1 bit | 17 bits | 110 bits |
s | mmmmmmmmmmmmmmmmm | tttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttt |
Besides the special cases infinities and NaNs there are four points relevant to understand the encoding of decimal128.
- BID vs. DPD encoding, Binary Integer Decimal using a binary coded positive integer for the significand, software centric and designed by Intel(r), vs. Densely Packed Decimal based on densely packed decimal encoding for all except the first digit of the significand, hardware centric and promoted by IBM(r), differences see below. Both alternatives provide exactly the same range of representable numbers: 34 digits of significand and 3 × 212 = 12288 possible exponent values. IEEE 754 allows these two different encodings, without a concept to denote which is used, for instance in a situation where decimal128 values are communicated between systems. CAUTION!: Be aware that transferring binary data between systems using different encodings will mostly produce valid decimal128 numbers, but with different value. Prefer data exchange in íntegral or ASCII 'triplets' for sign, exponent and significand.
- Because the significands in the IEEE 754 decimal formats is not normalized (in contrast to the binary formats), most values with less than 34 significant digits have multiple possible representations; 1000000 × 10-2=100000 × 10-1=10000 × 100=1000 × 101 all have the value 10000. These sets of representations for a same value are called cohorts, the different members can be used to denote how many digits of the value are known precisely.
- The encodings combine two bits of the exponent with the leading 3 to 4 bits of the significand in a 'combination field', different for 'big' vs. 'small' significands. That enables bigger precision and range, in trade-off that some simple functions like sort and compare, very frequently used in coding, do not work on the bit pattern but require computations to extract exponent and significand and then try to obtain an exponent aligned representation. This effort is partly balanced by saving the effort for normalization, but contributes to the slower performance of the decimal datatypes. Beware: BID and DPD use different bits of the combination field for that, see below.
- Different understanding of significand as integer or fraction, and acc. different bias to apply for the exponent (for decimal128 what is stored in bits can be decoded as base to the power of 'stored value for the exponent minus bias of 6143' times significand understood as d0 . d−1 d−2 d−3 ... d−31 d−32 d−33 (note: radix dot after first digit, significand fractional), or base to the power of 'stored value for the exponent minus bias of 6176' times significand understood as d33 d32 d31 ... d3 d2 d1 d0 (note: no radix dot, significand integral), both produce the same result [2019 version[2] of IEEE 754 in clause 3.3, page 18]. For decimal datatypes the second view is more common, while for binary datatypes the first, the biases are different for each datatype.)
In the case of Infinity and NaN, all other bits of the encoding are ignored. Thus, it is possible to initialize an array to Infinities or NaNs by filling it with a single byte value.
Binary integer significand field
[edit]This format uses a binary significand from 0 to 1034 − 1 = 9999999999999999999999999999999999 = 1ED09BEAD87C0378D8E63FFFFFFFF16 = 0111101101000010011011111010101101100001111100000000110111100011011000111001100011111111111111111111111111111111112. The encoding can represent binary significands up to 10 × 2110 − 1 = 12980742146337069071326240823050239 but values larger than 1034 − 1 are illegal (and the standard requires implementations to treat them as 0, if encountered on input).
If the 2 bits after the sign bit are "00", "01", or "10", then the exponent field consists of the 14 bits following the sign bit, and the significand is the remaining 113 bits, with an implicit leading 0 bit:
This includes subnormal numbers where the leading significand digit is 0.
If the 2 bits after the sign bit are "11", then the 14-bit exponent field is shifted 2 bits to the right (after both the sign bit and the "11" bits thereafter), and the represented significand is in the remaining 111 bits. In this case there is an implicit (that is, not stored) leading 3-bit sequence "100" in the true significand. Compare having an implicit 1 in the significand of normal values for the binary formats. The "00", "01", or "10" bits are part of the exponent field.
For the decimal128 format, all of these significands are out of the valid range (they begin with 2113 > 1.038 × 1034), and are thus decoded as zero, but the pattern is same as for decimal32 and decimal64.
Be aware that the bit numbering used in the tables for e.g. m16 … m0 is in opposite direction than that used in the paper for the IEEE 754 standard G0 … G16.
Combination Field | Exponent | Significand / Description | |||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
m16 | m15 | m14 | m13 | m12 | m11 | m10 | m9 | m8 | m7 | m6 | m5 | m4 | m3 | m2 | m1 | m0 | |||
combination field not! starting with '11', bits ab = 00, 01 or 10 | |||||||||||||||||||
a | b | c | d | m | m | m | m | m | m | m | m | m | m | e | f | g | abcdmmmmmmmmmm | (0)efgtttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttt
Finite number, all 'legal' significands 0 .. 9999999999999999999999999999999999 fit here. | |
combination field starting with '11', but not 1111, bits ab = 11, bits cd = 00, 01 or 10 | |||||||||||||||||||
1 | 1 | c | d | m | m | m | m | m | m | m | m | m | m | e | f | g | cdmmmmmmmmmmef | 100gtttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttttt
Theoretical case, all these signifiands are > 1.0384593717069655257060992658440191 × 10^34, thus > 10^34 - 1, 'illegal' and to be treated as zero. | |
combination field starting with '1111', bits abcd = 1111 | |||||||||||||||||||
1 | 1 | 1 | 1 | 0 | ±Infinity | ||||||||||||||
1 | 1 | 1 | 1 | 1 | 0 | quiet NaN | |||||||||||||
1 | 1 | 1 | 1 | 1 | 1 | signaling NaN (with payload in significand) |
In the above cases, the value represented is
- (−1)sign × 10exponent−6176 × significand
Densely packed decimal significand field
[edit]In this version, the significand is stored as a series of decimal digits. The leading digit is between 0 and 9 (3 or 4 binary bits), and the rest of the significand uses the densely packed decimal (DPD) encoding.
The encoding varies depending on whether the most significant 4 bits of the significand are in the range 0 to 7 (00002 to 01112), or higher (10002 or 10012).
2 bits of the exponent and the leading digit (3 or 4 bits) of the significand are combined into the five bits that follow the sign bit.
This twelve bits after that are the exponent continuation field, providing the less-significant bits of the exponent.
The last 110 bits are the significand continuation field, consisting of eleven 10-bit declets.[3] Each declet encodes three decimal digits[3] using the DPD encoding.
If the first two bits after the sign bit are "00", "01", or "10", then those are the leading bits of the exponent, and the three bits after that are interpreted as the leading decimal digit (0 to 7):
If the first two bits after the sign bit are "11", then the next two bits are the leading bits of the exponent, and the fifth bit is prefixed with "100" to form the leading decimal digit of the significand (8 or 9):
Combination Field | Exponent | Significand / Description | |||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
m16 | m15 | m14 | m13 | m12 | m11 | m10 | m9 | m8 | m7 | m6 | m5 | m4 | m3 | m2 | m1 | m0 | |||
combination field not! starting with '11', bits ab = 00, 01 or 10 | |||||||||||||||||||
a | b | c | d | e | m | m | m | m | m | m | m | m | m | m | m | m | abmmmmmmmmmmmm | (0)cde tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt
Finite number with small first digit of significand (0 … 7). | |
combination field starting with '11', but not 1111, bits ab = 11, bits cd = 00, 01 or 10 | |||||||||||||||||||
1 | 1 | c | d | e | m | m | m | m | m | m | m | m | m | m | m | m | cdmmmmmmmmmmmm | 100e tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt tttttttttt
Finite number with big first digit of significand (8 or 9). | |
combination field starting with '1111', bits abcd = 1111 | |||||||||||||||||||
1 | 1 | 1 | 1 | 0 | ±Infinity | ||||||||||||||
1 | 1 | 1 | 1 | 1 | 0 | quiet NaN | |||||||||||||
1 | 1 | 1 | 1 | 1 | 1 | signaling NaN (with payload in significand) |
The remaining two combinations (11110 and 11111) of the 5-bit field
are used to represent ±infinity and NaNs, respectively.
The 10-bit DPD to 3-digit BCD transcoding for the declets is given by the following table. b9 … b0 are the bits of the DPD, and d2 … d0 are the three BCD digits. Be aware that the bit numbering used here for e.g. b9 … b0 is in opposite direction than that used in the paper for the IEEE 754 standard b0 … b9, add. the decimal digits are numbered 0-based here while in opposite direction and 1-based in the IEEE 754 paper. The bits on white background are not counting for the value, but signal how to understand / shift the other bits. The concept is to denote which digits are small (0 … 7) and encoded in three bits, and which are not, then calculated from a prefix of '100', and one bit specifying if 8 or 9.
DPD encoded value | Decimal digits | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Code space (1024 states) |
b9 | b8 | b7 | b6 | b5 | b4 | b3 | b2 | b1 | b0 | d2 | d1 | d0 | Values encoded | Description | Occurrences (1000 states) | |
50.0% (512 states) |
a | b | c | d | e | f | 0 | g | h | i | 0abc | 0def | 0ghi | (0–7) (0–7) (0–7) | 3 small digits | 51.2% (512 states) | |
37.5% (384 states) |
a | b | c | d | e | f | 1 | 0 | 0 | i | 0abc | 0def | 100i | (0–7) (0–7) (8–9) | 2 small digits, 1 large digit |
38.4% (384 states) | |
a | b | c | g | h | f | 1 | 0 | 1 | i | 0abc | 100f | 0ghi | (0–7) (8–9) (0–7) | ||||
g | h | c | d | e | f | 1 | 1 | 0 | i | 100c | 0def | 0ghi | (8–9) (0–7) (0–7) | ||||
9.375% (96 states) |
g | h | c | 0 | 0 | f | 1 | 1 | 1 | i | 100c | 100f | 0ghi | (8–9) (8–9) (0–7) | 1 small digit, 2 large digits |
9.6% (96 states) | |
d | e | c | 0 | 1 | f | 1 | 1 | 1 | i | 100c | 0def | 100i | (8–9) (0–7) (8–9) | ||||
a | b | c | 1 | 0 | f | 1 | 1 | 1 | i | 0abc | 100f | 100i | (0–7) (8–9) (8–9) | ||||
3.125% (32 states, 8 used) |
x | x | c | 1 | 1 | f | 1 | 1 | 1 | i | 100c | 100f | 100i | (8–9) (8–9) (8–9) | 3 large digits, b9, b8: don't care |
0.8% (8 states) |
The 8 decimal values whose digits are all 8s or 9s have four codings each. The bits marked x in the table above are ignored on input, but will always be 0 in computed results. (The 8 × 3 = 24 non-standard encodings fill the unused range from 103 = 1000 to 210 - 1 = 1023.)
In the above cases, with the true significand as the sequence of decimal digits decoded, the value represented is
History
[edit]decimal128 was formally introduced in the 2008 revision of the IEEE 754 standard,[5] which was taken over into the ISO/IEC/IEEE 60559:2011 standard.[6]
Side effects, more info
[edit]Zero has 12288 possible representations (24576 when both signed zeros are included), (even many more if you account the 'illegal' significands which have to be treated as zeroes).
The gain in range and precision by the 'combination encoding' evolves because the taken 2 bits from the exponent only use three states, and the 4 MSBs of the significand stay within 0000 … 1001 (10 states). In total that is 3 × 10 = 30 possible values when combined in one encoding, which is representable in 5 bits ().
See also
[edit]- ISO/IEC 10967, Language Independent Arithmetic
- Q notation (scientific notation)
References
[edit]- ^ Cowlishaw, Mike (2007). "Decimal Arithmetic FAQ – Part 1 – General Questions". speleotrove.com. IBM Corporation. Retrieved 2022-07-29.
- ^ 754-2019 - IEEE Standard for Floating-Point Arithmetic ( caution: paywall ). 2019. doi:10.1109/IEEESTD.2019.8766229. ISBN 978-1-5044-5924-2. Archived from the original on 2019-11-01. Retrieved 2019-10-24.
- ^ a b Muller, Jean-Michel; Brisebarre, Nicolas; de Dinechin, Florent; Jeannerod, Claude-Pierre; Lefèvre, Vincent; Melquiond, Guillaume; Revol, Nathalie; Stehlé, Damien; Torres, Serge (2010). Handbook of Floating-Point Arithmetic (1 ed.). Birkhäuser. doi:10.1007/978-0-8176-4705-6. ISBN 978-0-8176-4704-9. LCCN 2009939668.
- ^ Cowlishaw, Michael Frederic (2007-02-13) [2000-10-03]. "A Summary of Densely Packed Decimal encoding". IBM. Archived from the original on 2015-09-24. Retrieved 2016-02-07.
- ^ IEEE Computer Society (2008-08-29). IEEE Standard for Floating-Point Arithmetic. IEEE. doi:10.1109/IEEESTD.2008.4610935. ISBN 978-0-7381-5753-5. IEEE Std 754-2008.
- ^ "ISO/IEC/IEEE 60559:2011". 2011. Archived from the original on 2016-03-04. Retrieved 2016-02-08.