# Floating-point Numbers Aren't Real

(Difference between revisions)
 Revision as of 22:27, 14 December 2008 (edit)← Previous diff Revision as of 00:48, 15 December 2008 (edit) (undo)Next diff → Line 1: Line 1: - Floating-point numbers are not "real numbers" in the mathematical sense. Real numbers have infinite precision and are therefore continuous and non-lossy; floating-point numbers have fixed precision and resemble "badly-behaved" integers, because they're not evenly spaced throughout their range and they have a limited range. + Floating-point numbers are not "real numbers" in the mathematical sense, even though they are called that way in some programming languages. Real numbers have infinite precision and are therefore continuous and non-lossy; floating-point numbers have fixed precision and resemble "badly-behaved" integers, because they're not evenly spaced throughout their range, and they have a limited range. To illustrate, assign 2147483647 (the largest signed 32-bit integer) to a 32-bit float variable (x, say), and print it. You'll see 2147483648. Now print x - 64. Still 2147483648. Now print x-65 and you'll get 2147483520! Why? Because the spacing between adjacent floats in that range is 128, and floating-point operations round to the nearest floating-point number. To illustrate, assign 2147483647 (the largest signed 32-bit integer) to a 32-bit float variable (x, say), and print it. You'll see 2147483648. Now print x - 64. Still 2147483648. Now print x-65 and you'll get 2147483520! Why? Because the spacing between adjacent floats in that range is 128, and floating-point operations round to the nearest floating-point number. IEEE floating-point numbers are fixed-precision numbers based on base-two scientific notation: 1.d1d2...dp-1 × 2e. ''p'' is the precision (24 for float, 53 for double). The spacing between two consecutive numbers is 21-p+e, which can be safely approximated by ε|x|, where ε is the ''machine epsilon'' (21-p). Knowing the spacing in the neighborhood of a floating-point number can help you avoid classic numerical blunders. For example, if you're performing an iterative calculation, such as searching for the root of an equation, there's no sense in asking for greater precision than the number system can give in the neighborhood of the answer. Make sure that the tolerance you request is no smaller than the spacing there; otherwise you'll loop forever. IEEE floating-point numbers are fixed-precision numbers based on base-two scientific notation: 1.d1d2...dp-1 × 2e. ''p'' is the precision (24 for float, 53 for double). The spacing between two consecutive numbers is 21-p+e, which can be safely approximated by ε|x|, where ε is the ''machine epsilon'' (21-p). Knowing the spacing in the neighborhood of a floating-point number can help you avoid classic numerical blunders. For example, if you're performing an iterative calculation, such as searching for the root of an equation, there's no sense in asking for greater precision than the number system can give in the neighborhood of the answer. Make sure that the tolerance you request is no smaller than the spacing there; otherwise you'll loop forever.

## Revision as of 00:48, 15 December 2008

Floating-point numbers are not "real numbers" in the mathematical sense, even though they are called that way in some programming languages. Real numbers have infinite precision and are therefore continuous and non-lossy; floating-point numbers have fixed precision and resemble "badly-behaved" integers, because they're not evenly spaced throughout their range, and they have a limited range.

To illustrate, assign 2147483647 (the largest signed 32-bit integer) to a 32-bit `float<code> variable (<code>x`, say), and print it. You'll see 2147483648. Now print `x - 64`. Still 2147483648. Now print `x-65` and you'll get 2147483520! Why? Because the spacing between adjacent floats in that range is 128, and floating-point operations round to the nearest floating-point number.

IEEE floating-point numbers are fixed-precision numbers based on base-two scientific notation: 1.d1d2...dp-1 × 2e. p is the precision (24 for `float`, 53 for `double`). The spacing between two consecutive numbers is 21-p+e, which can be safely approximated by ε|x|, where ε is the machine epsilon (21-p). Knowing the spacing in the neighborhood of a floating-point number can help you avoid classic numerical blunders. For example, if you're performing an iterative calculation, such as searching for the root of an equation, there's no sense in asking for greater precision than the number system can give in the neighborhood of the answer. Make sure that the tolerance you request is no smaller than the spacing there; otherwise you'll loop forever.