The number of digits used to perform a given computation. The concepts of accuracy and precision are both closely related and often confused. While the accuracy
of a number
is given by the number of significant decimal (or other) digits to the right
of the decimal point in
,
the precision of
is the total number of significant decimal (or other) digits.
In many programming language, numerical computations are done with some fixed precision.