This article says that NASA uses 15 digits after the decimal point, which I’m counting as 16 in total, since that’s how we count significant digits in scientific notation. If you round pi to 3, that’s one significant digit, and if you round it to 1, that’s zero digits.
I know that 22/7 is an extremely good approximation for pi, since it’s written with 3 digits, but is accurate to almost 4 digits. Another good one is √10, which is accurate to a little over 2 digits.
I’ve heard that ‘field engineers’ used to use these approximations to save time when doing math by hand. But what field, exactly? Can anyone give examples of fields that use fewer than 16 digits? In the spirit of something like xkcd: Purity, could you rank different sciences by how many digits of pi they require?
Answering my own question: I work in web development and my usual value for pi is the standard JavaScript Math.PI. JavaScript uses 64-bit floats, which are accurate to about 15 decimal places. But that’s how many digits the computer uses. For practical math, I don’t think I’ve ever needed more than 2 digits of accuracy in an equation involving pi.