All measurements have a degree of uncertainty regardless of precision and accuracy. This is caused by two factors, the limitation of the measuring instrument (systematic error) and the skill of the experimenter making the measurements (random error).
As the illustration below demonstrates:
(If you look at the slight curve the meniscus makes, the lowest part is where you actually record)
The graduated cylinder in the picture contains a certain amount of water to be measured. The amount of water is somewhere between 40 ml and 50 ml according to the marked lines. By checking to see where the bottom of the meniscus lies, referencing the ten smaller lines, the amount of water lies between 44 ml and 45 ml. The next step is to estimate the uncertainty between 44 ml and 45 ml. Making an approximate guess, the level is less than 44.5 ml but greater than 44.0 ml. We then report that the measured amount is approximately 44.1 ml. The graduated cylinder itself may be distorted such that the graduation marks contain inaccuracies providing readings slightly different from the actual volume of liquid present.
When we use tools meant for measurement, we assume that they are correct and accurate, however measuring tools are not always right. In fact, they have errors that naturally occur called systematic errors. An example of a systematic error is a weighing scale. There are two specific types of systematic errors:
Random error: Sometimes called human error, random error is determined by the experimenter's skill or ability to perform the experiment and read scientific measurements. These errors are random since the results yielded may be too high or low. Often random error determines the precision of the experiment or limits the precision. For example, if we were to time a revolution of a steadily rotating turnable, the random error would be the reaction time. Our reaction time would vary due to a delay in starting (an underestimate of the actual result) or a delay in stopping (an overestimate of the actual result).
Random vs. Systematic Errors
In this experiment a series of shots is fired at a target. Random errors are caused by anything that makes the shots inconsistent and arrive at the target at random different points. For example, the shooter has an unsteady hand or a change in the environment may distort the shooter's view. These errors would result in the scattering of shots shown by the right target in the figures to the left. A systematic error, on the other hand, would include consistent errors that always arise. For example, the gun may be misaligned or there may be some other type of technical problem with the gun. This type of error would yield a pattern similar to the left target with shots deviating roughly the same amount from the center area.
When measuring a defined length with a ruler, there is a source of uncertainty and the measurement may need estimation or rounding between two points. When doing this estimation, it is possible to over estimate and under estimate the measured value, meaning there is a possibility for random error. Also, the ruler itself may be too short or too long causing a systematic error. For example, the illustration to the right shows a pencil whose length lies between 25 cm and 26 cm. With an intermediate mark, the ruler shows in greater detail that the pencil length lies somewhere between 25.5 cm and 26 cm. Therefore, one may reasonably approximate that the length of the pencil is 25.7 cm. The presence of a systematic error, however, would likely be more subtle than a random error because the environment may affect the ruler in a difficult to notice way or the ruler itself may have slightly inaccurate markers.
Precision vs. Accuracy
Precision is often referred to as reproducibility or repeatability. For example, consider the precision with which the golf balls are shot in the figures below. A set of shots that are only precise would mean you are able to cluster your shots near each other on the green but you are not reaching your goal, which would be to get the golf balls into the hole. This concept is illustrated in the left picture of the two figures below. Accuracy, on the other hand, is how close a value is to the true or accepted value. The picture to the right demonstrates accuracy showing that the balls all get into the hypothetically large hole but are all at different corners of the hole. Therefore, the shots are not precise since they are relatively spread out but they are accurate because they all reached the hole. To sum up this concept, accuracy is the ability to hit the desired target area or measured value while precision is the agreement of shots or measured values with each other but not with the intended target or value.
Since equipment used in an experiment can only report a measured value with a certain degree of accuracy, calculating the extent to which a measurement deviates from the value accepted by the scientific community is often helpful in gauging the accuracy of equipment. Such a calculation is referred to as the percent error of a measurement and is represented by the following formula:
Percent Error = [(Experimental Result - Accepted value) / Accepted Value] X 100%
Consider this real world example to understand the role of percent error calculations in determining the accuracy of measuring equipment: A toy company that ships its products around the world must calculate fuel costs associated with transporting the weight of their standard 2 by 3 foot box. To predict shipping costs and create a reasonable budget, the company must obtain accurate mass measurements of their boxes. The accepted mass of a standard box is 0.525 kg. The company measures a sample of three dozen boxes with a sophisticated electronic scale and an analog scale each yielding an average mass of 0.531 kg and 0.49 kg, respectively. A calculation of percent error for each device yields the following results:
Percent Error of Electronic Scale = [(0.531kg - 0.525kg) / 0.525kg] X 100% = 1.14 %
Percent Error of Analog Scale = [(0.49kg - 0.525kg) / 0.525kg] X100% = -6.67%
Immediately, one notices that the electronic scale yields a far more accurate measurement with a percent error almost six times lower than the measurement obtained from the analog scale. Also note that percent error may take on a negative value as illustrated by the calculation for the analog scale. This simply indicates that the measured average lies 6.67% below the accepted value. Conversely, a positive percent error indicates that the measured average is higher than the accepted value.
While inaccuracies in measurement may arise from the systematic error of equipment or random error of the experimenter, there are methods that can be employed to reduce error:
Weighing by difference: Mass is an important measurement in many experiments and it is critical for labs to reduce error in mass measurements whenever possible. A simple way of reducing the systematic error of electronic balances commonly found in labs is to weigh masses by difference. This procedure entails the following:
(mass of container + mass of material) - (mass of container + mass of material after removing material) = mass of removed material
While most electronic balances have a "tare" or "zero" function that allows one to automatically calculate a mass by difference, equipment can be faulty so it is important to remember the fundamental logic behind weighing by difference.
An NSF funded Project