Mark-10 defines accuracy as a percentage of full scale of the instrument. To determine the measurement error as
an actual load value, multiply the accuracy percentage by the instrument’s capacity.
Example 1 –
M5-50 Force Gauge :
The accuracy is ±0.1% of full scale (FS). Multiply ±0.1% by 50 lbF, which equals ±0.05 lbF. This means that any
displayed reading may be higher or lower by up to 0.05 lbF. For example,
if the displayed value is 30.00 lbF, the true reading will be ≥29.95 lbF and ≤30.05 lbF.
Example 2 – Plug & Test ® Indicators and Sensors:
The accuracies of the sensor and the indicator must be added together. Models 7i and 5i indicators have accuracy
values of ±0.1% FS, while the Model 3i is rated at ±0.2% FS. Using the example of a
Series R50 Torque Sensor with Model 3i indicator, add ±0.35% to ±0.2%, which equals ±0.55%. In a specific example for the Model
MR50-12, the accuracy becomes ±0.55% x 135 Ncm = ±0.7425 Ncm.
Percentage of Reading:
Because of these fixed errors, lower measured values will be more inaccurate as a percentage of reading.
Further using the example of an M5-50 Force Gauge, a fixed error of ±0.05 lbF represents a higher error as a
percentage of reading for a load of 1.00 lbF than 30.00 lbF.
To calculate the error as a percentage of reading, divide the fixed error by the measured value. For a 1.00 lbF
load, the fixed error equals ±0.05 ÷ 1.00 lbF = ±5% of reading.
For a 30.00 lbF load, the fixed error equals ±0.05 ÷ 30.00 lbF = ±0.17% of reading.
Conclusion:
Because of the relationship between load and accuracy, the manufacturer recommends selecting an instrument
capacity as close as possible to the maximum measured load.