PRECISION METROLOGY

 

 

 

Introduction

 

Metrology is the scientific study of measurement.  One cannot embark on the pursuit of precision manufacturing without an equally passionate journey into the challenges (and perils!) of precision metrology.  So this document is intended to provide a brief introduction to and overview of this complex subject.  Here is a printed version of these notes.

 

Image result for micrometer tips

 

 

Precision Metrology is Hard Work!

 

The sooner you accept the wise words of Israelle Widjaja, that “properly measuring things is hard,” the sooner you’ll begin to understand how to make accurate and precise measurements.

 

memebetter.com-20180416172130

 

 

Rule of Ten

 

The Rule of Ten (or Rule of One to Ten) states the discrimination (resolution) of the measuring instrument should divide the tolerance of the characteristic to be measured into ten parts.  In other words, the gage or measuring instrument should be at least 10 times as accurate as the characteristic to be measured.  Many believe that this only applies to the instruments used to calibrate a gage or measuring instrument when in reality it applies to the choice of instrument for any precision measuring activity.  The whole idea here is to choose an instrument that is capable of detecting the amount of variation present in a given characteristic (i.e. part feature). 

 

To achieve reliable measurements, the instrument needs to be accurate enough to accept all good parts and reject all bad ones.  Conversely the gage should not reject good parts nor accept bad ones.  The real problem arises when an instrument is used that is only accurate enough to measure in thousandths and accepts parts based upon that result and the customer uses gages that discriminate to ten-thousandths and rejects parts sent to them for being 0.0008” over or under the specification limit.

 

Practically speaking, this means to reliably measure a part feature specified as +/- 0.0005” requires a measurement tool with a resolution and an accuracy of 0.0001”.

 

Image result for precision measuring instrument

 

 

Accuracy, Precision, and Reproducibility

 

Accuracy refers to how close a measurement is to a true (actual) value or a value accepted as being true.

 

Precision is a measure of the spread of different readings (i.e. repeatability), and is completely unrelated to accuracy.

 

Reproducibility is the degree to which a measurement can be reproduced or replicated by someone else working independently.

 

 

Got Calibration?

 

A measuring instrument is useless if not calibrated regularly against a reliably calibrated gage.

 

memebetter.com-20180416172352

 

 

Constant Force

 

A measuring instrument which offers no constant contact force method of measurement can never have the same level repeatability or reproducibility as one that does.  In addition, a measuring instrument that does provide constant contact force only works properly if the clutch or ratchet is rotated at consistent velocity, so technique still matters.

 

Image result for mitutoyo micrometers

 

 

NTP

 

Proper measurements should always be conducted as close to NTP (normal temp and pressure) as possible (68°F & 1atm (14.7 psia)).

 

 

Be Careful!

 

Whenever possible, measure in an environment that will not damage the part or measuring instrument if either is dropped.

 

Never touch precision ground surfaces (i.e. gage blocks, gage pins, calibration rings, precision measuring surfaces, etc.) with your bare hands, as doing so will cause them to rapidly corrode, ruining their accuracy.  Always wear gloves, remove any anti-corrosion protectant with WD-40 and a new blue shop towel, and reapply anti-corrosion protectant (LPS) after use.

 

 

Never force any measurement instrument.  If a caliper or micrometer won’t move freely, investigate why; most have a locking screw or cam, so check that it’s not tight before damaging the instrument.

 

 

Cleanliness is Key

 

Clean the contact jaws or tips with alcohol and a piece of tissue paper or a blue shop towel before use.

 

 

Got Zero?

 

Always remember to double check the zero of the measurement instrument before use.  This seems fundamental, but it’s surprisingly easy to overlook when paying attention to so many other things.  This means you will need to have calibration gages or standards for instruments which are not self-zeroing (like a 0-1” micrometers).

 

Image result for micrometer zero Image result for micrometer calibration Image result for micrometer calibration

 

Thermal Growth

 

Understand metals have a typical coefficient of linear expansion of 0.000010 in / (in-°F); therefore holding on to a measuring instrument and/or a part long enough will cause a 4″ nominal part to change length 0.0012″ due to temperature change alone (0.000010 in / (in-°F) x (4 in) x (30 °F) ≈ 0.0012 in)!

 

For this reason you should always (well, whenever practically possible) use an indicator stand to hold a precision measuring instrument and protect it from thermal growth due to body temperature.  In addition, you should always allow adequate time for the part(s) being measured to reach NPT.

 

 

Multiple Measurements

 

Always take at least three measurements to be “carelessly certain” of the ballpark value.  The deviation between these measurements should match the confidence you are seeking for the repeatability of your measurements.

 

 

Gage Blocks and Gage Pins

 

Become proficient with gage blocks and gage pins, as these are typically manufactured to ±0.000100″ or ±0.000050″ (depending on their grade rating), and are good for moderate precision calibrations.

 

When using them, always wear gloves, work over a safe surface in case you accidentally drop one (never over the open box!), and coat them with rust inhibitor (LPS) when finished.

 

 

 http://www.starrett-webber.com/RS%2081%20A1.jpg

LEFT: Instructions on how to use gage blocks (click image for video).  RIGHT: Use and care of gage blocks (click image for link).

 

 Precision Pin Gages

LEFT: Applications of gage pins (click image for video).  RIGHT: Example of gage pin set.

 

Image result for micrometer calibration https://cdn.mscdirect.com/global/images/ProductImages/9282804-21.jpg

LEFT and RIGHT: Using gage blocks to calibrate a micrometer and bore gage.

 

 

Abbé and Parallax Errors

 

Research Abbé error and parallax error to understand why calipers are not regarded very highly in metrology circles J.

 

Abbé principle states: “Only when datum lines of measuring system and measured workpiece are on the same line, is a measurement most accurate.”  As drawing shows, when there is distance (h) between measuring faces and reading axis line, there will be measuring error (ε = b-a = h tan θ).  Therefore, measuring force and tool distortion must be taken into accounted during such measurement.  Think about what happens when the jaws of a dial caliper are zeroed by bringing their flat surfaces into contact, and then a measurement is made without the jaws in flat contact against the artifact.

 

https://cdn3.bigcommerce.com/s-j9r5cx/product_images/uploaded_images/abbe-principium.png?t=1440188421 

Dial Caliper demonstrating Abbe error Dial Caliper demonstrating Abbe error

LEFT: Proper method of calibration using a length standard; RIGHT: Additional Abbé error introduced because of location of applied measurement force.

 

 

Parallax error is a perceived shift in an object’s position as it is viewed from different angles, and it is inherent in virtually every analog measurement. 

 

https://cdn.miniphysics.com/wp-content/uploads/2011/12/parallax-error.jpeg Image result for parallax error

Parallax error when reading a linear scale, as on a caliper (left) and when reading a vernier dial, as on a micrometer (right).

 

 

Indicators

 

Since I already have a document on indicators, I will simply include the link here.