Before we define what does calibration mean is, I want you to understand below scenario:
When you go to the market and buy some meat, you bought 1 kilo and paid it with your hard-earned money but when you checked it at home, it is only ¾ of a kilo, what would you feel?
You just gassed-up good for 2 days as your regular routine but only a day had passed and your meter indicates near empty – would this not make you mad?
You are baking a cake and the instruction tells you to set the temperature to 50 deg Celsius for 20 minutes, you followed the steps and specification but your cake turned into charcoal – will you not be upset?
Now, what do you feel if the above scenario happens to you? Of course, you may get angry, dismayed or worse complain to the services or product that you received and paid for, all because of the effect of the un-calibrated Instruments.
This is why calibration is important in our daily life not just inside a laboratory or within the manufacturing aspect. There is the involvement of quality, safety, and reliability.
Now let us define What calibration is.
Basics of Calibration
There are definitions of Calibrations by NIST or ISO that you can look for. But to easily understand, below is a simple calibration definition:
Calibration is simply the comparison of Instrument, Measuring and Test Equipment (M&TE), Unit Under Test (UUT), Unit Under Calibration (UUC), a Device Under Test (DUT) or simply a Test Instrument (TI) of unverified accuracy to an instrument or standards with a known (higher) accuracy to detect or eliminate unacceptable variations. It may or may not involve adjustment or repair.
It is making the instruments performs what it displays by referencing or adjusting it based on a Reference Standard.
- to ensure that you get what you have paid for;
- Satisfy your expectations;
- Create win-win situations
What is a Reference Standard?
A reference standard is also an Instrument, or equipment, or a measuring device with the highest metrological quality or accuracy than the Unit Under Calibration (UUC).
It is where we compare the UUC reading and where measurement values are derived. It is also calibrated, but by a higher level laboratory with traceability to a higher standard (See traceability below).
The reference standard is also known as Master Standard. Other terms that I sometimes hear, they refer to it as Master Calibrator or simple calibrator.
Why Calibrate – Reasons for Calibration
There are so many reasons why we need to perform a calibration. These reasons are:
- For public or consumer protection like the example above in order to get the value of the money we spent for a product or services.
- For a technical reason, we need to calibrate because as components aged or equipment undergo environmental or mechanical stress, its performance gradually degrades.
This degradation is what we call the ‘drift‘. When this happens, the results or performance generated by certain equipment will be unreliable where design and quality suffer. We cannot eliminate drift, but through the process of calibration, it can be detected and contained.
- There are also practical reasons for implementing calibration. Calibration will eliminate the doubts and provide confidence when we encounter below situations with our instruments:
- when there is a newly installed or purchased intruments
- instruments that are mishandled during transfer (example: dropped or fell down)
- when instrument performance is questionable
- calibration period is overdue
- kept to an unstable environment for too long (exposed to vibrations or too high/low temperatures)
- when a new setting, repair, and/or adjustment is performed
While detecting an inaccuracy is one of the main reason for calibration, some other reasons are:
- Customer requirements – they want to ensure that the product they buy is within the expected specifications.
- Requirements of a government or statutory regulations – they want to make sure that products produce are safe and reliable for the public
- Audit requirements– as a requirement for achieving a certification like ISO 9001:2015 certification and ISO 17025 accreditations
- Quality and Safety requirements– a reliable and accurate operation through proper use of inspection instruments provides a great deal of confidence for everyone
- Process requirements – to ensure that the product produced is the most accurate and reliable, some operations will not be executed unless the equipment has passed the calibration and verification process, used in equipment or product qualifications as part of quality control.
Importance of Instrument Calibration
- To establish and demonstrate traceability (I will explain the traceability below). Through calibration, the measurement established by the instrument is the same where ever you are, it means that a 1 Kg of weights in one place is also 1 Kg in other places or wherever it reaches. You can use instruments regardless of the units or parameters it measures on different occasions.
- To determine and ensure the accuracy of instrument readings (through calibration, you can determine how close the actual value to the true or reference value) – resulting in product quality and safety
- To ensure readings from the instrument are consistent with other measurements. Which means that you have the same measurement results regardless of what measuring instrument you use that is compatible in the process.
- To establish the reliability of the instrument making sure that they function in the way they are intended to be – resulting in more confidence on the expected output.
- Provides customer satisfaction by having a product that meets what they have paid for – high quality of product.
What needs calibration?
- All inspection, measuring and test equipment that can affect or determine product quality. This means that if you are using the instruments to verify the acceptance of a product whether to pass or fail based on the measured value you have taken, the instrument should be calibrated.
- Equipment which, if out of calibration, would produce unsafe products.
- Equipment which requires calibration because of an agreement. An example is a customer , where before progressing into a contract, they need to ensure that the equipment that will produce their product is calibrated.
- All measuring and testing equipment (standards) having an effect on the accuracy or validity of calibrations. These are the master standards, go-no-go jigs, check masters, reference materials and related instruments that we used to verify other instruments or measuring equipment for their accuracy.
What does a calibrated Instrument Looks like?
When you have your measuring instruments calibrated, see to it that it has a calibration label where details of its calibration
date and due date are seen, also includes a serial number, certificate number, and person-in-charge of the calibration which depends on the calibration lab. Also, if needed, avoid seal is placed to protect it from an unauthorized adjustment.
A calibrated Instrument with labels are useless if the calibration certificate is not available, so be sure to keep it safe and readily available once requested.
Be cautious to check calibration certificates once received, not all calibrated instruments are performing the same as you expected, some have a limited use based on the result of the calibration. you must learn how to read or interpret the results in a calibration certificate.
When is Calibration Not Required?
Every measuring instrument needs calibration but not all measuring instruments are required to be calibrated. Below are some of the reasons or criteria to consider before having an instrument calibrated. This may save you some time and money.
- It is not critical in your process ( just to display a certain reading for the purpose of functionality check).
- It functions as an indicator only (for example: high or low and close or open).
- As an accessory only to support the main instrument. For example, a coil of wire used to amplify current. Current is measured, but the amplification is not that critical, used as an accessory only to amplify a measured current.
- Its accuracy is established by a higher or reference to a higher or more accurate instrument within a group. (for example a set of pressure gauges that are connected in series in which one of them is a more accurate gauge where they are compared or referenced to).
- If the instruments are verified regularly or continuously monitored by a calibrated instrument that is documented in a measurement assurance process. For example, a room thermometer that is verified by a calibrated thermometer regularly.
- If the instrument is a part of a system or integrated into a system where the system is calibrated as a whole. For example, a thermocouple that is permanently connected to the oven (some thermocouples are detached after usage then transferred to other units).
Please visit this link for more details regarding this topic.
How To Perform Adjustment of Calibration Interval
Every measuring instrument needs calibration and every calibrated instrument needs to be re-calibrated. This means that there is a due date for a re-calibration period that we need to establish.
This recalibration period is a scheduled calibration that is based on the calibration interval that we set. Before we send our instruments for calibration, we need to set on advance our initial interval or our final fixed interval. This will be communicated with the calibration lab.
These initial calibration interval may be based on the below aspects:
- Manufacturer Requirements – recommended by the manufacturer
- On the frequency of use – the more it is used, the shorter the calibration interval
- Required by the regulatory bodies (example: required by the government)
- Past experience of the user with the same type of instrument
- Based on the criticality of use. – more critical instruments have higher accuracy or very strict tolerance, therefore shorter calibration interval
- Customer Requirements
- Conditions of the environment where it is being used.
- Published Documents
Now that we have an initial calibration interval, we need to set the fixed calibration interval based on the performance of our instruments. We will review the performance by gathering all the calibration past records and plot it in a graph in order to study its stability or the drift encountered by the instruments.
Based on this performance history, we can decide if we need to extend or reduce the calibration interval.
I have shown how to do this in this link >> CALIBRATION INTERVAL: HOW TO INCREASE THE CALIBRATION FREQUENCY OF INSTRUMENTS
Basic Calibration Terms and Principles
What is the Accuracy of an Instrument?
Accuracy is a reference number (usually given in percent (%) error) that shows the degree of closeness to the true value. There is a true value which means that you have a source of known value to compare with.
The closer your measuring instrument reads or measure to the true value, the more accurate your instrument. Or, in other word, “the smaller the error, the more accurate the instrument”.
How to calculate accuracy of measurement?
In order to find the value of accuracy, you need to calculate the error, and to gauge the “degree of closeness to the true value”, you need to calculate the % error.
To calculate the error and % error, below are the formulas:
Error = Measured value – True value
And for the percent error:
% error = Error/True value x 100
We are performing a calibration to check for the accuracy of instruments, to determine how close (or far) the reading of our instruments compared to the reading of the reference standard.
Where can we find the Accuracy of Instruments?
- Original Equipment Manufacturer (OEM) Specifications in manual or brochure
- Publish standards or handbook
- Calibration Certificates
When calibrating, make sure that the reference standard has a higher accuracy than the Unit Under Calibration (UUC), usually, a good rule of thumb is to have an accuracy ratio of 4:1.
This means that if your UUC has an accuracy of 1, the reference standard to be used should have at least a 0.25 accuracy, four times more accurate than the UUC (1/0.25 =4). To make it more clear, since most of the time, accuracy is expressed in percentage (% error), the closer the value is to ZERO, the more accurate the instrument.
What is the Tolerance of Instruments?
Tolerance is closely related to accuracy at some point but for clarity, it is the permissible deviation or the maximum error to be expected from a manufactured component and expressed usually in measurement units ( example are psi, volts, meter, etc.).
As per JCGM 106:2012:
Tolerance = difference of upper and lower tolerance limits
Tolerance limit = specified upper or lower bound of permissible values of a property
Tolerance Interval = Intervals of permissible values
A pressure gauge with a full scale reading of 20 psi with a tolerance limit of +/- 0.5 ( 20 +/-0.5) psi has a tolerance of 1 psi, see below photo.
After calibration, we should perform a verification to determine if the reading is within the tolerance specified, it may be less accurate but if it is within the specified tolerance limit, it is still acceptable. This is where a pass or fail decision can be brought out.
The tolerance needed should be determined by the user of the UUC, which are the combinations of many factors which includes:
- Process requirements
- The capability of measurement equipment
- Manufacturer’s tolerance specifications (Related to accuracy)
- Published Standards
See more explanation and presentation in this link >> Differences Between Accuracy, Error, Tolerance, and Uncertainty in a Calibration Results
What is Precision in Measurement?
Precision is the closeness of a repeated measurement to each other. Precision signifies good stability and repeatability of instruments but not accuracy.
A measuring instrument can be highly precise but cannot be accurate. Our goal during calibration for a measuring instrument is to have a good accuracy and precision. Precision can be determine without using a reference standard.
How to Determine Precision in Measurements?
Precision is closely related to repeatability, you cannot determine precision of you cannot have a repeated measurements.
In order to calculate and determine precision, follow below steps:
1. With the same method of measurement, get a repeated reading. For example, a digital caliper measuring a 10 mm gauge block (or any stable material) for 10 times. See below table.
2. using the excel worksheet (this will simplify and make the calculation easy) plot all the measured points and calculate the standard deviation. Follow below instructions to calculate in excel.
The smaller the value, the more precise the instrument.
Below are the relationships of Accuracy, Precision, and Tolerance to understand better:
What is Stability and How to Determine Stability?
Stability is the ability of the instrument to maintain its output within a defined limits over a period of time. A highly stable instruments can be determine by collecting its output data in a fixed interval.
There are many ways to do it but I will share the most basic and simple to do. Just remember that the goal for stability calculation is to determine that the instrument or standard is functioning within specifications on a defined period of time.
Use this formula:
To determine stability, refer to the control chart below (or the error part on the table). Observe the peak to peak value (the highest and the lowest value in the control chart below) then perform below calculation.
Highest positive error = 0.2
Highest negative error = -0.3
Stability = [0.2 -(-0.3)]/2 = 0.25
Therefore, stability = 0.25 psi
Another method is by determining the standard deviation. If you are using excel sheet, this is the simplest to calculate.
Basing it on the table below, just use the formula =STDEV() (see also excel calculation of precision above) then highlight or choose the ‘UUC Actual Value’. A standard deviation equal to ‘0.192’ will be calculated as the ‘stability’.
How to Monitor Stability?
Below are the things that you can do to monitor stability of your reference standard.
1. Collect all the past calibration certificates of your reference standards to be evaluated, the more the better. Check the data results, or..
2. Stability is also determined by collecting data on a fixed interval. This could be a data from your intermediate check. In can be daily, monthly or every 3 months.
3. Organize your data on a table, see below example.
I have collected data through an intermediate check performed every three months on a Test Gauge.
Below are the data.
Determine the error, calculate the mean, and the standard deviation
Plot in the control chart.
As long as the error is within the control limit, we can be sure that the reference standard is very stable and in control.
Traceability in Calibration
As it was stated, calibration is the comparison of an instrument to a higher or more accurate instrument. These higher accuracy instruments are called the reference standard and sometimes also known as the calibrator, a master or a reference.
This reference standard that is used to calibrate your instruments has also a more accurate master being used to calibrate it and so on and so forth until it reaches the main source of that certain highest accuracy (measurement result).
There is an unbroken chain of comparison being linked from top to the bottom of the chain. It is passed to local from international standards or the BIPM (Bureau International des Poids et Measures (International Bureau of Weights & Measures), the top most source of traceability in the comparison chain ( as shown in the figure above.
This means that the 1 kilogram you used is also 1 kilogram no matter where you go. There is unity to every measurement. Traceability can be determined through its calibration certificate indicating the results and reference standard used to calibrate your instrument.
Why is Traceability in Measurements Necessary?
- For companies engaged in manufacturing and engineering, ensures that parts produce or supplied have the same or acceptable specifications when used by customers anywhere. Compatibility is not an issue.
- Traceability provides confidence to our measurement process because the validity of the measurement results are ensured for its accuracy.
- Traceability has a value, this value can be seen in a calibration certificate as the measurement uncertainty results. With this results, you can determine how accurate the measurement intruments are.
- A requirement by relevant laws and regulations to guarantee product quality.
- A requirement from a contract agreed by two parties (contractual provisions)- a traceable calibration
- Statutory requirements for safety – even though we have different units of measurements, we are confident that compatibility in terms of size, form or level is not an issue anywhere it goes.
Where can we find the traceability information of a Calibrated Instrument?
Since traceability is very important, we should have knowledge on how to determine or check the traceability information of certain calibrated instruments. We can find it through its calibration certificate, this is one of the check items in a calibration certificate once it is calibrated by an authorized laboratory.
In a calibration certificate, there is traceability information written usually in the middle part together with the reference standard used or at the bottom and or even both.
This is a requirement so it must not be neglected if the laboratory is a competent one. Moreover, the most important one is the result of its measurement uncertainty, it should be reflected with the data results to ensure that you have a traceable calibration done by an accredited laboratory.
To see more examples and evidence of traceability that you need to know, please visit my other post in this link.
What are the Differences Between Calibration, Verification, and Validation?
Calibration, verification, and validation are the terms that are most confusing if you are not aware of their differences and true meaning when it comes to the measurement process.
To differentiate these terms, below are the main points to remember:
- Calibration is simply the “comparison” of the unknown reading of a UUC to a known reading of a Reference Standard, also known as the Master.
- Verification is a process of “confirming” that a given specification is fulfilled.
- Validation is for “ensuring” the acceptability of the implemented measurement process. Focusing on the final output of the measurement process.
To learn more about their differences including a concrete example, visit my other post in this link>>
What is Measurement Uncertainty?
Calibration is not complete without Measurement uncertainty or Uncertainty of Measurement. This is where the unbroken chain of comparison is being connected or linked in.
Measurement Uncertainty is the value being displayed to quantify the doubt that exists on a specified measurement result. Since no measurement is exact, there is always an error that is associated with every measurement. And to determine the degree, effect, or quantity of this error that exists in every measured parameter, we compute or estimate Measurement Uncertainty.
During uncertainty computation or estimation, we identify all the valid sources of errors that have an influence our measurement system. It can be from our procedures, instruments, environment and many more. We evaluate and quantify the value of each error and combine them to a single computed value.
Where Can we find measurement uncertainty? We can find measurement uncertainty results in the calibration certificates usually in the data results page.
Visit this link from WIKIPEDIA to learn more.
Difference Between Measurement Uncertainty and Tolerance
Measurement Uncertainty (MU)
Usually defined as the quantification of the doubt. If you measure something, there is always an error (a doubt – no result is perfect) included in the final result since there are no exact measurement results.
Since there are no exact measurement results, what we can do is to determine the range where the true value is located, this range can be determined by adding or subtracting the limits of uncertainty (or the measurement uncertainty result) to the measurement result.
We do not know the true value is, but because of the measurement uncertainty result, it will show to us that the true value lies within the limits of the calculated measurement uncertainty.
“The smaller the measurement uncertainty, the more accurate or exact our measurement results.”
For example, based on calibration certificate of a pressure switch, it has a measurement result of 10 psi with a calculated measurement uncertainty of +/- 0.3 psi.
As we can see, it has an exact value of 10 psi, but in reality, the exact value is located between the range ‘9.7 to 10.3 psi’.
It is the maximum error or deviation that is allowed or acceptable as per the design of the user for its manufactured product or components.
If we perform a measurement, tolerance value will tell us if the measurement we have is acceptable or not.
If you know the tolerance, it will help you answer the questions like:
1. How do you know that your measurement result is within the acceptable range?
2. Is the final product specification pass or fail?
3. Do we need to perform adjustments?
“The bigger the tolerance, the more product or measurement results will pass or accepted.”
For example, a pressure switch is set to turn on at 10 psi. The process tolerance limit is 1 psi. Therefore, the acceptable range for the switch to turn on is between 9 to 11 psi, beyond this range, we need to perform adjustment and calibration.
See more explanation and presentation in this link >> Differences Between Accuracy, Error, Tolerance, and Uncertainty in a Calibration Results
ISO 17025 – Calibration Laboratory Quality Management System
What is ISO 17025?
- International Standard ISO/IEC 17025:2017, General requirements for the competence of testing and calibration laboratories
- As the title implies, it is a standard for laboratory competence, to differ from ISO 9001, which is for certification.
- It is an accreditation standard used by accreditation bodies where a demonstration of a calibration laboratory competency is assessed with regards to its scope and capabilities. Accreditation is simply the formal recognition of a demonstration of that competence.
- It is also a Quality Management System that is comparable to ISO 9001 but it is designed for Calibration Laboratory specifically 3rd Party or External Calibration Labs.
- The usual contents of the quality manual follow the outline of the ISO/IEC 17025 standard.
- It can be divided into two principal parts,
1. Management System Requirements- similar to those specified in ISO 9001:2015, primarily related to the operation and effectiveness of the quality management system within the laboratory.
2. Technical Requirements – include factors which determine the correctness and reliability of the tests and calibrations performed in the laboratory.
- These requirements are implementing systems and procedures to be met by testing and calibration labs in their organization and management of their quality system particularly when seeking accreditation.
- Since ISO 17025 is a quality management system specifically for calibration laboratories, it is also a good tool or a guide if you are managing an in-house or internal lab. Following its requirements will help you achieve most of the auditors’ requirements during internal or customer audits.
To learn more about the requirements of ISO 17025:2017, visit my other post in the below link:
ISO/IEC 17025:2017 Requirements: List of Documents Outline and Summary
Learn the basic elements regarding In-House Calibration Management Implementation by visiting this link. >>> ELEMENTS IN IMPLEMENTING IN-HOUSE CALIBRATION.
Thank you for visiting, please subscribe and share
You can also connect with me on my Facebook page.