Part 3. Calibration of Geotechnical Instruments (Including Case Study)

Overview: This new 5 part series has been written to explain the hardware, software and instrumentation used in the testing of soil and rock. The series comprises of 5 chapters (see below). The series is aimed at people interested in gaining a better understanding of geotechnical laboratory equipment.

5. Principles of Testing Machine Control Feedback.

Read below or download the PDF file here

INTRODUCTION

As we have seen, the accuracy of transducers is transferred to them by the process of calibration against (or comparison with) some standard. The standard will itself have a specified accuracy. Doebelin (2003) points out that the standard must have an accuracy higher than the accuracy of the instrument being calibrated, and that the standard itself must have been established by acceptable means. Sydenham et al. (1989) make the important point that “… Associated with the calibration are two costs – that of making it and that of not making it”.

It is evident that the calibration means of transferring the accuracy must encompass the whole measurement system involved and will be affected by resolution, stability and repeatability – and above all be part of the laboratory culture.

It is desirable that transducers have a near-linear relationship between the set standard and the measured quantity so that the relationship can be expressed as a parameter in 
engineering units per transducer output which is usually in mV e.g. kPa/mV. The variability in this relationship can be expressed as an accuracy or linearity such as say 0.1%. But it is not always possible, and indeed with modern software, not always necessary for linearity to be achieved.

There follows three examples of calibration methods. The use of the dead-weight tester for calibrating pressure transducers, pressure sources and load cells is probably common to most large soil mechanics laboratories. The use of laser interferometry to calibrate LVDTs (Linear Variable Displacement Transformers) and measure small strains on triaxial test specimens of stiff soils and soft rocks, however, will be very much a glimpse of the future of advanced soil testing. There also follows an example of how machines can be calibrated too and how this can be important for loading frames, particularly when testing hard soils and soft rocks. Finally we comment on the importance of verification.
Budenberg Dead-weight Tester

The calibration of pressure transducers, pressure sources and load cells can be carried out using the Budenberg dead-weight tester. This is shown in Fig. 3. The principle of operation is a screw pump that pressurises low viscosity oil. Oil can be drawn into the screw pump via a set of valves from an oil reservoir. The pressurised oil causes a vertical piston to float up through an open vertical cylinder. The top of the piston where it emerges from the piston is provided with a platen that fits into an interlocking vertical stack of nested plate-like weights. The stack of plates and the piston together are lifted by operation of the screw pump so that they float above the rest position. In this floating state, the weights can be slowly set spinning by hand or by an electric motor. Oil slowly flows past the piston and into an overflow cup. The operation of the piston is therefore virtually frictionless much like a rotating bushing except here the piston rotates and the bushing or cylinder is stationary. Stops come into action if the pressure is too high or too low and it is essential that the weights should be spinning freely when taking readings.


Fig. 3 Budenberg dead-weight tester showing stack of standard calibrating weights, oil-water change-over pot, and load cell calibrating frame (permission of DH-Budenberg Ltd).


Fig. 4 Calibration of “slave” pressure controllers from a production batch by a “master” pressure controller calibrated by the Budenberg dead-weight tester.

The masses of the weights together with the mass of the piston and the piston area are so arranged as to give a number of increments in the calibrating pressure. The pressure datum of the tester is at the base of the piston and is marked on the piston units. The tester is certified by the manufacturer to be balanced against an assembly calibrated
 by the National Physical Laboratory whence they certify that the error of the tester when used at standard temperature of 20°C ± 1°C does not exceed 0.05% of the pressure being measured. The manufacturer provides equations and software for correcting for variations in temperature (e.g. deducting 0.0027% from the nominal pressure for each 1°C rise in temperature and adding correspondingly for each 1°C fall) and gravity (e.g. the calibration for the tester is only exact at a place where the acceleration due to gravity is 9.80665m/s2).

For example, the set-up for calibrating pressure sources such as pressure controllers is shown in Fig. 4. The oil outlet of the tester is connected to an oil-water change-over pot made of clear acrylic polymer so the oil-water interface can be seen. This enables the pressure connection to the controllers to be filled with water. At GDS, when calibrating a production run of say 20 controllers, one of the controllers is selected that has a very high linearity on-board pressure transducer (pressure transducers are typically specified by the manufacturer as having a minimum linearity of 0.5% but when supplied in batches will typically range from 0.4% to 0.1%). This is designated the “master” controller and is calibrated using the dead weight tester. First, the master controller “hard” zero is set in hardware (by adjusting a small potentiometer on the controller printed circuit board) against the zero of the dead-weight tester by holding a water-filled open tube that is connected to the controller so the open end of the water-filled tube is at the same elevation as the tester zero. The tube is then connected into the outlet of the tester oil-water pot. The full range is set on the tester, say 2000kPa, and the “hard” range of the master controller is also set in hardware (again by adjusting a small potentiometer on the controller printed circuit board). The master is then calibrated by increasing the pressure of the tester in steps of 100kPa in the range from 0 to 2000kPa and in steps back to zero again. At each step, the pressure measured by the master controller is stored in computer and a calibration table is made in software. The master controller is then connected by small bore nylon tubing to a manifold connecting the remaining 19 “slave” controllers of the production batch. A water-filled open tube is connected to the manifold, the open end held at the elevation of the dead-weight tester zero, and the slave controllers are also hard-zeroed. The open tube of water is then disconnected from the manifold and the master controller is set to full range, say 2000kPa, and the hard ranges set on the slaves, by adjusting another small potentiometer on the controller printed circuit board. The calibration software is then run and the computer calibrates the slave controllers against the master controller by setting target pressures on the master and measuring the corresponding pressures for each slave and thus producing a calibration certificate with a table of readings and linear regression analyses for each controller of the batch.

Calibration of machines – Virtual Infinite Stiffness (VIS) 

Machines can be calibrated too. This can very useful to correct machines or machine generated readings for machine compliance. At GDS this has lead to the development of Virtual Infinite Stiffness (VIS) applied to loading systems. To the observer, and in terms of the test specimen, it allows the axial loading system to appear to have infinite stiffness.

Loading frames have a display of platen displacement. This is usually derived from the loading system mechanical characteristics such as counting steps for the driving stepping motor. This does not take into account the machine compliance that includes strain in the load cell and side columns, bending flexure of the cross beams, and distortion within the motorised mechanical transmission. For testing hard soils and soft rocks (and particularly for medium to hard rocks) the machine itself can have a stiffness comparable to the material being tested – the so-called “soft machine”. This leads to the machine overestimating the displacement that occurs between the load cell and machine platen i.e. within the domain of the test cell.

To overcome this, for the entire loading range, both the measurement and control of platen displacement
is automatically corrected so that it corresponds to the deformation that occurs between the platen and the load button of the load cell. In this way, the platen displacement is corrected for machine compliance. These measurements are made with the adjustable upper cross beam in the maximum and minimum positions. For each position, measurements are made with the platen at each end of its travel.

The calibration data is loaded into the read only memory (ROM) of the system which constantly monitors the axial load and uses the calibration to apply a correction to the platen displacement. Therefore, it appears to the observer (or controlling computer) that the measurement of platen displacement (resolved to 0.1μ) is derived from a machine with infinite stiffness. In this way the system has the characteristic of “Virtual Infinite Stiffness”. Of course, the loading frame (or loading assembly in the case of a force actuator used in direct shear tests for example) is not infinitely stiff – it only appears to be infinitely stiff by having the machine displacement corrected for machine compliance both on the display and at the computer interface.

Verification

Following the calibrations of all test equipment and transducers (and machines) using the measurement systems of the test, it might seem that the testing programme can begin. Indeed it can. But what if on subsequent tests and before the next calendar date of re-calibration one of the transducers malfunctions? With modern transducers monitored by computer logging with on-line test data presentation, a malfunctioning transducer might not be immediately obvious or it might be misunderstood, for example, as anomalous soil behaviour. This is why verification is important. For example, a GDS digital pressure controller can be easily provided with a Bourdon tube type mechanical pressure gauge that can readily be checked during the test against the display of pressure on the controller and on the computer screen and GDS can supply their controllers with pressure gauges (for example in Hong Kong – see HOKLAS requirements below). While the performance of load cells and displacement transducers, however, cannot usually be checked during a test, they can be verified before each test. A displacement transducer can easily be moved by either a set amount (say by inserting a small block of known size to cause a displacement of the armature) or by an approximate amount “by eye” (say ≈10mm) and the corresponding movement recorded by the logger can be compared to verify that the readings make sense (this can only confirm that the transducer does or does not work – not that the calibration is OK). With a load cell it is a little more complicated and less precise. A rough check can be made by extracting the submersible load cell and ram from a triaxial cell, or by removing the external load cell from a triaxial loading frame say, and loading it manually by pushing it against the floor by hand. About half one’s weight or more can easily be applied in this way corresponding to an approximate force somewhere in the range of about half to one kN. By noting the output of the load cell it can soon be seen if the reading is about right i.e. a rough verification has been made (which is better than none at all). Of course if the laboratory has a Budenberg dead-weight tester and load cell calibration rig. In-house verification (and calibration) means and procedures are a 
necessary requirement for laboratory accreditation. In the UK there is the United Kingdom Accreditation Service (UKAS). UKAS is a member of the European co-operation for Accreditation (EA), the International Accreditation Forum (IAF) and the International Laboratory Accreditation Cooperation (ILAC). A search on the UKAS web site under “construction/soils and stabilised soils/effective shear strength tests” returns seven company names.

The Hong Kong Laboratory Accreditation Scheme (HOKLAS) publish detailed guidelines. For example, in their on-line publication “HOKLAS Supplementary Criteria No.18. Calibration/verification procedure or guidance documents and equipment requirements. General soil and rock tests1” they state clearly (and most helpfully): 

“Pressure transducers – transducer types (for triaxial testing only). 

Check against calibrated values of the pressure gauge Bourdon tube type at three points, one at the middle of the range, one at the lower (not less than 200kPa) and one at the uppermost range. Carry out a full calibration if the drift of the gauge exceeds the test requirement.” 

In their on-line publication on measurement uncertainty1 (or error), UKAS make the very important point that uncertainty of measurement has particular implications for specification, regulation and simply for comparison of test results from different laboratories. They make the point that: 

“Uncertainty is an unavoidable part of any measurement and it starts to matter when results are close to a specified limit. A proper evaluation of uncertainty is good professional practice and can provide laboratories and customers with valuable information about the quality and reliability of the result. Although common practice in calibration, there is some way to go with expression of uncertainty in testing, but there is growing activity in the area and, in time, uncertainty statements will be the norm.”
Reference:
Doebelin, E. O. (1983). Measurement systems – application and design. McGraw-Hill, 1078p.

Sydenham, P. H., Hancock, N. H. and Thorn, R. (1989). Introduction to measurement science and engineering. Chichester: Wiley, 327p.