Glossary

This glossary includes terms and definitions for IBM® Maximo® Calibration.

The following cross-references are used in this glossary:
  • See refers you from a term to a preferred synonym, or from an acronym or abbreviation to the defined full form.
  • See also refers you to a related or contrasting term.

To view glossaries for other IBM products, go to www.ibm.com/software/globalization/terminology.

analog device

An instrument that provides continuous results, such as a pressure gauge.

as-found value

The calibration value of an instrument before the instrument is adjusted.

as-left value

The calibration value of an instrument after the instrument is adjusted.

asset function data

The calibration data for one or more functions that an asset performs.

asset template

A record that specifies asset information that is shared by multiple asset records. An asset template can be applied to existing asset records or it can be used to generate multiple new asset records for similar assets, such as a fleet of vehicles.

calibration

The comparison and adjustment of an instrument to a standard of known accuracy.

calibration drift

The difference between the as left value of the previous measurement compared to the as found value of the current measurement for an asset.

calibration point

The calibration requirement for each asset function.

data sheet

A template that defines the specifications of an asset for the calibration process.

deadband

The difference between the set point and reset point of a discrete device.

desired output

The output that is required for a given input. For example, 0 psi input = 4 mA output.

direction

An indicator of whether a specified calibration point is increasing or decreasing with respect to the reference point.

discrete device

An instrument that has only two settings, such as, a switch that is either on or off or a valve that is either open or closed.

field resolution

The minimum number of required decimal places for a numeric field value.

guardband method

A method that calculates the reliability of the calibration of an asset to ensure that the asset meets a specified level of quality.

input

The process variable that is applied to a device under test.

input range

The upper and lower input values of an asset that is being calibrated.

instrument

A device, such as a gauge, transmitter, or switch, that is used to monitor and control plant processes.

loop calibration

A process that uses a set of grouped and calibrated instruments to determine the pass or fail status of the calibration.

measuring and test equipment (M&TE)

A calibrating tool that has a higher degree of accuracy than the asset being calibrated.

nominal input

The user-specified input value for each analog calibration point.

output

A reading of the input for a device under test.

output range

The upper and lower output values of an asset that is being calibrated.

range-on-nominal

The user-specified acceptable limits for input values for each calibration point.

reverse traceability

The ability to identify which assets were calibrated by particular measuring and test equipment.

scaling factor

The conversion factor from output engineering units to user-specified units.

set point

The input value at which a discrete device activates.

set point action

The wanted behavior of the discrete device, such as a switch, at the specified set point, such as on or off.

tolerance

An acceptable value or value range for a specified output value.

traceability

The ability to identify the measuring and test equipment that was used to calibrate an asset.