Definitions

# Ohmmeter

[ohm-mee-ter]
An ohmmeter is an electrical instrument that measures electrical resistance, the opposition to an electric current. Micro-ohmmeters (microhmmeter or microohmmeter) make low resistance measurements. Megohmmeters (aka megaohmmeter or megger) measure large values of resistance.

The original design of an ohmmeter provided a small battery to apply a voltage to a resistance. It used a galvanometer to measure the electric current through the resistance. The scale of the galvanometer was marked in ohms, because the fixed voltage from the battery assured that as resistance decreased, the current through the meter would increase.

A more accurate type of ohmmeter has an electronic circuit that passes a constant current (I) through the resistance, and another circuit that measures the voltage (V) across the resistance. According to the following equation, derived from Ohm's Law, the value of the resistance (R) is given by:

$R = frac\left\{V\right\}\left\{I\right\}$

For high-precision measurements the above types of meter are inadequate. This is because the meter's reading is the sum of the resistance of the measuring leads, the contact resistances and the resistance being measured. To reduce this effect, a precision ohmmeter has four terminals, called Kelvin contacts. Two terminals carry the current from the meter, while the other two allow the meter to measure the voltage across the resistor. With this type of meter, any current drop due to the resistance of the first pair of leads and their contact resistances is ignored by the meter. This four terminal measurement technique is called Kelvin sensing, after William Thomson, Lord Kelvin, who invented the Kelvin bridge in 1861 to measure very low resistances. Four-point-probe can also be ultilized to conduct accurate measurements of low resistances.