From Wikipedia, the free encyclopedia  View original article
Ohm  

A laboratory oneOhm standard resistor.  
Unit information  
Unit system  SI derived unit 
Unit of  Electrical resistance 
Symbol  Ω 
Named after  Georg Ohm 
In SI base units:  1 Ω = 1 kg·m^{2}·s^{3}·A^{2} 
Ohm  

A laboratory oneOhm standard resistor.  
Unit information  
Unit system  SI derived unit 
Unit of  Electrical resistance 
Symbol  Ω 
Named after  Georg Ohm 
In SI base units:  1 Ω = 1 kg·m^{2}·s^{3}·A^{2} 
The ohm (symbol: Ω) is the SI derived unit of electrical resistance, named after German physicist Georg Simon Ohm. Although several empirically derived standard units for expressing electrical resistance were developed in connection with early telegraphy practice, the British Association for the Advancement of Science proposed a unit derived from existing units of mass, length and time and of a convenient size for practical work as early as 1861. The definition of the "ohm" unit was revised several times. Today the value of the ohm is expressed in terms of the quantum Hall effect.
The ohm is defined as a resistance between two points of a conductor when a constant potential difference of 1.0 volt, applied to these points, produces in the conductor a current of 1.0 ampere, the conductor not being the seat of any electromotive force.^{[1]}
where:
In many cases the resistance of a conductor in ohms is approximately constant within a certain range of voltages, temperatures, and other parameters; one speaks of linear resistors. In other cases resistance varies (e.g., thermistors).
Commonly used multiples and submultiples in electrical and electronic usage are the microohm, milliohm, kilohm, megohm, and gigaohm,^{[2]} the term "gigohm", though not official, is in common use for the latter.^{[3]}
In alternating current circuits, electrical impedance is also measured in ohms.
The siemens (symbol: S) is the SI derived unit of electric conductance and admittance, also known as the mho (ohm spelled backwards, symbol is ℧); it is the reciprocal of resistance in ohms.
The power dissipated by a resistor may be calculated from its resistance, and the voltage or current involved. The formula is a combination of Ohm's Law and Joule's law:
where:
A linear resistor has a constant resistance value over all applied voltages or currents; many practical resistors are linear over a useful range of currents. Nonlinear resistors have a value that may vary depending on the applied voltage (or current). Where alternating current is applied to the circuit,(or where the resistance value is a function of time), the relation above is true at any instant but calculation of average power over an interval of time will require integration of "instantaneous" power over that interval.
The rapid rise of electrotechnology in the last half of the 19th century created a demand for a rational, coherent, consistent, and international system of units for electrical quantities. Telegraphers and other early uses of electricity in the 19th century needed a practical standard unit of measurement for resistance. Resistance was often expressed as a multiple of the resistance of a standard length of telegraph wires; different agencies used different bases for a standard, so units were not readily interchangeable. Electrical units so defined were not a coherent system with the units for energy, mass, length, and time, requiring conversion factors to be used in calculations relating energy or power to resistance.^{[4]}
Two different methods of establishing a system of electrical units can be chosen. Various artifacts, such as a length of wire or a standard electrochemical cell, could be specified as producing defined quantities for resistance, voltage, and so on. Alternatively, the electrical units can be related to the mechanical units by defining, for example, a unit of current that gives a specified force between two wires, or a unit of charge that gives a unit of force between two unit charges. This latter method ensures coherence with the units of energy. Defining a unit for resistance that is coherent with units of energy and time in effect also requires defining units for potential and current. It is desirable that one unit of electrical potential will force one unit of electrical current through one unit of electrical resistance, doing one unit of work in one unit of time, otherwise all electrical calculations will require conversion factors.
Since socalled "absolute" units of charge and current are expressed as combinations of units of mass, length, and time, dimensional analysis of the relations between potential, current and resistance show that resistance is expressed in units of length per time  a velocity. Some early definitions of a unit of resistance, for example, defined a unit resistance as one quadrant of the Earth per second.
"Absolute" units system related magnetic and electrostatic quantities to metric base units of mass, time, and length. These units had the great advantage of simplifying the equations used in the solution of electromagnetic problems, and eliminated conversion factors in calculations about electrical quantities. However, the CGS units turned out to have impractical sizes for practical measurements.
Various artifact standards were proposed as the definition of the unit of resistance. In 1860 Werner Siemens published a suggestion for a reproducible resistance standard in Poggendorffs Annalen der Physik und Chemie.^{[5]} He proposed a column of pure mercury, of one square millimetre cross section, one metre long. However, this unit was not coherent with other units. One proposal was to devise a unit based on a mercury column that would be coherent  in effect, adjusting the length to make the resistance one ohm. Not all users of units had the resources to carry out metrology experiments to the required precision, so working standards notionally based on the physical definition were required.
In 1861, Latimer Clark and Sir Charles Bright presented a paper at the British Association for the Advancement of Science meeting ^{[6]} suggesting that standards for electrical units be established and suggesting names for these units derived from eminent philosophers, 'Ohma', 'Farad' and 'Volt'. The BAAS in 1861 appointed a committee including Maxwell and Thomson to report upon Standards of Electrical Resistance.^{[7]} Their objectives were to devise a unit that was of convenient size, part of a complete system for electrical measurements, coherent with the units for energy, stable, reproducible and based on the French metrical system.^{[8]} In the third report of the committee, 1864, the resistance unit is referred to as "B.A. unit, or Ohmad".^{[9]} By 1867 the unit is referred to as simply Ohm.^{[10]}
The B.A. ohm was intended to be 10^{9} CGS units but owing to an error in calculations the definition was 1.3% too small. The error was significant for preparation of working standards.
On September 21, 1881 the Congrès internationale d'électriciens (international conference of electricians) defined a practical unit of Ohm for the resistance, based on CGS units, using a mercury column at zero deg. Celsius, similar to the apparatus suggested by Siemens.
A legal ohm, a reproducible standard, was defined by the international conference of electricians at Paris in 1884 as the resistance of a mercury column of specified weight and 106 cm long; this was a compromise value between the B. A. unit (equivalent to 104.7 cm), the Siemens unit (100 cm by definition), and the CGS unit. Although called "legal", this standard was not adopted by any national legislation. The "international" ohm was defined as a mercury column 106.3 cm long of mass 14.4521 grams and 0 C at the International Electrical Conference 1893 in Chicago. This definition became the basis for the legal definition of the ohm in several countries. In 1908, the next Electrical Conference confirmed this definition. The mercury column standard was maintained until the 1948 General Conference on Weights and Measures, at which the ohm was redefined in absolute terms instead of as an artifact standard.
By the end of the 19th century, units were wellunderstood and consistent. Definitions would change with little effect on commercial uses of the units. Advances in metrology allowed definitions to be formulated with a high degree of precision and repeatability.
Unit^{[11]}  Definition  Value in B.A. ohms  Remarks 

Absolute foot/second x 10^{7}  using Imperial units  .3048  considered obsolete even in 1884 
Thomson's unit  using Imperial units  .3202  100 million feet/second, considered obsolete even in 1884 
Jacobi  A specified copper wire 25 feet long weighing 345 grains  .6367  
Weber's absolute unit × 10^{7}  Based on the metre and the second  0.9191  
Siemens mercury unit  column of pure mercury  .9537  100 cm and 1 mm² cross section at 0 °C 
British Association (B.A.) "ohm"  1.000  
Digney, Breguet, Swiss  9.266–10.420  Iron wire 1 km long and 4 square mm cross section  
Matthiessen  13.59  One mile of 1/16 inch diameter pure annealed copper wire at 15.5 °C  
Varley  25.61  One mile of special 1/16 inch diameter copper wire  
German mile  57.44  A German mile (8,238 yard) of iron wire 1/6th inch diameter  
Abohm  10^{−9}  Electromagnetic absolute unit in cmgramsecond units  
Statohm  Electrostatic absolute unit in cmgramsecond units 
The mercury column method of realizing a physical standard ohm turned out to be difficult to reproduce, owing to the effects of nonconstant cross section of the glass tubing. Various resistance coils were constructed by the British Association and others, to serve as physical artifact standards for the unit of resistance. The longterm stability and reproducibility of these artifacts was an ongoing field of research, as the effects of temperature, air pressure, humidity, and time on the standards were detected and analyzed.
Artifact standards are still used, but metrology experiments relating accuratelydimensioned inductors and capacitors provided a more fundamental basis for the definition of the ohm. Since 1990 the quantum Hall effect has been used to define the ohm with high precision and repeatability. The quantum Hall experiments are used to check the stability of working standards that have convenient values for comparison.^{[12]}
When preparing electronic documents, some document editing software will attempt to use the Symbol typeface to render the Ω character. Where the font is not supported, a W is displayed instead ("10 W" instead of "10 Ω", for instance). As W represents the watt, the SI unit of power, not resistance, this can lead to confusion.
An "R" can be used instead of the Ω symbol if it is not supported, thus, a 10 Ω resistor can also be represented as 10R. This is the British standard BS 1852 code. It is used in many instances where the value has a decimal place. For example, 5.6 Ω would be listed as 5R6. One advantage of this method is that it is relatively easy to "rub off" a decimal point symbol ".", changing the apparent value, compared to the "R" symbol, which would require more effort.
Unicode encodes the symbol as U+2126 Ω ohm sign, distinct from Greek omega among letterlike symbols, but it is only included for backwards compatibility and the Greek uppercase omega character U+03A9 Ω greek capital letter omega (HTML Ω
· Ω
) is preferred.^{[13]} In DOS and Windows, the alt code ALT 234 may produce the Ω symbol. In Mac OS, ⌥ Opt+Z does the same.
