From Wikipedia, the free encyclopedia - View original article
|This article needs additional citations for verification. (April 2008)|
Image resolution can be measured in various ways. Basically, resolution quantifies how close lines can be to each other and still be visibly resolved. Resolution units can be tied to physical sizes (e.g. lines per mm, lines per inch), to the overall size of a picture (lines per picture height, also known simply as lines, TV lines, or TVL), or to angular subtenant. Line pairs are often used instead of lines; a line pair comprises a dark line and an adjacent light line. A line is either a dark line or a light line. A resolution 10 lines per millimeter means 5 dark lines alternating with 5 light lines, or 5 line pairs per millimeter (5 LP/mm). Photographic lens and film resolution are most often quoted in line pairs per millimeter.
The resolution of digital cameras can be described in many different ways.
Resolution is the capability of sensor to observe or measure the smallest object clearly with distinct boundaries. There is difference between resolution and pixel. Pixel is actually a unit of digital image. Resolution depends upon the size of pixel. Smaller the size of pixel, higher will be the resolution and more clear will the object in image. Image having smaller pixel size occupy more size on disk.
The term resolution is often used for a pixel count in digital imaging, even though American, Japanese, and international standards specify that it should not be so used, at least in the digital camera field.
An image of N pixels high by M pixels wide can have any resolution less than N lines per picture height, or N TV lines. But when the pixel counts are referred to as resolution, the convention is to describe the pixel resolution with the set of two positive integer numbers, where the first number is the number of pixel columns (width) and the second is the number of pixel rows (height), for example as 7680 by 6876. Another popular convention is to cite resolution as the total number of pixels in the image, typically given as number of megapixels, which can be calculated by multiplying pixel columns by pixel rows and dividing by one million. Other conventions include describing pixels per length unit or pixels per area unit, such as pixels per inch or per square inch. None of these pixel resolutions are true resolutions, but they are widely referred to as such; they serve as upper bounds on image resolution.
According to the same standards, the number of effective pixels that an image sensor or digital camera has is the count of elementary pixel sensors[clarification needed] that contribute to the final image, as opposed to the number of total pixels, which includes unused or light-shielded pixels around the edges.
Below is an illustration of how the same image might appear at different pixel resolutions, if the pixels were poorly rendered as sharp squares (normally, a smooth image reconstruction from pixels would be preferred, but for illustration of pixels, the sharp squares make the point better).
An image that is 2048 pixels in width and 1536 pixels in height has a total of 2048×1536 = 3,145,728 pixels or 3.1 megapixels. One could refer to it as 2048 by 1536 or a 3.1-megapixel image.
Unfortunately, the count of pixels isn't a real measure of the resolution of digital camera images, because color image sensors are typically set up to alternate color filter types over the light sensitive individual pixel sensors. Digital images ultimately require a red, green, and blue value for each pixel to be displayed or printed, but one individual pixel in the image sensor will only supply one of those three pieces of information. The image has to be interpolated or demosaiced to produce all three colors for each output pixel.
The measure of how closely lines can be resolved in an image is called spatial resolution, and it depends on properties of the system creating the image, not just the pixel resolution in pixels per inch (ppi). For practical purposes the clarity of the image is decided by its spatial resolution, not the number of pixels in an image. In effect, spatial resolution refers to the number of independent pixel values per unit length.
The spatial resolution of computer monitors is generally 72 to 100 lines per inch, corresponding to pixel resolutions of 72 to 100 ppi. With scanners, optical resolution is sometimes used to distinguish spatial resolution from the number of pixels per inch.
In remote sensing, spatial resolution is typically limited by diffraction, as well as by aberrations, imperfect focus, and atmospheric distortion. The ground sample distance (GSD) of an image, the pixel spacing on the Earth's surface, is typically considerably smaller than the resolvable spot size.
In astronomy, one often measures spatial resolution in data points per arcsecond subtended at the point of observation, since the physical distance between objects in the image depends on their distance away and this varies widely with the object of interest. On the other hand, in electron microscopy, line or fringe resolution refers to the minimum separation detectable between adjacent parallel lines (e.g. between planes of atoms), while point resolution instead refers to the minimum separation between adjacent points that can be both detected and interpreted e.g. as adjacent columns of atoms, for instance. The former often helps one detect periodicity in specimens, while the latter (although more difficult to achieve) is key to visualizing how individual atoms interact.
In Stereoscopic 3D images, spatial resolution could be defined as the spatial information recorded or captured by two viewpoints of a stereo camera (left and right camera). The effects of spatial resolution on overall perceived resolution of an image on a person's mind are yet not fully documented. It could be argued that such "spatial resolution" could add an image that then would not depend solely on pixel count or Dots per inch alone, when classifying and interpreting overall resolution of a given photographic image or video frame.
Color images distinguish light of different spectra. Multispectral images resolve even finer differences of spectrum or wavelength than is needed to reproduce color. That is, multispectral images have higher spectral resolution than normal color images.
Movie cameras and high-speed cameras can resolve events at different points in time. The time resolution used for movies is usually 24 to 48 frames per second (frames/s), while high-speed cameras may resolve 50 to 300 frames/s, or even more.
Many cameras and displays offset the color components relative to each other or mix up temporal with spatial resolution:
CRT (shadow mask)
Radiometric resolution determines how finely a system can represent or distinguish differences of intensity, and is usually expressed as a number of levels or a number of bits, for example 8 bits or 256 levels that is typical of computer image files. The higher the radiometric resolution, the better subtle differences of intensity or reflectivity can be represented, at least in theory. In practice, the effective radiometric resolution is typically limited by the noise level, rather than by the number of bits of representation.
This is a list of traditional, analog horizontal resolutions for various media. The list only includes popular formats, not rare formats, and all values are approximate (rounded to the nearest 10), since the actual quality can vary machine-to-machine or tape-to-tape. For ease-of-comparison, all values are for the NTSC system. (For PAL systems, replace 480 with 576.)