Добавил:
Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:
литература / Digital_Video_and_HD_Second_Edition_Algorithms_and_Interfaces.pdf
Скачиваний:
0
Добавлен:
13.05.2026
Размер:
38.02 Mб
Скачать

Don’t confuse point spread function (PSF) with progressive segmented-frame (PsF), to be described on page 94.

Image structure

7

A naïve approach to digital imaging treats an image as a matrix of independent pixels, disregarding the spatial distribution of light power across each pixel. You might think that optimum image quality is obtained when there is no overlap between the distributions of neighboring pixels; many computer engineers hold this view. However, continuous-tone images are best reproduced with a certain degree of overlap between pixels; sharpness is reduced slightly, but pixel structure is made less visible and image quality is improved.

The distribution of intensity across a displayed pixel is referred to as its point spread function (PSF). A onedimensional slice through the center of a PSF is colloquially called a spot profile. A display’s PSF influences the nature of the images it reproduces. The effects of a PSF can be analyzed using filter theory, discussed for one dimension in the chapter Filtering and sampling, on page 191, and for two dimensions in Image digitization and reconstruction, on page 237.

Historically, the PSFs of greyscale (“black-and-white”) CRTs were roughly Gaussian in shape: Intensity distribution peaked at the center of the pixel, fell off over

a small distance, and overlapped neighboring pixels to some extent. The scanning spot of colour CRTs had this shape, too; but the PSF was influenced by the shadow mask or aperture grille. The introduction of direct-view colour CRTs shifted the requirement for spatial filtering to the viewer: The assumption was introduced that the viewers were sufficiently distant from the screens that the viewers’ visual systems would perform the spatial integration necessary to obscure the triad structure.

75

Figure 7.1“Box” reconstruction of a bitmapped graphic image is shown.

Figure 7.2 Gaussian reconstruction is shown for the same bitmapped image as Figure 7.1. I will detail the onedimensional AGaussian function on page 200.

Modern direct view fixed-pixel displays (FPDs) such as LCD and PDP displays have more or less uniform light emission over most of the area corresponding to each colour component (subpixel); their modulated light has a spatial structure comparable to that of

a direct-view colour CRT, and similarly depends upon the viewers being located at a sufficient distance that their visual characteristics perform the spatial intergation necessary to obscure the triad structure.

A pixel whose intensity distribution uniformly covers a small square area of the screen has a point spread function referred to as a “box.”

Image reconstruction

Figure 7.1 reproduces a portion of an idealized bitmapped (bilevel) graphic image, part of a computer’s desktop display. Each sample is either black or white. The element with horizontal “stripes” is part of a window’s titlebar; the checkerboard background is intended to integrate to grey. Figure 7.1 shows reconstruction of the image with a “box” distribution. Each pixel is uniformly shaded across its extent; there is no overlap between pixels. This figure exemplifies a rasterlocked image as displayed on an LCD. By raster-locked, I refer to image data having the underlying image elements aligned with the pixel array.

A CRT’s electron gun produces an electron beam that illuminates a spot on the phosphor screen. The beam is deflected to form a raster pattern of scan lines that traces the entire screen, as I will describe in the following chapter. The beam is not perfectly focused when it is emitted from the CRT’s electron gun, and is dispersed further in transit to the phosphor screen. The intensity produced for each pixel at the face of the screen has a “bell-shaped” distribution resembling

a two-dimensional Gaussian function. With a typical amount of spot overlap, the checkerboard area of this example will display as a nearly uniform grey as depicted in Figure 7.2. You might think that the blur caused by overlap between pixels would diminish image quality. However, for continuous-tone (“contone”) images, some degree of overlap is not only desirable but necessary, as you will see from the following examples.

76

DIGITAL VIDEO AND HD ALGORITHMS AND INTERFACES

Figure 7.3 Diagonal line reconstruction. At the left is a nearvertical line slightly more than 1 pixel wide, rendered as an array 20 pixels high that has been reconstructed using a box distribution. At the right, the line is reconstructed using a Gaussian distribution. Between the images I have placed a set of markers to indicate the vertical centers of the image rows.

Figure 7.4 Contone image reconstruction. At the left is a contin- uous-tone image of 16× 20 pixels that has been reconstructed using a box distribution. The pictured individual cannot be recognized. At the right is exactly the same image data, but reconstructed by a Gaussian function. The reconstructed image is very blurry but recognizable. Which reconstruction function do you think is best for continuous-tone imaging?

Visual acuity is detailed in

Contrast sensitivity function (CSF), on page 251.

Figure 7.3 shows a 16× 20-pixel image of a dark line slightly more than one pixel wide, 7.2° off the vertical. At the left, the image data is reconstructed using a box distribution; a jagged and “ropey” nature is evident. At the right, the image data is reconstructed using a Gaussian. It is blurry, but less jagged.

Figure 7.4 shows two ways to reconstruct the same 16× 20 pixels (320 bytes) of continuous-tone greyscale image data. The left-hand image is reconstructed using a box function, and the right-hand image with a Gaussian. The example was constructed so that each image is 4 cm (1.6 inches) wide. At typical reading distance of 40 cm (16 inches), a pixel subtends 0.4°, where visual acuity is near its maximum. At this distance, when reconstructed with a box function, the pixel structure of each image is highly visible; visibility of the pixel structure overwhelms the perception of the image itself. The right image is reconstructed using a Gaussian distribution. It is blurry, but easily recognizable as an American

CHAPTER 7

IMAGE STRUCTURE

77

Figure 7.5 One frame of an animated sequence, reconstructed with a “box” filter.

Figure 7.6 A Moiré pattern is a form of aliasing in two dimensions that results when a sampling pattern (here the perforated square) has a sampling density that is too low for the image content (here the dozen bars, 14° offvertical). This figure is adapted from Fig. 3.12 of Wandell’s Foundations of Vision (cited on page 195).

cultural icon. This example shows that sharpness is not always good, and blurriness is not always bad!

Figure 7.5 in the margin shows a 16× 20-pixel image comprising 20 copies of the top row of Figure 7.3 (left). Consider a sequence of 20 animated frames, where each frame is formed from successive image rows of Figure 7.3. The animation would depict a narrow vertical line drifting rightward across the screenat a rate of 1 pixel every 8 frames. If image rows of Figure 7.3 (left) were used, the width of the moving line would appear to jitter frame-to-frame, and the minimum lightness would vary. With Gaussian reconstruction, as in Figure 7.3 (right), motion portrayal is much smoother.

Sampling aperture

In a practical image sensor, each element acquires information from a finite region of the image plane; the value of each pixel is a function of the distribution of intensity over that region. The distribution of sensitivity across a pixel of an image capture device is referred to as its sampling aperture, sort of a PSF in reverse – you could call it a point “collection” function. The sampling aperture influences the nature of the image signal originated by a sensor. Sampling apertures used in continuous-tone imaging systems usually peak at the center of each pixel, fall off over a small distance, and overlap neighboring pixels to some extent.

In 1915, Harry Nyquist published a landmark paper stating that a sampled analog signal cannot be reconstructed accurately unless all of its frequency components are contained strictly within half the sampling frequency. This condition subsequently became known as the Nyquist criterion; half the sampling rate became known as the Nyquist rate. Nyquist developed his theorem for one-dimensional signals, but it has been extended to two dimensions. In a digital system, it takes at least two elements – two pixels or two scanning lines – to represent a cycle. A cycle is equivalent to a line pair of film, or two “TV lines” (TVL).

In Figure 7.6, the black square punctured by a regular array of holes represents a grid of small sampling apertures. Behind the sampling grid is a set of a dozen black bars, tilted 14° off the vertical, representing image information. In the region where the image is sampled,

78

DIGITAL VIDEO AND HD ALGORITHMS AND INTERFACES

Figure 7.7 Bitmapped graphic image, rotated.

you can see three wide dark bars tilted at 45°. Those bars represent spatial aliases that arise because the number of bars per inch (or mm) in the image is greater than half the number of apertures per inch (or mm) in the sampling lattice. Aliasing can be prevented – or at least minimized – by imposing a spatial filter in front of the sampling process, as I will describe for one-dimen- sional signals in Filtering and sampling, on page 191, and for two dimensions in Image presampling filters, on page 242.

Nyquist explained that an arbitrary signal can be reconstructed accurately only if more than two samples are taken of the highest-frequency component of the signal. Applied to an image, there must be at least twice as many samples per unit distance as there are image elements. The checkerboard pattern in Figure 7.1 (on page 76) doesn’t meet this criterion in either the vertical or horizontal dimensions. Furthermore, the titlebar element doesn’t meet the criterion vertically. Such elements can be represented in a bilevel image only when they are in precise registration – “locked” – to the imaging system’s sampling grid. However, images captured from reality almost never have their elements precisely aligned with the grid!

Point sampling refers to capture with an infinitesimal sampling aperture. This is undesirable in continuoustone imaging. Figure 7.7 shows what would happen if a physical scene like that in Figure 7.1 were rotated 14°, captured with a point-sampled camera, and displayed with a box distribution. The alternating on-off elements are rendered with aliasing in both the checkerboard portion and the titlebar. (Aliasing would be evident even if this image were to be reconstructed with

a Gaussian.) This example emphasizes that in digital imaging, we must represent arbitrary scenes, not just scenes whose elements have an intimate relationship with the sampling grid.

A suitable presampling filter would prevent (or at least minimize) the Moiré artifact of Figure 7.6, and prevent or minimize the aliasing of Figure 7.7. When image content such as the example titlebar and the desktop pattern of Figure 7.2 is presented to a presampling filter, blurring will occur. Considering only bitmapped images such as Figure 7.1, you might think

CHAPTER 7

IMAGE STRUCTURE

79

Соседние файлы в папке литература