Добавил:
kiopkiopkiop18@yandex.ru t.me/Prokururor I Вовсе не секретарь, но почту проверяю Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:
Ординатура / Офтальмология / Английские материалы / Computational Analysis of the Human Eye with Applications_Dua, Acharya, Ng_2011.pdf
Скачиваний:
0
Добавлен:
28.03.2026
Размер:
20.45 Mб
Скачать

Michael Dessauer and Sumeet Dua

filter response at each location (x, y). Although matched filters require high computation, the method can be parallelized for faster computation time.

We have detailed several methods for retinal image anatomy segmentation that have proven successful when coupled with proper preand postprocessing steps. Choosing a proper segmentation method should depend on specific domain knowledge, the level of accuracy required, and computational resources available, with no one specific method standing out as producing the best results for every situation. The retinal anatomy creates unique challenges for successful segmentation, due to “distracters” that include blood vessel occlusion, spatially varying albedo, the presence of pathologies such as lesions and cotton wool spots, and residual non-uniform illumination effects.1

2.4. Feature Representation Methods for Classification

Along with segmentation, discriminative feature extraction is a prerequisite for accurate pathology classification. In this section, the term feature will describe a quantitative representation of the detected, localized, or segmented regions of the image, which are often specific parts of the retinal anatomy. We chose features that can best classify a retinal image for the specific problem, whether it is binary such as “diseased” or “nondiseased,” or varying degrees of a specific pathology. Most of these methods provide input into pattern recognition algorithms that use machine-learning methods for classification. Segmentation and image registration problems can also use feature extraction to localize or detect parts of the retinal anatomy. We will discuss several feature extraction methods that convert low-level information (e.g. pixel intensities) into robust numerical descriptors that can classify retinal pathologies and/or segment anatomy with high levels of specificity.

2.4.1. Statistical Features

After an image has been successfully segmented into its constituent parts, these regions of interest’s size, shape, color, and texture can provide simple and powerful values for discriminative classification. These descriptors are scale and rotationally invariant, which increases the robustness

70

Computational Methods for Feature Detection in Optical Images

and accuracy of classifying non-registered or orientation-corrected images. Microaneursym detection and classification using these types of features has been used to classify the degrees of diabetic retinopathy.23 We will provide the details on how to compute the three most common methods of statistical descriptors.

2.4.1.1. Geometric descriptors

First, we will describe geometric descriptors. Assuming that a region of connected components has been established through segmentation, characteristics of the geometry can provide discriminative classification specificity. These values only take into account the overall shape from linked boundary pixels, providing illumination invariant features sets. These scalar descriptors, listed below, also make entry into a classification algorithm straightforward.

Area is a simple sum of pixels contained within a bounded region. Although not overly useful alone, it is typically needed for finding more interesting shape statistics.

Centroid is the center of mass or the geometric center of a region, which provides insight on region localization at a single point (x, y). This location can be used as an axis for orientation adjustment for registration or feature extraction that is not rotationally invariant.

Eccentricity measures how circular a region’s shape is. Enclosing a region in an ellipse, the eccentricity is measured as the scalar value of the ratio between the distance from the center to the focus of the ellipse, c, and its major axis length, a:

Eccentricity = c .

a

This value will be zero if the region is a line and one if the region is a circle. This shape descriptor is useful when classifying hemorrhages.22

Euler number is a scalar value, E, that tabulates the number of objects, O, in a region subtracted from the number of holes, H : E = O H . Knowing the number of holes in a region provides insight that can be used to decide how to extract descriptors.

Extent is a scalar ratio of pixels between the bounding box and region area. This value gives an idea of the overall density of the region.

71

Michael Dessauer and Sumeet Dua

Axis lengths are the major and minor axis lengths, in pixels, of the elliptical representation of the region, used in computing orientation.

Orientation is the angle between the x-axis and the major axis. Perimeter gives the length of the boundary of the region. Compactness is the scalar measure of the ratio of the square of the

perimeter, p, and the area, a, given as: compactness = p2/a, which provides a minimal value for circular regions.

We use the shape of boundary segments to create a boundary histogram, p(vi), by representing the curve of a boundary as a 1D function, s(x), connecting its end-points and rotating to a horizontal access (Fig. 2.23). The amplitude at each boundary point is the distance from the x-axis to the boundary. We select an arbitrary number, A, of amplitude bins for the histogram and use p(vi) as the estimate of the probability of amplitude bin value vi occurring in the boundary segment. The nth moment of v about its mean is8:

 

A1

 

µn(v) =

 

 

(vi m)np(vi),

(2.41)

where

i=0

 

 

 

 

A1

 

m =

 

 

vi p(vi),

(2.42)

i=0

showing that m is the mean value of v and the second moment gives us the variance. These rotationally invariant shape descriptors retain physical boundary information, requiring only the first few moments to represent unique shape characteristics. Figure 2.23 shows the creation of 1D boundary histogram.

Fig. 2.23. (a) Segmented region, (b) top and bottom boundaries aligned on the x-axis, and (c) boundary magnitude histogram.

72

Computational Methods for Feature Detection in Optical Images

Although most of these geometrical descriptors are straightforward to calculate, they provide insight for refining choices on which regions are suitable for further analysis and as features for classification. These descriptors are not the only methods that can characterize regional information of boundaries and shapes, but do provide a good introduction into how scalar values can represent complex shapes.

2.4.1.2. Texture features

Another important statistical description of a segmented region of interest is the intra-region intensity distribution, which is loosely referred to as the texture. The texture of an image region can have multiple values, depending on how the internal pixels are represented. We will discuss two sets of texture feature descriptors that provide discrimination for retinal pathology classification: histogram-based texture features and spatial textures through co-occurrence matrices.

Using the gray-level histogram (explained in Sec. 2.2.2.3), we can calculate n-level moments, each of which provides insight into the characteristics of the intensity distribution, often a discriminate descriptor for classification. First, let z be a random variable with an ROI histogram p(zi), i = 0, 1, 2, . . . , L1. We use the moment equation to find the nth moment,8

L1

 

 

 

µn(z)

(zi m)np(zi),

(2.43)

i=0

 

where m is the mean value of z:

 

 

m =

L1

 

zi p(zi).

(2.44)

 

i=0

 

 

 

 

The most straightforward moment is the second moment, better known as the standard deviation. This value measures the intensity contrast, characterizing the relative smoothness of the ROI. The third moment is a measure of skewness, and the fourth moment describes the relative flatness. Other histogram-based texture descriptors include “uniformity,” which is formulated as

L1

 

 

 

U = p2(zi),

(2.45)

i=0

73

Michael Dessauer and Sumeet Dua

which has a maximum value when all pixels are the same (or “uniform”). Entropy, on the other hand, is a calculation of variability and is zero for a constant intensity, given as

L1

 

 

 

e = − p(zi) log2 p(zi).

(2.46)

i=0

Histogram-based texture features are useful due to their rotational invariability, but lack discrimination for instances where the spatial textures (pixel intensity positions) provide characteristic class information.

Representing a segmented region as pixels with location and intensity I(x, y), we create a set of co-occurrence matrices that tabulate how often a pair of pixels with similar intensity values “occur” in a specific orientation from one another. The user can define an orientation of 0(pixels I(x, y) and I(x + 1, y)), 45(pixels I(x, y) and I(x + 1, y 1)), 90(pixels I(x, y) and I(x, y 1)), and 45(pixels I(x, y) and I(x + 1, y + 1)), that will produce a square matrix, G, the size of intensity bins, k, in each dimension. The value at each location Gi,j will be the total number of times two pixels of a chosen orientation having similar intensity values “occur” within the region. After constructing co-occurrence matrices in the principal directions (typically 0, 45, 90, and 45), we normalize the matrices to give a joint probability occurrence of pixel pairs with the corresponding orientation and intensity range. We can find a set of spatially dependent texture descriptors termed Harilick features24:

Contrast:

 

|i j|2p(i, j),

 

 

i, j

 

 

 

 

Correlation:

 

 

(i µi)(j µj)p(i, j)

,

 

 

 

 

 

 

i, j

 

σiσj

 

 

 

 

 

 

Energy:

 

p(i, j)2 and

 

 

 

 

i, j

 

 

 

 

Homogeneity:

p(i, j)

 

i, j

1 + |i j|

.

 

Contrast measures intensity differences between neighboring pixels over the image, where correlation returns a measure of how correlated a pixel

74