- •CONTENTS
- •PREFACE
- •Abstract
- •1. Introduction
- •2.1. Differential Geometry of Space Curves
- •2.2. Inverse Problem Formulation
- •2.3. Reconstruction of Unique Space Curves
- •3. Rigid Motion Estimation by Tracking the Space Curves
- •4. Motion Estimation Using Double Stereo Rigs
- •4.1. Single Stereo Rig
- •4.2. Double Stereo Rigs
- •5.1. Space-Time or Virtual Camera Generation
- •5.2. Visual Hull Reconstruction from Silhouettes of Multiple Views
- •5.2.1. Volume Based Visual Hull
- •5.2.1.1. Intersection Test in Octree Cubes
- •5.2.1.2. Synthetic Model Results
- •5.2.2. Edge Base Visual Hull
- •5.2.2.1. Synthetic Model Results
- •Implementation and Exprimental Results
- •Conclusions
- •Acknowledgment
- •References
- •Abstract
- •Introduction: Ocular Dominance
- •Demography of Ocular Dominance
- •A Taxonomy of Ocular Dominance
- •Is Ocular Dominance Test Specific?
- •I. Tests of Rivalry
- •II. Tests of Asymmetry
- •III. Sighting Tests
- •Some Misconceptions
- •Resolving the Paradox of Ocular Dominance
- •Some Clinical Implications of Ocular Dominance
- •Conclusion
- •References
- •Abstract
- •1. Introduction
- •2. Basic Teory
- •3. Bezier Networks for Surface Contouring
- •4. Parameter of the Vision System
- •5. Experimental Results
- •Conclusions
- •References
- •Abstract
- •Introduction
- •Terminology (Definitions)
- •Clinical Assessment
- •Examination Techniques: Motility
- •Ocular Motility Recordings
- •Semiautomatic Analysis of Eye Movement Recordings
- •Slow Eye Movements in Congenital Nystagmus
- •Conclusion
- •References
- •EVOLUTION OF COMPUTER VISION SYSTEMS
- •Abstract
- •Introduction
- •Present-Day Level of CVS Development
- •Full-Scale Universal CVS
- •Integration of CVS and AI Control System
- •Conclusion
- •References
- •Introduction
- •1. Advantages of Binocular Vision
- •2. Foundations of Binocular Vision
- •3. Stereopsis as the Highest Level of Binocular Vision
- •4. Binocular Viewing Conditions on Pupil Near Responses
- •5. Development of Binocular Vision
- •Conclusion
- •References
- •Abstract
- •Introduction
- •Methods
- •Results
- •Discussion
- •Conclusion
- •References
- •Abstract
- •1. Preferential Processing of Emotional Stimuli
- •1.1. Two Pathways for the Processing of Emotional Stimuli
- •1.2. Intensive Processing of Negative Valence or of Arousal?
- •2. "Blind" in One Eye: Binocular Rivalry
- •2.1. What Helmholtz Knew Already
- •2.3. Possible Influences from Non-visual Neuronal Circuits
- •3.1. Significance and Predominance
- •3.2. Emotional Discrepancy and Binocular Rivalry
- •4. Binocular Rivalry Experiments at Our Lab
- •4.1. Predominance of Emotional Scenes
- •4.1.1. Possible Confounds
- •4.2. Dominance of Emotional Facial Expressions
- •4.3. Inter-Individual Differences: Phobic Stimuli
- •4.4. Controlling for Physical Properties of Stimuli
- •4.5. Validation of Self-report
- •4.6. Summary
- •References
- •Abstract
- •1. Introduction
- •2. Algorithm Overview
- •3. Road Surface Estimation
- •3.1. 3D Data Point Projection and Cell Selection
- •3.2. Road Plane Fitting
- •3.2.1. Dominant 2D Straight Line Parametrisation
- •3.2.2. Road Plane Parametrisation
- •4. Road Scanning
- •5. Candidate Filtering
- •6. Experimental Results
- •7. Conclusions
- •Acknowledgements
- •References
- •DEVELOPMENT OF SACCADE CONTROL
- •Abstract
- •1. Introduction
- •2. Fixation and Fixation Stability
- •2.1. Monocular Instability
- •2.2. Binocular Instability
- •2.3. Eye Dominance in Binocular Instability
- •3. Development of Saccade Control
- •3.1. The Optomotor Cycle and the Components of Saccade Control
- •3.4. Antisaccades: Voluntary Saccade Control
- •3.5. The Age Curves of Saccade Control
- •3.6. Left – Right Asymmetries
- •3.7. Correlations and Independence
- •References
- •OCULAR DOMINANCE
- •INDEX
In: Binocular Vision |
ISBN: 978-1-60876-547-8 |
Editors: J. McCoun et al, pp. 81-105 |
© 2010 Nova Science Publishers, Inc. |
Chapter 3
THREE-DIMENSIONAL VISION BASED ON
BINOCULAR IMAGING AND APPROXIMATION
NETWORKS OF A LASER LINE
J. Apolinar Muñoz-Rodríguez*
Centro de Investigaciones en Optica, A. C. Leon, Gto, 37150 Mexico.
Abstract
We present a review of our computer vision algorithms and binocular imaging for shape detection optical metrology. The study of this chapter involves: laser metrology, binocular image processing, neural networks, and computer vision parameters. In this technique, the object shape is recovered by means of laser scanning and binocular imaging. The binocular imaging avoids occlusions, which appear due to the variation to the object surface. A Bezier approximation network computes the object surface based on the behavior of the laser line. By means of this network, the measurements of the binocular geometry are avoided. The parameters of the binocular imaging are computed based on the Bezier approximation network. Thus, the binocular images of the laser line are processed by the network to compute the object topography. By applying Bezier approximation networks, the performance of the binocular imaging and the accuracy are improved. It is because the errors of the measurement are not added to the computational procedure, which performs the shape reconstruction. This procedure represents a contribution for the stripe projection methods and the binocular imaging. To describe the accuracy a mean square error is calculated.
* E-mail address: munoza@foton.cio.mx. Tel: (477) 441 42 00.
82 |
J. Apolinar Muñoz-Rodríguez |
|
|
This technique is tested with real objects and its experimental results are presented. Also, the time processing is described.
Keywords: Binocular imaging, laser line, Bezier approximation networks.
1. Introduction
In computer vision, optical systems have been applied for shape detection. The use of structured illumination makes the system more reliable and the acquired data are easier to interpret. A particular technique is the laser line projection [1-3]. When a laser line is projected on an object, the line position is shifted in the image due to the surface variation and the camera position. From the line position and the geometry of the optical setup, the object contour is deduced. The main aim of the line projection is the detection of the line behavior in the image [4-6]. Also, the geometry of the setup is measured to obtain the object topography. When the surface variation produces an occlusion, the line position can not be detected. Therefore, in the area of the line occlusion the topographic data are not retrieved [7-8]. In the proposed technique here, occluded areas are retrieved by binocular imaging. To extract the object surface, the object is moved in the x-axis and scanned by a laser line. Based on the line behavior in the image, the object shape is retrieved. In each step of the scanning, a pair of binocular images of the line is captured. When the surface produced a line occlusion, the occluded area appears only one of the two images. Thus, the data missed in each image can be retrieved from each other. In this manner, the line occlusion is avoided by the binocular imaging. Typically, the disparity is needed to retrieve the object depth in a binocular system [9]. In the proposed technique, the disparity is determined by detecting the line position in each pair of the binocular images. In the proposed setup, the position of the line disparity is proportional to the object depth. By means of a Bezier network, a mathematical relationship between the line position and the object depth is generated. In this manner, the measurements of the focal length, and the distance between the two cameras are avoided. The Bezier network is constructed using the disparity of the line, which is projected on the objects with known dimensions. The disparity is detected measuring the position of the line displacement in the image. Thus, the network calculates the depth dimension for a position of a stripe displacement given. The line position is detected by means of Bezier curves with a resolution of a fraction of pixel. This position is processed by the network to determine the object shape. In this manner, all steps of the proposed technique are performed automatically by
Three-Dimensional Vision Based on Binocular Imaging… |
83 |
|
|
computational algorithms. Thus, the physical measurements on the setup are avoided. This kind of computational process improves the performance and the accuracy. Thus, a contribution is provided in the binocular imaging methods. In the proposed technique, the object is moved in the x-axis and scanned by a laser line. From the scanning, a set of binocular images of the line is captured. Each one of these images is processed by the Bezier curves to determine the position of the line disparity. This position is processed by the network to determine the object depth. The structure of the network consists of an input vector, a hidden layer and an output layer. The input vector includes: object dimensions, line position and parametric data. The hidden layer is constructed by neurons of Bezier Basis functions. The output layer is formed by the summation of the neurons, which are multiplied by a weight. The produced information by a pair of the binocular images corresponds to a transverse section of the object. The data of transverse sections are stored in an array memory to obtain the complete object shape. The results obtained in this technique are achieved with very good repeatability.
2. Basic Teory
The proposed setup figure 1 consists of a line projector, two CCD cameras, an electromechanical device and a computer. In this arrangement, the object is fixed on a platform of the electromechanical device. The platform moves the object in the x-axis. A laser stripe is projected onto the object surface by a laser diode to perform the scanning. In each step of the scanning, the laser line is digitized by two CCD cameras. In each image, the line is deformed in the x-axis according to the object surface. The profilometric method is based on the line deformation. By detecting the position of the line deformation, the object depth is determined. Therefore, the line position is main parameter to perform the object contouring. The object contour is broken when an occlusion appears. To retrieve the complete object contour, the binocular imaging is applied. In this system, the occluded line in the first camera can be observed by the second camera. Also, the occluded line in the second camera can be observed by the first camera. Thus, the binocular system overcomes the occlusion in the line projection. In this manner, the object contour can be computed completely. The contouring is described based on the geometry shown in figure 2. On the reference plane are located the x-axis, y-axis and the object depth is indicated by h(x, y) in the z-axis. The points o and p correspond to line projected on the reference plane and object surface, respectively. The focal length is indicated by F, d is the distance between the two cameras and the image center of each
84 |
J. Apolinar Muñoz-Rodríguez |
|
|
camera is indicated by ac and ec, respectively. When a laser line is projected on the surface, the line position is moved from ao to ap in the image plane of the camera a. At the same time, the line position is moved from eo to ep in the camera e. The line displacement in the image plane for each camera is represented by
as(x,y)= ao - ap. |
(1) |
es(x,y)= ep - e0. |
(2) |
By means of the positions ep and ap, the line displacement respect to the reference point “o” is given by as(x,y) for the camera a and es(x,y) for the camera e. Based on the pinhole camera model [10], the surface zi in a binocular system is deduced by
zi = |
d F |
, |
(3) |
δ +ε |
from this equation, δ+ε is the disparity. The distances of the disparity are deduced by δ = ac- ap and ε = ec – ep, respectively. To compute the surface zi by means of Eq.(3), the constants F, d, ac, ec and camera orientation should be known. Typically, the parameters d and F are obtained by an external procedure to the contouring. Then, these parameters are given to the computer system to reconstruct the object shape. This means that Eq.(3), can not be computed in the image processing of the laser line. For the proposed technique, the object depth h(x,y) is computed directly in the image processing of line position. In this procedure, a Bezier network provides a function that computes the object depth based on the line displacement. Also, this network provides information of the camera orientation, the focal lens, the center coordinates, the distance between tow cameras, disparity and center coordinates. Thus, the performance for object contouring is improved. In our contouring system, the object depth is computed by the network based on the line position using only one camera of the binocular system. This means that the network produces the depth h(x,y) using as(x,y) or es(x,y). The image processing provides a least one of the two displacements. When a line occlusion appears in the camera a, as(x,y) is missing. Then, the ep of the line disparity is used to compute the object depth h(x,y). In this manner, the occlusion problem of the laser line is solved by the binocular imaging. To detect the disparity, the line position is measured in every row of the binocular images. This position is computed by detecting the maximum of the line intensity in each
Three-Dimensional Vision Based on Binocular Imaging… |
85 |
|
|
row of the image. The intensity projected by a laser diode is a Gaussian distribution in the x-axis [11]. The intensity in every row of the image is represented by (x0, z0), (x1, z1), (x , z2),.......,(xn, zn), where xi is the pixel position and zi is the pixel intensity. To detect the line position, the maximum intensity measured in the image. To carry it out, Bezier curves and peak detection are used. The nth-degree Bezier function is determined by n+1pixels [12]. The nth-degree Bezier function is determined by two parametric equations, which are described by
n |
|
(1 |
- u)n u0 x0 |
n |
(1 - u) |
||
x(u) = |
|
|
+ |
|
|||
|
0 |
|
|
|
|
|
|
|
|
|
|
|
1 |
|
|
|
n |
(1 - u) 0 un xn, |
0 ≤ u ≤ 1. |
||||
+ |
|
||||||
|
|
|
|
|
|
|
|
|
n |
|
|
|
|
||
n |
|
(1 |
- u)n u0 z0 |
n |
(1 - u) |
||
z(u) = |
|
|
+ |
|
|||
|
0 |
|
|
|
|
|
|
|
|
|
|
|
1 |
|
|
|
n |
(1 - u) 0 un zn, |
0 ≤ u ≤ 1. |
||||
+ |
|
||||||
|
|
|
|
|
|
|
|
|
n |
|
|
|
|
||
n-1u x1 +
(4)
n-1u z1 +
(5)
Eq.(4) represents the pixel position and Eq.(5) represents the pixel intensity. To fit the Bezier curve shown in figure 3, x0, x1, x2,......,xn, are substituted into Eq. (10) and z0, z1, z2,....,zn, are substituted into Eq. (5). Then, these equations are evaluated in the interval 0≤ u≤1. In this case, the second derivative z”(u) > 0 in the interval 0≤ u≤1. Therefore, the maximum is detected by the first derivative equal to zero z´(u)=0 [13] via bisection method. Beginning with a pair of values ui = 0 and us =1, because z(u) is defined for the interval 0 ≤ u ≤ 1, u* is halfway between ui and us. If z´(u) evaluated at u = u* is positive, then ui = u*. If z´(u) evaluated at u = u* is negative, then us=u*. Next, u* is taken as the mid point of the last pair values that converges to the root. The value u* where z´(u) = 0 is substituted into Eq.(5) to obtain maximum position x*. The result is x* = 34.274 and the stripe position is ap = 34.274 pixels, which is shown in figure 3. The procedure of stripe detection is applied to all rows of the image. Then, the line position is processed by the network to obtain the object contour.
86 |
J. Apolinar Muñoz-Rodríguez |
|
|
Figure 1. Experimental setup.
Figure 2. Geometry of the experimental setup.0.
