Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Full space coordinate measurement method with a portable light pen by using hybrid light field imaging

Open Access Open Access

Abstract

The light pen is one of the most useful spatial coordinate measurement methods. However, due to the principle of vision measurement, the traditional light pen has the disadvantages such as limited measurement range, large structure size and complex calibration process. In order to revolutionize the design of light pen, this paper proposes a full space coordinate measurement method with a portable light pen by using hybrid light field imaging. Firstly, a light field imaging system with depth estimation ability has been introduced to simplify the light pen structure to three target points by using P3P method. In order to improve the accuracy of the depth estimation, a hybrid imaging method and EPI method have been used. Secondly, a full spatial measurement method with a rotation platform has been presented in detail. Finally, the experiments have been designed to verify the effectiveness and accuracy of the proposed method.

© 2021 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Space coordinate measurement is an essential technology for modern manufacturing and processing industry. With the development of the large-scale equipment manufacturing such as aircrafts ships, and automobiles, there is a great need for large-scale, easy-to-use and efficient methods of coordinate measurement [1].

The traditional coordinates measurement method mainly includes contact measurement and non-contact measurem1ent. Coordinate measuring machining (CMM) is one of the most widely used contact measurement method [2,3]. The measuring accuracy CMM is as high as 1µm with 1∼2 meters measured range. However, due to the contact measurement, mechanical transmission mode and limited measured range, the measurement efficiency and the application field is totally restricted. In order to extend the measured range and obtain non-contact measurement mode, laser tracker is one of the optional measurement methods [4]. The principle of laser tracker system is interference of light, and the measured range can be extended to more than 100 meters with 10µm precision. However, the main disadvantage of laser tracker is that only one special point can be measured at each test process. Meanwhile, the expensive price limits the applications of the laser tracker. Another popular coordinate measurement method is Indoor GPS which is a measurement network with several transmitters and portable receivers. However, for large-scale measurements, iGPS requires many signal transmitters to be installed indoors in order to obtain high accuracy which limits its portability [5]. Other 3D coordinates measurement methods such as laser scanner, laser radar can only obtain the point cloud data of measured piece, rather than single point. Therefore, the light pen technology based on vision measurement with its flexibility, real-time, non-contact measurement mode and excellent accuracy is the most suitable 3D coordinates measurement method in most industrial applications.

Many scholars have studied the light pen using the single CCD camera or binocular stereo vision principle [6]. The light pen with two cameras is based on the principle of binocular intersection, and the main advantage is simple in design. However, it can only measure coordinates within overlapped field-of-views of these two cam-eras and the cameras should be pre-calibrated. Conversely, the light pen with a single camera which doesn’t need to be calibrated each time offers good portability and ease of use. However, due to the principle of PNP which is used to solve the coordinates of the light pen with a single camera, at least six sets of points are required for calculation [79]. So, the structure of the light pen is much more complex, and the portability is seriously affected. To simplify the light pen with single camera, the minimum number of the points on the light pen with P3P algorithm is three while the distance information has been obtained. Laser tracking light pen measurement systems combined with vision such as the T-Probe from Leica in Switzerland and the Handy PROBE from Creaform in Canada capture the depth information from laser tracker and provide a suitable solution. However, the addition of the laser tracker makes the light pen system expensive and poorly portable. Another similar measurement method based on the rotating laser scanning without cameras can calculate the special coordinates by using the angular information obtained from the intersection laser beam [10]. Due to the similarity with the principle of monocular vision measurement, the minimum number of the marker points is six and the measurement system is much more complicated. Therefore, the light pen with a single camera by using P3P algorithm is the best solution for portability while the depth information can be obtained simultaneously.

Plenoptic imaging is an emerging technique that allows instantaneous 3D imaging of a scene using a single camera [11]. In contrast to conventional digital images, light field images can separately record light from multiple passing points [1214]. It greatly simplifies the depth estimation process. Although Tao et al. accomplishing the goal of depth estimation with light-field camera by combining off-focus and corresponding depth cues [15,16], the accuracy of depth estimation is greatly limited by the resolution of the light-field camera. Therefore, the low accuracy of depth obtained directly using the light-field camera limits its application in light pen measurements [17].

In order to extend the application field of the light pen based on vision measurement principle with only three points, this paper presents a full-space three-point light pen measurement system based on hybrid imaging [18]. The system consists of a light field camera and a conventional industrial camera, a motorised turntable and a three-point light pen. Then the depth measurement accuracy of the light pen can be improved significantly by using the hybrid imaging system and the full space dimensional measurement can be achieved with the turntable rotation method. Therefore, the portable light pen with the advantages of full space, high accuracy, large scale measurement, multi-tasking will be accomplished.

This paper is organized as follows. Section 2 describes the principles of the proposed method, including hybrid light field images, Epipolar Surface Image (EPI) depth estimation [1921], spinning parallelogram operator (SPO) measurements [2224] and the P3P method. In Section 3, experiments are conducted based on a platform composed of cameras and get the accuracy of the method is analysed according to the experimental results. Finally, in section 4, we present concluding remarks and a brief overview of future improvements.

2 Method

2.1 Light field image fusion

The light field camera is consisted of a main lens, an image sensor and a microlens array which is mounted front the image sensor [25]. It is possible to obtain the depth information directly from the light field image by using the refocus method. However, the accuracy is limited by the size of the microlens array [26].

In order to improve the depth estimation accuracy, this paper proposes a hybrid light field imaging method. With Homography correspondence, it is possible to fuse light field images with images obtained by other methods of measurement to achieve hybrid imaging. As shown in Fig. 1, the overall hybrid light field imaging measurement system consists of these components. The nib of the light pen is used as the position point to be measured, and the traditional CCD and the light field camera are fixed side-by-side on a rotating stage. The traditional CCD is fixed parallel to the light field camera. The plane mirror and 50-50 beam splitter are parallel, and located directly in front of the two cameras at an angle of 45° to ensure that the imaging centers of the two cameras are in the same position. These two imaging systems shoot the light pen through a 50-50 beam splitter and a plane mirror to achieve the measurement of the nib position. The 50-50 beam splitter enables the transmitted and reflected light to be of the same intensity, so that the images captured by both cameras can be considered at the same angle. In this way, two sets of images can be obtained from these two cameras.

 figure: Fig. 1.

Fig. 1. Schematic diagram of optical field hybrid imaging

Download Full Size | PDF

To ensure the accuracy of the hybrid light field imaging, the correspondence between the large pixels in the light field image and the pixels in the traditional CCD image will be obtained according to the homography principle. Firstly, by photographing the same checkerboard grid, the best (relatively similar) set of features of the two camera images can be obtained. Then the matching points can be used to calculate the Homography matrix that combinations the two images together. The relationship between the coordinates of the checkerboard grid in CCD image and light field image can be represented by Eq. (1). Finally, according to the known parameters, the correspondence of the other points can be calculated.

$$\left( {\begin{array}{c} {{x_{CCD}}}\\ {{y_{CCD}}}\\ 1 \end{array}} \right) = {H_{3 \times 3}} \times \left( {\begin{array}{c} {{x_{LF}}}\\ {{y_{LF}}}\\ 1 \end{array}} \right).$$
$({x_{CCD}},{y_{CCD}})$, $({x_{LF}},{y_{LF}})$ refer to the pixel coordinate values of a point in each of the two images. It can be seen that the Homography matrix, as a $3 \times 3$ matrix, has 8 unknown quantities that can be calculated using at least 4 sets of points. It turns out that the Homography matrix is very sensitive to the quality of the data which is passed to it, and the more points that fit the requirement for a given fixed plane, the higher the accuracy obtained by the calculation. In this paper, each subaperture image from the light field camera and the CCD camera image can be used to obtain the Homography matrix. Therefore, a higher resolution collection of light-field subaperture images can be obtained from the conventional camera images.

Figure 2 gives an illustration of the pixel fusion process between the light field images and the high-resolution traditional CCD images. The light field image contains $m \times n$ macro pixels which contains the pixel points covered under a single micro lens. Meanwhile, the high precision traditional CCD contains $p \times q$ effective pixels. Then the macro pixels can be mapped to the corresponding CCD imaging pixels by the Homography matrix, while the non-directly mapped pixel points can be calculated from the macro pixel information of the adjacent area. Finally, the high positional and angular resolution can be achieved simultaneously by the fusion algorithm from the hardware level. For each macro pixel in the light field image, a correspond pixel in high spatial resolution image can be found. Then these two images can be fused as one high resolution image in which some points have light field information and traditional CCD information while some points have only CCD information.

 figure: Fig. 2.

Fig. 2. Schematic diagram of the pixel fusion process

Download Full Size | PDF

As illustrated in Fig. 2, when a traditional CCD image with $p \times q$ pixels and a light field image with $m \times n$ macro pixels have been captured, the light field information can be calculated from the neighboring yellow macropixels for each blue pixel $(i,j)$ as formula (2):

$$\overrightarrow {L{F_{i,j}}} = \frac{{\mathop \sum \limits_{p \in \mathrm{\Omega }} {G_{{\sigma _r}}}\left( {d_p^{CCD}} \right) \cdot {G_{{\sigma _x}}}\left( {d_p^{xy}} \right) \cdot \overrightarrow {L{F_p}} }}{{\mathop \sum \limits_{p \in \mathrm{\Omega }} {G_{{\sigma _r}}}\left( {d_p^{CCD}} \right) \cdot {G_{{\sigma _x}}}\left( {d_p^{xy}} \right)}}.$$
where $\overrightarrow {L{F_{i,j}}}$ denotes the light field information vector in pixel $(i,j)$, the dimension of the vector is just the angular resolution of the light field. $\Omega $ represents the neighborhood of pixel $(i,j)$, $p \in \Omega $ denotes the pixel p in the neighborhood $\Omega $ to calculate $\overrightarrow {L{F_{i,j}}}$. $\overrightarrow {L{F_p}}$ denotes the light filed information vector in pixel p. ${G_\sigma }$ represents the Gaussian operator with zero mean and variance $\sigma $ . $d_p^{CCD}\textrm{ }$ and $d_p^{xy}$ denotes the Euclidean distance between the pixels $(i,j)$ and p in CCD image space and $(x,y)$ space, respectively.

2.2 EPI depth estimation

The light field image contains multi-view information from the scene, which makes depth estimation possible. In order to obtain high precision depth estimation, the EPI (Epipolar plane image) method has been introduced in this paper. By fixing the rows or columns of the subaperture array of a light field camera, a certain row or column of the light field subaperture image can be arranged in the corresponding position to obtain an EPI.

As shown in Fig. 3, the Epipolar plane view shows that there are “boundaries” in the image. Taking into account that each point on the sloping line in the EPI image is obtained from a different view of the same point in the real world. Considering this particularity of EPI, the spinning parallelogram operator (SPO) can be used to find the angle of a sloping line. A parallelogram operator is created in the Epipolar plane with a direction equal to the possible slope $\theta $. In this EPI image, the tilt angle of the segmentation line is $\theta $. Due to the limitation of computer computing power and the limitation of the spatial resolution of the real light field camera, 32 labels are chosen in this paper. The method measures the slopes by maximizing distribution distances between two parts of the parallelogram window to extract depth information.

 figure: Fig. 3.

Fig. 3. Schematic diagram of the pixel fusion process

Download Full Size | PDF

This correct matching of points as well as the line will split the parallelogram into two different parts, so the difference between these two parts can be used to predict the direction of the line. The center point of the parallelogram is set as the reference point and the orientation of the center line through the center point is the difference to be estimated. The center lines with different orientations, divide the window into two parts of the same size. The correct line, indicating the disparity information, can be figured out by finding the maximum distance between the distributions of pixel values on either side of the lines.

Firstly, the size of parallelogram operator will be defined by using a weighted function. The weights in the window are used to measure each pixel’s contribution to the distance. The weighting function $w(i,j)$ will be created by using the derivative of a Gaussian, as illustrated in Fig. 4. Specifically, for the reference point $P({x_r},{u_r})$ in the EPI ${I_{y,v}}(x,u)$ the corresponding pixels in the defined window are weighted by formula (3):

$${\omega _\theta }({i,j} )= c \cdot {d_\theta }({i,j} )\cdot {e^{\frac{{ - d_\theta ^2({i,j} )}}{{2{\alpha ^2}}}}},$$
where ${d_\theta }(i,j) = i - ({x_r} + (j - {u_r}) \cdot \tan \theta )$ is the distance between the pixel and the matching point P in the same row, and $({x_r} + (j - {u_r}) \cdot \tan \theta )$ is the hypothetical matching line’s position in the row j . The scale parameter α is determined by the complexity of the data, and it is found by different choices that the best results are obtained by choosing a parallelogram width of when α is chosen to be 0.8. c is a normalizing constant. The height of the window is defined as the number of subaperture in one angle, i.e. angular resolution. The angle $\theta $ in the figure represents the angle label. The folded line indicates the probability of distribution of different labels here, and the probability of distribution under label ${\theta _1}$ is the largest currently. For example, the EPI result in Fig. 4 is the angle value under label ${\theta _1}$.

 figure: Fig. 4.

Fig. 4. Schematic diagram of SPO

Download Full Size | PDF

Then the weight of the point has been set according to its horizontal distance from the hypothetical matching line so that the point which has the same distance from the line will contribute the same amount of information to the distance measurement. Therefore, this method separates the window based on the pixel’s distance to the hypothetical line. After dealing with the large disparities which contains stairstep structures due to the sparse sampling, the weighted window shows the similar stairstep structure to estimate the slope. Compared with the other slope estimation methods which need the densely sampling of the large slope, the SPO can be applied immediately.

2.3 Distribution rule of the P3P method’s multiple solutions

When the relatively accuracy distance between the measurement system and the light pen has been estimated by using the above method, the coordinates of the nib of the light pen can be calculated through PNP method (Perspective N points). The PNP method is an algorithm for solving camera poses based on 3D to 2D point pairs. While a set of points have been known to correspond to projections onto an image plane and given the intrinsic camera parameters, the transformation matrix (three rotation parameters and three translation parameters) between the object frame to the camera frame can be calculated. In the light pen measurement systems, while the world and image coordinates of the characteristic laser points on the light pen have been captured, the camera’s poses can be solved through the PNP method.

It has been proved that when the number of point pairs is greater than 3, the PNP problem has unique solution while the points on the light pen are not distributed on the same plane, and the commercially available light pen always has more than 6 feature points. However, as the number of the feature points increases, the portability deteriorates. Therefore, 3 feature points light pen has been designed in our method. When the number of point pairs is reduced to 3, the PNP algorithm is simplified to the P3P algorithm. Only 3 pairs of 3D-2D matched points are required to solve the external reference matrix of the camera to find the spatial coordinates of the measured point.

As shown in Fig. 5, these three-imaging points a, b and c corresponds the three spatial points A,B and C on the light pen. The center of the camera optical axis is the point O.

 figure: Fig. 5.

Fig. 5. Schematic diagram of P3P projection

Download Full Size | PDF

In the spatial triangle, Eq. (4) can be obtained through the cosine theorem:

$$\left\{ {\begin{array}{l} {O{A^2} + O{B^2} - 2|{OA} |\cdot |{OB} |\cdot \cos \left\langle {a,b} \right\rangle = A{B^2}}\\ {O{B^2} + O{C^2} - 2|{OB} |\cdot |{OC} |\cdot \cos \left\langle {b,c} \right\rangle = B{C^2}}\\ {O{A^2} + O{C^2} - 2|{OA} |\cdot |{OC} |\cdot \cos \left\langle {a,c} \right\rangle = A{C^2}} \end{array}} \right.,$$
when $|{OA} |= x|{OC} |$, $|{OB} |= y|{OC} |$, $u = A{B^2}/O{C^2}$, $uv = B{C^2}/O{C^2}$ and $uw = A{C^2}/O{C^2}$, Eq. (5) can be deduced:
$$\left\{ {\begin{array}{l} {(1 - v){y^2} - v{x^2} - 2y\cos \left\langle {b,c} \right\rangle + 2vxy\cos \left\langle {a,b} \right\rangle + 1 = 0}\\ {(1 - w){x^2} - w{y^2} - 2x\cos \left\langle {a,c} \right\rangle + 2wxy\cos \left\langle {a,b} \right\rangle + 1 = 0} \end{array}} \right..$$

The unknown elements x,y can be solved by Eq. (5) and then the known quantities $\cos \left\langle {a,b} \right\rangle$, $\cos \left\langle {a,c} \right\rangle$, $\cos \left\langle {b,c} \right\rangle$, v and w can be calculated. The P3P calculation is a system of binary quadratic Equations which have up to 4 sets of solutions. Moreover, there are two sets of imaginary roots, which do not conform to the length proportional relationship. Therefore, two sets of RT matrices corresponding to two special planes can be calculated through P3P problem.

The RT matrix allows the calculation of the coordinates of any point in the plane in this camera coordinate system. In order to eliminate the wrong solution, the depth of these points relative to the measurement system will be used.

Assuming that the optical center of the camera coordinate system and the origin of the light pen coordinate system are both on the optical axis line. As shown in Fig. 6, the Euclidean distance is d, and the light pen system is rotated around its $x$-axis by $\alpha $. The rotation matrix $\textrm{R}$ of the light pen coordinate system with respect to the camera coordinate system can be calculated through Eq. (6):

$$\textrm{R} = \left[ {\begin{array}{ccc} 1&0&0\\ 0&{\cos \alpha }&{ - \sin \alpha }\\ 0&{\sin \alpha }&{\cos \alpha } \end{array}} \right].$$

 figure: Fig. 6.

Fig. 6. The relationship between the optical axis of the camera and the measuring object

Download Full Size | PDF

The translation vector T is $T = {[0\textrm{,}0\textrm{,}d]^T}$, and sets d to 1000 mm, 1500 mm, 2000 mm and 2500 mm. The rotation angle α is at the range of $[ - 30^\circ ,30^\circ ]$ and the step size is set to $1^\circ $. The light pen coordinate system is established with the central light point as the origin, and the center of the light pen tip used in this paper is located 270 mm from the origin. Figure 6 shows that there are two results for P3P to solve the tilted surface, and the simulation is used to understand the difference between these two answers. The simulation results of the difference between the two sets of solutions caused by d difference and the tilt angles in the camera coordinate system are shown in Fig. 7.

 figure: Fig. 7.

Fig. 7. the simulation result of different d

Download Full Size | PDF

From the analysis of the simulation results, the greater the tilt angle, the greater the difference in depth of the P3P multiple solutions within a reasonable range. Hence, in the subsequent light field data collection, in order to increase the robustness of the algorithm and the accuracy of selecting the correct solution, the light pen should be tilted at a certain angle with respect to the camera coordinate system to increase the depth difference of the P3P multi-solution. Then the wrong solution can be excluded and the robustness of the algorithm can be improved.

2.4 Full spatial measurement method with the light pen

For the camera measurement method, the measurement range is limited by the shooting range of the camera. So a rotation stage for measurement is used to solve this problem. As shown in Fig. 8, the camera coordinate system cannot be guaranteed to match the rotation plane exactly due to the assembly. The actual optical center of the camera is located near the rotation axis, and there will be an assembly angle between the rotation axis and the camera coordinate axis, and the camera motion relationship on the rotation axis can be obtained by calibrating the relationship between them.

 figure: Fig. 8.

Fig. 8. The relation between the measurement system and the rotation axis

Download Full Size | PDF

Since the camera is fixed on the turntable, the rotation axis of the camera around the turntable is uncertain. Therefore, the external parameters of the measurement system before and after the change of rotation cannot be obtained directly. According to the external reference of the camera at different positions with respect to the world coordinate system, the measurement data at different angles can be unified in the same camera coordinate system.

When the camera rotates around a fixed axis the captured checkerboard grid image can be seen as a checkerboard grid rotating around a fixed rotation axis while the camera optical center position remains unchanged. As illustrated in Fig. 9, O is the rotation center, N is the rotation axis, and ${\textbf{R}_1}{\textbf{T}_1},{\textbf{R}_2}{\textbf{T}_2}, \cdot{\cdot} \cdot ,{\textbf{R}_n}{\textbf{T}_n}$ are the pose parameters of the camera at different angles in the coordinate system of the calibration plate, where $\textbf{R}$ is the rotation matrix and $\textbf{T}$ is the translation vector.

 figure: Fig. 9.

Fig. 9. Schematic of rotating shaft

Download Full Size | PDF

The monoclinic relationship between the world and pixel coordinates of the corner points on the calibration plate is obtained by combining the camera internal parameters according to the Eq. (7):

$$s\left[ {\begin{array}{c} {{x_i}}\\ {{y_i}}\\ 1 \end{array}} \right] = {\mathbf K}[{{{\mathbf R}_{\mathbf i}}{{\mathbf T}_{\mathbf i}}} ]\left[ {\begin{array}{c} {{X_i}}\\ {{Y_i}}\\ {{Z_i}}\\ 1 \end{array}} \right],$$
where s is the scale factor, $[{x_i},{y_i}]$ is the flush coordinate of the corner point of the calibration plate in the pixel coordinate system, $\textrm{K}$ is the internal reference matrix, and $[{X_i},{Y_i},{Z_i}]$ is the coordinate of the point in the calibration plate coordinate system.

By converting the camera coordinate systems of different angles into the first camera coordinate system, the pose of the $i$-th camera with respect to the first camera can be calculated by Eq. (8):

$$\left\{ {\begin{array}{l} {{{\mathbf R}_{i1}} = {{\mathbf R}_i} \times {\mathbf R}_1^{ - 1}}\\ {{{\mathbf T}_{i1}} = {{\mathbf T}_1} - {{\mathbf R}_1}{\mathbf R}_i^{ - 1}{{\mathbf T}_i}} \end{array}} \right..$$

By reading the coordinates of the checkerboard grid vertices at different angles, it is actually the position of the camera's optical center in a certain checkerboard grid coordinate system. As shown in Fig. 10, the optical center of the camera will form a circle which is distributed on a plane in three-dimensional space, and the camera's axis of rotation is a vector through the center of the circle and perpendicular to the plane. Therefore, the direction of rotation axis can be fitted using these coordinates. Due to manufacturing and assembly errors between the camera and the turntable, the axis of the turntable is not precisely vertical for the measurement work, and the camera's optical center does not pass through the axis of rotation. The actual structure of the camera's optical and rotational axes mounted on the turntable is shown in Fig. 10.

 figure: Fig. 10.

Fig. 10. Schematic diagram of rotating shaft

Download Full Size | PDF

The circle center of the camera rotation is at the intersection of the rotation plane and the rotation axis. Then a three-dimensional spatial circle fitting method is used to obtain the coordinates of the center of the circle. Assume that the fitted plane equation is $ux + vy + wz + d = 0$, where $N = (u,v,w)$ is the directional quantity of the plane. The projection points ${O_i} = ({x_i},{y_i},{z_i})$ are obtained by projecting the 3D coordinates $( {x^{\prime}_i},{y^{\prime}_i},{z^{\prime}_i})$ of the camera optical center at different positions onto the plane through Eq. (9).

$$\left\{ {\begin{array}{c} {{{x^{\prime}}_i} = uk + {x_i}}\\ {{{y^{\prime}}_i} = vk + {y_i}}\\ {{{z^{\prime}}_i} = wk + {z_i}} \end{array}} \right..$$
where $k ={-} u{x_i} - v{y_i} - w{z_i} - d$. Then The problem of fitting a spatial circle can be transformed into the problem of fitting a planar circle.

The first point in the sequence of the optical center is taken as the origin of the plane coordinate system while ${X_0} = {x^{\prime}_0},{Y_0} = {y^{\prime}_0},{Z_0} = {z^{\prime}_0}$.The positive direction of the X-axis is defined as the direction from the first camera photocenter point to the second point, and the positive direction of the Z-axis as the normal vector of the fitted plane, so that the direction of the Y-axis is also obtained. Finally, the three sets of the directions for the new axis can be solved through Eqs. (10)–(12):

$${x_1} = {x^{\prime}_1} - {x^{\prime}_0},\textrm{ }{y_1} = {y^{\prime}_1} - {y^{\prime}_0},\textrm{ }{z_1} = {z^{\prime}_1} - {z^{\prime}_0},$$
$${x_3} = u,\textrm{ }{y_3} = v,{z_3} = w,$$
$${x_2} = ({ - {z_1} - {y_1}{y_2}} )/{x_1},\textrm{ }{y_2} = \frac{{{x_1}{x_3} - {x_3}{z_1}}}{{{x_3}{y_1} - {x_1}{z_3}}},\textrm{ }{z_2} = 1.$$

The unit vectors of the three axes are $({{n_{1x}},{n_{1y}},{n_{1z}}} )$, $({{n_{2x}},{n_{2y}},{n_{2z}}} )$ and $({{n_{3x}},{n_{3y}},{n_{3z}}} )$, and the conversion matrix from 2D to 3D coordinates can be obtained through Eq. (13):

$${\mathbf H} = \left[ {\begin{array}{cccc} {{n_{1x}}}&{{n_{2x}}}&{{n_{3x}}}&{{X_0}}\\ {{n_{1y}}}&{{n_{2y}}}&{{n_{3y}}}&{{Y_0}}\\ {{n_{1z}}}&{{n_{2z}}}&{{n_{3z}}}&{{Z_0}}\\ 0&0&0&1 \end{array}} \right].$$

In this way the three-dimensional photocentric coordinates can be transformed to the same plane as Eq. (14):

$$\left[ {\begin{array}{c} {{{x^{\prime\prime}}_i}}\\ {{{y^{\prime\prime}}_i}}\\ 0\\ 1 \end{array}} \right] = {{\rm H}^{ - 1}}\left[ {\begin{array}{c} {{{x^{\prime}}_i}}\\ {{{y^{\prime}}_i}}\\ {{{z^{\prime}}_i}}\\ 1 \end{array}} \right].$$

Then the objective function is obtained as Eq. (15) using the least squares method for plane circle fitting.

$$\mathop U\limits_{\min } = \mathop \sum \limits_{i = 0}^{N - 1} {\left[ {{{\left( {{{x^{\prime\prime}}_i} - {a_0}} \right)}^2} + {{\left( {{{y^{\prime\prime}}_i} - {b_0}} \right)}^2} - {r^2}} \right]^2}.$$

By solving ${a_0},{b_0},r$, the final fitted 3D circle center coordinates $O = (a,b,c)$ can be obtained by Eq. (14).

The normal vector of the axis of rotation is $N = (u,v,w)$ and the position of the axis of rotation is $O = (a,b,c)$. When the camera is rotated by an angle $\theta $ around the rotation axis, the rotation matrix of the camera can be obtained as Eq. (16):

$${{\mathbf R}_\theta } = \left[ {\begin{array}{ccc} {{u^2} + ({{v^2} + {w^2}} )\cos \theta }&{uv(1 - \cos \theta ) - w\sin \theta }&{uw(1 - \cos \theta ) + v\sin \theta }\\ {uv(1 - \cos \theta ) + w\sin \theta }&{{v^2} + ({{u^2} + {w^2}} )\cos \theta }&{vw(1 - \cos \theta ) + u\sin \theta }\\ {uw(1 - \cos \theta ) - v\sin \theta }&{vw(1 - \cos \theta ) - u\sin \theta }&{{w^2} + ({{u^2} + {v^2}} )\cos \theta } \end{array}} \right].$$

The center of light of the camera is not over the axis of rotation, so there is also some translation. The translation matrix is shown as Eq. (17):

$${{\mathbf T}_\theta } = \left[ {\begin{array}{l} {({a({{v^2} + {w^2}} )- u(bv + cw)} )(1 - \cos \theta ) + (bw - cv)\sin \theta }\\ {({b({{u^2} + {w^2}} )- v(au + cw)} )(1 - \cos \theta ) + (cu - aw)\sin \theta }\\ {({c({{u^2} + {v^2}} )- w(au + bv)} )(1 - \cos \theta ) + (av - bu)\sin \theta } \end{array}} \right].$$

Therefore, once the camera is fixed to the rotation stage, the images taken by the rotated camera can be unified under the same coordinate system by simply calibrating the camera in relation to the spindle.

3. Experimental validation and analysis of results

In order to verify the effectiveness of the proposed method, a set of experiments has been designed. As shown in Fig. 11, the light pen was fixed to the three-dimensional translation stage at a titled angle. The light field camera and traditional CCD were fixed on both sides of the 50-50 beam splitter on the electric rotation stage, ensuring that both cameras were in the same view as far as possible.

 figure: Fig. 11.

Fig. 11. Measurement experiment system

Download Full Size | PDF

The light field camera was Lytro illum, with a microlens array of $625 \times 434$ and $15 \times 15$ pixels behind each microlens. The angular resolution is $15 \times 15$ which means 225 different directions of light emanating from a point can be recorded. This is equivalent to taking a picture of the object at the same time from 225 nearby locations behind the object plane. The reconstructed light field image is 3 channels $9375 \times 6510$, with a fixed focal length of 35 mm, an equivalent focal length of 70 mm, an equivalent conversion factor of 3.19, which means the nominal focal length of f = 21.944 mm, and a sensor image element size of 17.312 um.

This experiment used the HY-LW18-01A Lightweight Network Head, which uses a lightweight worm gear to allow continuous rotation in full horizontal space with a positioning accuracy of 0.1°.

The 5-dimensional packets captured by the light field camera were decoded to obtain the subaperture images and calibrated to obtain the camera internal reference matrix. The Homography matrix of the corresponding subaperture was obtained by calibrating the single-strain relationship captured by the traditional CCD, and the fusion algorithm was used to interpolate the traditional CCD image into the enlarged subaperture image using the single-strain principle to obtain a collection of subaperture images with higher resolution and reconstruct the light field image. Then the depth information of the light pen was obtained based on the light field EPI depth estimation. Two different sets of RT transformation matrices were obtained by the P3P algorithm. The depth values calculated by using EPI were compared with the depth values obtained from the two sets of coordinate transformations respectively, so as to exclude an erroneous solution and finally obtained the positional matrix of the light pen relative to the camera and the exact coordinates of the light pen tip.

The accuracy of the measurement was influenced by the precision of the electric rotation stage and the camera. Firstly, measurements were taken on the tiny moving light pen without rotating cameras. Secondly, the light pen was moved over a wide area and cameras are rotated through the rotation stage to take the measurement.

A checkerboard was fixed on the translation stage and the protation relationship between electric rotation stage and camera by rotating the camera could be obtained. Figure 12 shows the images of a fixed checkerboard captured by the camera at different angles.

 figure: Fig. 12.

Fig. 12. Checkerboard images from different angles

Download Full Size | PDF

The conversion matrix between the camera coordinate system of different angles and the world coordinate system could be calculated by using the monoclinic relationship between the phase plane and the tessellation plane in Eq. (7), and then the conversion matrix $\textbf{R}, \textbf{T}$ of the $i$-th camera relative to the camera coordinate system at the first view could be obtained by using Eq. (8). Taking the camera coordinate system with a 2° change as an example, the transformation matrix of the pose in the camera coordinate system at the first shot was shown in Eq. (18):

$${\textbf{R}_{21}} = \left[ \begin{array}{l} 0.99939\textrm{ }0.00711\textrm{ }0.03416\\ - 0.00708\textrm{ }0.99998\textrm{ } - 0.00086\\ - 0.03416\textrm{ } - 0.00086\textrm{ }0.99942 \end{array} \right],\textrm{ }{\textbf{T}_{21}} = \left[ \begin{array}{l} 1.23876\\ - 0.03501\\ - 0.03896 \end{array} \right].$$

The translation vector at the first camera coordinate system after the camera optical center rotation angle is shown in Table 1.

Tables Icon

Table 1. Camera translation

After obtaining different angular camera translations, the rotation parameters to the camera calculated using the plane fitting algorithm and the spatial circle fitting algorithm are shown in Table 2.

Tables Icon

Table 2. Rotor parameters

The errors in the rotation axis calibration algorithm of this method can be obtained from the rotation axis parameters in Table 2. Where the calculated points are fitted to the same plane, the root mean square error from the point to the fitted result plane is 0.0027mm, and then the circle is fitted on this plane, the root mean square error of the radius is 0.3124mm.

The effect of fitting the rotating axis in space is shown in Fig. 13.

 figure: Fig. 13.

Fig. 13. The protation relationship between electric rotation stage and camera coordinate system

Download Full Size | PDF

The images obtained from the light field camera and the traditional CCD were fused using Eq. (2) to obtain a high-resolution light-field image. Figure 14 compares the hybrid light field images obtained with this method. The EPI plane is selected and the depth of the synthesized light field image by rotating the parallelogram operator (SPO) was estimated. Figure 15 shows the depth values by SPO operator at each point on a horizontal line. This value indicates the depth difference between the point and the adjacent point.

 figure: Fig. 14.

Fig. 14. image of light field and hybrid light field (a) original light field image (b) hybrid light field image

Download Full Size | PDF

 figure: Fig. 15.

Fig. 15. the Epipolar plane image of the image and the corresponding to the maximum label

Download Full Size | PDF

According to Eq. (5), the calculation results of the two groups could be obtained. The relationship between the position of the light pen placement can be obtained through Fig. 15, and the wrong solution can be excluded from the calculation results. When the camera was not rotating and the small dimensions were measured from 0-200mm, the results can be obtained as shown in the Table 3.

Tables Icon

Table 3. Experimental results of light pen translational measurements (The first one is correct) (unit: mm)

When the turntable is introduced, large scale measurements can be implemented. When the moving distance exceeds 2000mm, the corresponding matrix could be obtained according to the rotation angle using Eqs. (16)–(17). The measurement results are shown in Table 4.

Tables Icon

Table 4. Large scale measurement experiment (unit: mm)

The above table records the measurement results of the traditional method and this paper's method for large-size equipment, where Method 1 is the traditional PNP measurement method and Method 2 is the proposed measurement method. From the table, it shows that the traditional light pen measurement system cannot measure beyond a certain range due to the small field of view.

It can be seen from Table 3 and Table 4, the measurement accuracy is mainly depended on the P3P algorithm when the rotation angle is during a proper range. Therefore, according to the simulation and Refs. [27], the measurement accuracy of the light pen is mainly affected by the measured distance. As demonstrated in Fig. 16, the measurement system was perpendicular to the movement path of the light pen and the vertical distance is about 3m. So the distance from the measurement system to the light pen would change from 3m to 4m. Therefore, there was no significant change in measurement error which is illustrated in Table 4. In the comparison experiment, the minimum measurement error of the traditional light pen measurement system is 0.16mm, the maximum is 0.23mm, and the average error is 0.18mm. The minimum error of the proposed measurement system is 0.12mm, the maximum error is 0.34mm, the average error is 0.22mm, and the root mean square error is 0.24mm.

 figure: Fig. 16.

Fig. 16. large scale measurement system

Download Full Size | PDF

4. Conclusion

A full-space hybrid optical field CMM system based on a precision rotary table to address the problems of traditional three-point optical pen vision measurement system has been proposed in this paper. This measurement method compensates for the limited resolution of the plenoptic camera and has higher accuracy when using the P3P method. From the experimental results, the measurement accuracy is mainly depended on the P3P algorithm when the rotation angle is during a proper range. The measurement error of the system is about 0.3 mm in the range of 5 m. This method allows the use of three marker points in a limited field of view to exclude incorrectly calculated values from traditional CCD by mixing light field images. Compared to binocular depth estimation, the proposed system does not require a large spacing and complex calibration process. The introduced rotation stage can effectively expand the measurement range of the three-point light pen measurement system. Finally, the full space coordinate measurement method with portable light pen has been accomplish.

Funding

National Natural Science Foundation of China (52005500); The Basic Science-research Funds of National University (3122019088); Foundation of Tianjin educational committee (2018KJ242); Tianjin Graduate Scientific Research Innovation Project (2020YJS006).

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. G. N. Peggs, P. G. Maropoulos, E. B. Hughes, A. B. Forbes, S. Robson, M. Ziebart, and B. Muralikrishnan, “Recent Developments in Large-Scale Dimensional Metrology,” P. I. Mech. Eng. B-j. Eng 223(6), 571–595 (2009). [CrossRef]  

2. R. V. Arencibia, C. C. Souza, H. L. Costa, and A. Piratelli-Filho, “Simplified model to estimate uncertainty in CMM,” J. Braz. Soc. Mech. Sci 37(1), 411–421 (2015). [CrossRef]  

3. G. X. Zhang, Y. H. Zhang, S. M. Yang, and Z. Li, “A Multipoint Method for Spindle Error Motion Measurement,” Cirp. Ann-manuf. Techn 46(1), 441–445 (1997). [CrossRef]  

4. M. Vincze, J. P. Prenninger, and H. Gander, “A Laser Tracking System to Measure Position and Orientation of Robot End Effectors Under Motion,” Ind. Robot 13(4), 305–314 (1994). [CrossRef]  

5. Z. Chen and F. Du, “Measuring principle and uncertainty analysis of a large volume measurement network based on the combination of iGPS and portable scanner,” Measurement 104, 263–277 (2017). [CrossRef]  

6. S. Liu, H. Zhang, Y. Dong, S. Tang, and Z. Jiang, “Portable light pen 3d vision coordinate measuring system- probe tip center calibration,” Meas. Sci. Rev 13(4), 194–199 (2013). [CrossRef]  

7. S. Fu, L. Zhang, N. Ye, W. Zhang, and S. Liu, “A flexible approach to light pen calibration for a monocular-vision-based coordinate measuring system,” Meas. Sci. Technol. 25(12), 125006 (2014). [CrossRef]  

8. R. Zhang, S. Liu, S. Wang, and X. Song, “Stylus Tip Center Position Self-Calibration Based on Invariable Distances in Light-Pen Systems,” Sensors 17(12), 131 (2017). [CrossRef]  

9. P. Zhao and G. Q. Ni, “Precise curved surface area measurement with a light-pen vision measurement system,” Optik 121(20), 1848–1852 (2010). [CrossRef]  

10. B. Chao, L. Yong, F. Jian-guo, G. Xia, L. Lai-peng, and D. Pu, “Calibration of laser beam direction for optical coordinate measuring system,” Measurement 73, 191–199 (2015). [CrossRef]  

11. E. M. Hall, T. W. Fahringer, D. R. Guildenbecher, and B. S. Thurow, “Volumetric calibration of a plenoptic camera,” Appl. Opt. 57(4), 914–923 (2018). [CrossRef]  

12. E. H. Adelson and J. Y. Wang, “Single lens stereo with a plenoptic camera,” IEEE Trans. Pattern Anal. Machine Intell. 14(2), 99–106 (1992). [CrossRef]  

13. M. Gutsche, H. Schilling, M. Diebold, and C. Garbe, “Surface Normal Reconstruction from Specular Information in Light Field Data,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Workshops (IEEE, 1735–1742 (2017).

14. C. Chen, H. Lin, Z. Yu, S. B. Kang, and J. Yu, “Light Field Stereo Matching Using Bilateral Statistics of Surface Cameras,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 1518–1525 (2014).

15. R. C. Bolles, H. H. Baker, and D. H. Marimont, “Epipolar-plane image analysis: An approach to determining structure from motion,” Int. J. Comput 1(1), 7–55 (1987). [CrossRef]  

16. Z. Li, Z. Xu, R. Ramamoorthi, and M. Chandraker, “Robust energy minimization for BRDF-invariant shape from light fields,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 5571–5579 (2017).

17. F. Dai, J. Lu, Y. Ma, and Y. Zhang, “High-resolution light field camera based on a hybrid imaging system,” Proc. SPIE 9273, 92730K (2014). [CrossRef]  

18. T. T. Ngo, H. Nagahara, K. Nishino, R. I. Taniguchi, and Y. Yagi, “Reflectance and shape estimation with a light field camera under natural illumination,” Int. J. Comput. Vis. 127(11-12), 1707–1722 (2019). [CrossRef]  

19. S. Wanner and B. Goldluecke, “Globally consistent depth labeling of 4D light fields,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2012), pp. 41–48.

20. S. Zhang, H. Sheng, C. Li, J. Zhang, and Z. Xiong, “Robust depth estimation for light field via spinning parallelogram operator,” Comput. Vis. Image. Und 145, 148–159 (2016). [CrossRef]  

21. L. Wei, Y. Wang, and Y. Liu, “Tensor-based light field compressed sensing and epipolar plane images reconstruction via deep learning,” in Proceedings of IEEE Access (IEEE, 2020), pp. 134898–134910.

22. H. Lv, K. Gu, Y. Zhang, and Q. Dai, “Light field depth estimation exploiting linear structure in EPI,” in Proceedings of IEEE International Conference on Multimedia & Expo Workshops (IEEE, 2015), pp. 1–6.

23. S. Wanner and B. Goldluecke, “Variational light field analysis for disparity estimation and super-resolution,” in Proceedings of IEEE Transactions on Pattern Analysis and Machine Intelligence (IEEE, 2013), pp. 606–619.

24. W. Chantara and Y. S. Ho, “Initial depth estimation using EPIs and structure tensor,” in Proceedings of IEEE Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (IEEE, 2017), pp. 255–258.

25. M. Levoy, “Light fields and computational imaging,” Computer 39(8), 46–55 (2006). [CrossRef]  

26. H. G. Jeon, J. Park, G. Choe, J. Park, Y. Bok, Y. W. Tai, and I. S. Kweon, “Accurate depth map estimation from a lenslet light field camera,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2015), pp.1547–1555.

27. S. Wang, S. Liu, and Q. Mao, “A CMM-Based Method of Control Point Position Calibration for Light Pen Coordinate Measuring System,” Sensors 20(19), 5592 (2020). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (16)

Fig. 1.
Fig. 1. Schematic diagram of optical field hybrid imaging
Fig. 2.
Fig. 2. Schematic diagram of the pixel fusion process
Fig. 3.
Fig. 3. Schematic diagram of the pixel fusion process
Fig. 4.
Fig. 4. Schematic diagram of SPO
Fig. 5.
Fig. 5. Schematic diagram of P3P projection
Fig. 6.
Fig. 6. The relationship between the optical axis of the camera and the measuring object
Fig. 7.
Fig. 7. the simulation result of different d
Fig. 8.
Fig. 8. The relation between the measurement system and the rotation axis
Fig. 9.
Fig. 9. Schematic of rotating shaft
Fig. 10.
Fig. 10. Schematic diagram of rotating shaft
Fig. 11.
Fig. 11. Measurement experiment system
Fig. 12.
Fig. 12. Checkerboard images from different angles
Fig. 13.
Fig. 13. The protation relationship between electric rotation stage and camera coordinate system
Fig. 14.
Fig. 14. image of light field and hybrid light field (a) original light field image (b) hybrid light field image
Fig. 15.
Fig. 15. the Epipolar plane image of the image and the corresponding to the maximum label
Fig. 16.
Fig. 16. large scale measurement system

Tables (4)

Tables Icon

Table 1. Camera translation

Tables Icon

Table 2. Rotor parameters

Tables Icon

Table 3. Experimental results of light pen translational measurements (The first one is correct) (unit: mm)

Tables Icon

Table 4. Large scale measurement experiment (unit: mm)

Equations (18)

Equations on this page are rendered with MathJax. Learn more.

( x C C D y C C D 1 ) = H 3 × 3 × ( x L F y L F 1 ) .
L F i , j = p Ω G σ r ( d p C C D ) G σ x ( d p x y ) L F p p Ω G σ r ( d p C C D ) G σ x ( d p x y ) .
ω θ ( i , j ) = c d θ ( i , j ) e d θ 2 ( i , j ) 2 α 2 ,
{ O A 2 + O B 2 2 | O A | | O B | cos a , b = A B 2 O B 2 + O C 2 2 | O B | | O C | cos b , c = B C 2 O A 2 + O C 2 2 | O A | | O C | cos a , c = A C 2 ,
{ ( 1 v ) y 2 v x 2 2 y cos b , c + 2 v x y cos a , b + 1 = 0 ( 1 w ) x 2 w y 2 2 x cos a , c + 2 w x y cos a , b + 1 = 0 .
R = [ 1 0 0 0 cos α sin α 0 sin α cos α ] .
s [ x i y i 1 ] = K [ R i T i ] [ X i Y i Z i 1 ] ,
{ R i 1 = R i × R 1 1 T i 1 = T 1 R 1 R i 1 T i .
{ x i = u k + x i y i = v k + y i z i = w k + z i .
x 1 = x 1 x 0 ,   y 1 = y 1 y 0 ,   z 1 = z 1 z 0 ,
x 3 = u ,   y 3 = v , z 3 = w ,
x 2 = ( z 1 y 1 y 2 ) / x 1 ,   y 2 = x 1 x 3 x 3 z 1 x 3 y 1 x 1 z 3 ,   z 2 = 1.
H = [ n 1 x n 2 x n 3 x X 0 n 1 y n 2 y n 3 y Y 0 n 1 z n 2 z n 3 z Z 0 0 0 0 1 ] .
[ x i y i 0 1 ] = H 1 [ x i y i z i 1 ] .
U min = i = 0 N 1 [ ( x i a 0 ) 2 + ( y i b 0 ) 2 r 2 ] 2 .
R θ = [ u 2 + ( v 2 + w 2 ) cos θ u v ( 1 cos θ ) w sin θ u w ( 1 cos θ ) + v sin θ u v ( 1 cos θ ) + w sin θ v 2 + ( u 2 + w 2 ) cos θ v w ( 1 cos θ ) + u sin θ u w ( 1 cos θ ) v sin θ v w ( 1 cos θ ) u sin θ w 2 + ( u 2 + v 2 ) cos θ ] .
T θ = [ ( a ( v 2 + w 2 ) u ( b v + c w ) ) ( 1 cos θ ) + ( b w c v ) sin θ ( b ( u 2 + w 2 ) v ( a u + c w ) ) ( 1 cos θ ) + ( c u a w ) sin θ ( c ( u 2 + v 2 ) w ( a u + b v ) ) ( 1 cos θ ) + ( a v b u ) sin θ ] .
R 21 = [ 0.99939   0.00711   0.03416 0.00708   0.99998   0.00086 0.03416   0.00086   0.99942 ] ,   T 21 = [ 1.23876 0.03501 0.03896 ] .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.