Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Risley-prism-based tracking model for fast locating a target using imaging feedback

Open Access Open Access

Abstract

Fast imaging tracking technology exhibits attractive application prospects in the emerging fields of target tracking and recognition. Smart and compact tracking model with fast and flexible tracking strategy can play a decisive role in improving system performance. In this paper, an effective imaging tracking model from a target to a rotation Risley prism pair embedded with a camera is derived by the beam vector propagation method. A boresight adjustment strategy using the inverse ray tracing and iterative refinement method is established to accomplish the function of fast locating a target. The influence of system parameters on boresight adjustment accuracy and even the dynamic characteristics of the tracking system are investigated to reveal the coupling mechanisms between prism rotation and imaging feedback. The root-mean-square tracking error is below 4.5 pixels by just once adjustment in the static target experiment, while the error in the dynamic experiment is below 8.5 pixels for a target moving at the speed of 50 mm/s, which validates the feasibility of the proposed method for fast imaging tracking applications.

© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

The imaging tracking technology is increasingly applied for target tracking and recognition in many emerging fields, such as advanced robotics, autonomous navigation, security surveillance and augmented reality [1,2]. Smart and compact tracking model with fast and flexible tracking strategy can play a decisive role in improving system performance. Therefore, many efforts are engaged in accomplishing rapid-response and strong-robustness imaging tracking function during the system design. At present, the multi-axis turntable has been widely used for boresight adjustment in imaging tracking applications, in which a typical case is the Pan-Tilt-Zoom (PTZ) camera [36]. However, the imaging tracking performance of multi-axis turntable is limited by its dynamic lag, vibration sensitivity and error accumulation. It has also been reported that the tracking component can be directly driven by an ultrasonic motor to achieve fast response and high precision [7,8]. But the coupling effect between the rotor and the stator of the ultrasonic motor will make the system design difficult to some extent.

In addition to translating or rotating the camera body as described above, an alternative approach to imaging boresight adjustment is to add an auxiliary optical device in front of the camera such that the boresight can be changed through refraction or reflection while keeping the camera stationary. Hu et al. [9] proposed an ultrafast mirror-drive vision system featured by time division and multithreading techniques to acquire multi-perspective information as two or more cameras. But such a reflective optical system is limited in that it is usually large in volume and relatively sensitive to the undesired deviation of camera boresight. Compared with reflective components, the refractive prism has the inherent advantages of compact structure, small moment of inertia and vibration insensitivity [10], which can offer great potential for boresight adjustment in imaging tracking applications. It has been demonstrated in most previous studies that prisms can be utilized to steer the camera boresight and enlarge the imaging field of view (FOV). Lavigne et al. [11] developed a step-stare imaging system based on rotation double prisms, in which the boresight could be continuously altered through the coaxial rotation of two prisms. Wang et al. [12] proposed a Risley-prism-based imaging system for super-resolution imaging and FOV extension. Tao et al. [13] established an active imaging system using Risley prisms and a scanning mirror, which can provide a flexible view for microassembly operation. These previous studies are mostly devoted to the application of prism-based imaging systems [1015], but few are focused on the problems associated with boresight control strategy in the dynamic imaging tracking field, especially the nonlinear relationship between prism rotation and boresight adjustment.

We have ever proposed a closed-loop visual tracking system using Risley prisms, in which a radial-circumferential decoupling control strategy was demonstrated [16]. Regardless of the initial orientations of two prisms, the method can achieve boresight adjustment with satisfactory precision but at the time-consuming cost, which is not always allowed in dynamic applications. The combination of two pairs of Risley prisms has also been explored to enlarge the tracking FOV [17] or achieve high-precision tracking by a coarse-fine coupling method [18]. However, the control strategy of two cascaded prism pairs becomes more complicated and may sacrifice the real-time tracking performance of the system, which limits its applicability to the dynamic imaging tracking field. Another related work is on the iterative refinement method (IRM) for finding the inverse solutions of Risley prisms given the desired beam pointing direction, which takes several milliseconds to calculate the prism orientations with the accuracy better than 0.1 µm [19]. Since the IRM has never been applied for Risley prisms in imaging tracking, more efforts need be concentrated on the modelling of Risley prism-based imaging system and the formulation of fast boresight control strategy to track a moving target in real time.

Based on the previous research, this paper further developed an effective imaging tracking model from a target to a rotation Risley prism pair embedded with a camera. Combined with the inverse ray tracing method and iterative refinement method, a novel boresight adjustment strategy is proposed to accomplish direct and fast control of Risley prisms for tracking a near-distance target, especially for most industrial applications within tracking ranges from a few meters to tens of meters. Moreover, the influence of system parameters on the dynamic characteristics of the tracking system, as well as the imaging tracking accuracy, are investigated to reveal the internal mechanisms for Risley-prism-based imaging tracking prototype.

This paper is organized as follows. In Section 2, the rigorous theoretical model of the Risley-prism-based imaging tracking is established, followed by the systematic analysis on the influence of camera and prism parameters on the model. In Section 3, the boresight adjustment process for fast locating a target is investigated, including the demonstration of the boresight adjustment principle and the discussion on the dynamic characteristics of the system. In Section 4, the imaging tracking experiments are carried out to validate the feasibility of the tracking model. Conclusions are finally drawn in Section 5.

2. Theoretical model

As illustrated in Fig. 1, the imaging tracking system consists of a CCD camera and a rotation double-prism system. Through independent rotation of two Risley prisms, the camera boresight can be altered to a specific cone to change the FOV. The parameters of two prisms, such as refractive index, wedge angle, thinnest-end thickness and clear aperture, are properly matched to ensure the camera imaging FOV at the maximum deviation angle of camera boresight.

 figure: Fig. 1.

Fig. 1. Schematic and coordinate diagram of the imaging tracking system using Risley prisms.

Download Full Size | PDF

The Cartesian coordinate system OCXCYCZC and OPXPYPZP are established as shown in Fig. 1, where the coordinate origins OC and OP are located at the optical center of the camera and the center of the plane side of the first prism, respectively. The ZC-axis and ZP-axis are oriented along the optical axis of the camera. The two Risley prisms are sequentially named as prism Π1 and prism Π2, each of which can rotate independently around ZP-axis with the rotation angles defined as θr1 and θr2, and the angular velocities defined as ωr1 and ωr2. The pitch angle and the azimuth angle of the adjusted boresight are defined as ρ and φ, respectively.

The beam propagation model of the rotation double prism system is established to illustrate how it is used in imaging tracking. On this basis, the inverse ray tracing model is developed to obtain the light path from the imaging feedback of a target, which is essential for demonstrating boresight adjustment principle. Moreover, the influence of camera and prism parameters on the results of the inverse ray tracing process is analyzed in this section.

2.1 Modelling of the beam propagation in rotation double-prism system

Figure 2 indicates the propagation of the refracted beam in the system. Each prism is designated with thinnest-end thickness d0, wedge angle α, refraction index n and clear aperture Dp. D1 is the distance between the plane surfaces of two prisms, while D2 is the distance between the target and surface 4. Without specific explanation, the parameters of the system are given as d0 = 5 mm, α = 10°, Dp = 80 mm, n = 1.517 and D1 = 67 mm in the following analysis in this paper.

 figure: Fig. 2.

Fig. 2. Schematic diagram of a beam passing through the rotation double-prism system.

Download Full Size | PDF

The normal vectors at each surface are respectively defined as N1, N2, N3 and N4 [20],

$$\begin{array}{l} {{\boldsymbol {N}}_{\bf 1}} = {[0,0,1]^\textrm{T}},\;{{\boldsymbol {N}}_2} = {[\cos ({\theta _{r1}})\sin (\alpha ),\sin ({\theta _{r1}})\sin (\alpha ),\cos (\alpha )]^\textrm{T}},\;\\ {{\boldsymbol {N}}_3} = {[ - \cos ({\theta _{r2}})\sin (\alpha ), - \sin ({\theta _{r2}})\sin (\alpha ),\cos (\alpha )]^\textrm{T}},\;{{\boldsymbol {N}}_4} = {[0,0,1]^\textrm{T}}. \end{array}$$
Supposing that the unit incident beam vector is defined as A0 = [xa0, ya0, za0]T, the unit vectors of the refracted beam passing through the four surfaces are represented as Ai = [xai, yai, zai]T, i = 1∼4. According to Snell’s law, the following formula can be obtained [20],
$$\begin{array}{c} {{{\boldsymbol {A}}_i}\textrm{ = }\frac{1}{n}{{\boldsymbol {A}}_{i - 1}} + \left\{ {\sqrt {1 - {{(\frac{1}{n})}^2}[1 - {{({\boldsymbol {A}}_{i - 1}^\textrm{T}{{\boldsymbol {N}}_i})}^2}]} - \frac{1}{n}{\boldsymbol {A}}_{i - 1}^\textrm{T}{{\boldsymbol {N}}_i}} \right\}{{\boldsymbol {N}}_i}\textrm{ = [}{x_{ai}},{y_{ai}},{z_{ai}}{]^\textrm{T}},\;(i = 1,3),}\\ {{{\boldsymbol {A}}_i}\textrm{ = }n{{\boldsymbol {A}}_{i - 1}} + \left\{ {\sqrt {1 - {n^2}[1 - {{({\boldsymbol {A}}_{i - 1}^\textrm{T}{{\boldsymbol {N}}_i})}^2}]} - n{\boldsymbol {A}}_{i - 1}^\textrm{T}{{\boldsymbol {N}}_i}} \right\}{{\boldsymbol {N}}_i}\textrm{ = [}{x_{ai}},{y_{ai}},{z_{ai}}{]^\textrm{T}},\;(i = 2,4).} \end{array}$$

Based on the coordinates of the incident point M1 (xm1, ym1, zm1), the intersection point of the refracted beam with respect to surface 2∼5 are represented as Mi (xmi, ymi, zmi), i = 2∼5 [20].

$$\left\{ {\begin{array}{c} {{x_{mi}} = {x_{i - 1}}{t_{i - 1}} + {x_{m({i - 1} )}}{\kern 1pt} {\kern 1pt} {\kern 1pt} }\\ {{y_{mi}} = {y_{_{i - 1}}}{t_{i - 1}} + {y_{m({i - 1} )}}}\\ {{z_{mi}} = {z_{i - 1}}{t_{i - 1}} + {z_{m({i - 1} )}}{\kern 1pt} } \end{array}} \right.$$
where, ti (i = 2∼5) represent the different proportionality coefficients.

2.2 Modelling of the image-based inverse ray tracing process

Figure 3 intuitively illustrates the image-based inverse ray tracing process according to a usual pinhole camera model, where the center of camera FOV and the spatial position of the target are represented by points C and M5, respectively. The imaging point of the target locates in point m in the image coordinate system oxyz. In this paper, the target distance in the range of few meters to tens of meters is far larger the focus length of tens of millimeters, and the actual image point of the target on the imaging plane is close to the optical center through continuous boresight adjustment. Therefore, the ideal pinhole model is used to establish the image-based inverse ray tracing model [21]. For the convenience of research, the rays are traced inversely from the imaging point m to the actual target position M5. The incident beam vector from the target is represented by vector A4 while the emergent beam vector is represented by vector A0, which is also the incident beam vector of the camera. The prisms’ orientations are respectively given as θr10 and θr20 at this moment.

 figure: Fig. 3.

Fig. 3. Schematic diagram illustrating the inverse ray tracing process.

Download Full Size | PDF

Assuming that the coordinate of imaging point b is (xm, ym) and the focal length of camera is f, the vector A0 can be derived from the following formula,

$${{\boldsymbol A}_0} = \frac{1}{{\sqrt {x_m^2 + y_m^2 + {f^2}} }}{[{{x_m},{y_m},f} ]^T}.$$
In the coordinate system OPXPYPZP, vector A0 is expressed as ${\mathbf A}_0^P$,
$${\boldsymbol A}_0^P = \frac{1}{{\sqrt {x_m^2 + y_m^2 + {f^2}} }}{[{ - {y_m},{x_m},f} ]^T}.$$
The intersection point of vector Ab0 and surface 1 is expressed as M1 (xm1, ym1, zm1),
$${[{{x_{m1}},{y_{m1}},{z_{m1}}} ]^T} = {[{ - {y_m}({d_t} - f)/f,{x_m}({d_t} - f)/f,0} ]^T}.$$
where dt represents the distance between the image sensor of camera and surface 1 of prism Π1.

By substituting the current orientations of two prisms into Eqs. (2) and (3), the vector A4 (xa4, ya4, za4) and the intersection point M4 (xm4, ym4, zm4) on surface 4 can be derived.

2.3 Analysis of the influence of system parameters on inverse ray tracing results

With the introduction of rotation double prisms into the camera model, the coupling effect between the camera and the prism pair must be clarified for the further optimization of the system. Two important parameters, including the focal length of the camera and the wedge angle of the prism, are taken as typical examples to reveal the influence of system parameters on the inverse ray tracing results. Without loss of generality, this method can be applied for the analysis of other related parameters as well.

In the coordinate system OPXPYPZP, the unit vector of the emergent beam (from surface 4) and the exiting point derived by ray tracing method are given as A4 (xa4, ya4, za4) and M4 (xm4, ym4, zm4), respectively. Since za4 is determined from xa4 and ya4 with the length constraint of A4, and zm4 is a constant value for a given system. Hence, only the influences of system parameters on (xa4, ya4) and (xm4, ym4) need to be taken into account. These influences are evaluated by the partial derivatives of xa4, ya4, xm4 and ym4 with respect to f, expressed as ηfax, ηfay, ηfmx and ηfmy. More comprehensive evaluation of these influences can be given by ηfa and ηfm as follows

$${\eta _{fax}} = \left|{\frac{{\partial {x_{a4}}}}{{\partial f}}} \right|,\;{\eta _{fry}} = \left|{\frac{{\partial {y_{a4}}}}{{\partial f}}} \right|,\;{\eta _{fmx}} = \left|{\frac{{\partial {x_{m4}}}}{{\partial f}}} \right|,\;{\eta _{fmy}} = \left|{\frac{{\partial {y_{m4}}}}{{\partial f}}} \right|.$$
$${\eta _{fa}} = \sqrt {{{\left( {\frac{{\partial {x_{a4}}}}{{\partial f}}} \right)}^2} + {{\left( {\frac{{\partial {y_{a4}}}}{{\partial f}}} \right)}^2}} ,\;{\eta _{fm}} = \sqrt {{{\left( {\frac{{\partial {x_{m4}}}}{{\partial f}}} \right)}^2} + {{\left( {\frac{{\partial {y_{m4}}}}{{\partial f}}} \right)}^2}} .$$
Supposing that the initial orientations of two prisms are given as (θr10, θr20) = (135°, 135°), the pixel deviations of the target imaging point are given as (Δu, Δv) = (100, 100) and dt = 60 mm. Figure 4 depicts the above partial derivatives as a function of the camera focal length f. It is worth mentioning that the camera focal length has approximately the same influence on the emergent beam vector and the exiting point position. The reason is that the incident beam happens to fall in the principal section of two prisms in this specific case. On the other hand, it is noteworthy that as the focal length increases, its influences on the emergent beam vector and the exiting point position decrease gradually. Therefore, a long focal lens can be applied to minimize the influence of focal length error on the ray tracing results. Whereas increasing the focal length will result in the reduction of FOV, a compromise should be reached between the focal length and the FOV of the camera. It is also observed from the comparison between ηfa and ηfm that the focal length has less influence on the emergent beam vector than on the exiting point position. Nevertheless, it should be noted that small error in the emergent beam vector will lead to a relatively large deviation of the boresight adjustment, and the deviation tends to get larger as the target distance increases. On the contrary, the boresight adjustment deviation caused by the position error of the exiting point remains invariable with the increment of target distance, and the deviation is negligible in any far-distance tracking situation.

 figure: Fig. 4.

Fig. 4. The influence of camera focal length f on the inverse ray tracing results. (a) The influence on the unit emergent beam vector, including ηfax, ηfay and ηfa. (b) The influence on the exiting point position, including ηfmx, ηfmy and ηfm.

Download Full Size | PDF

Similarly, assuming that f = 16 mm and other parameters remain unchanged, the attention is turned to the influences of prism wedge angle α on the inverse ray tracing results. The partial derivatives of xa4, ya4, xm4 and ym4 with respect to α are represented as ηαax, ηαay, ηαmx and ηαmy, and the comprehensive influences are respectively represented as ηαa and ηαm, as follows

$${\eta _{\alpha ax}} = \left|{\frac{{\partial {x_{a4}}}}{{\partial \alpha }}} \right|,\;{\eta _{\alpha ay}} = \left|{\frac{{\partial {y_{a4}}}}{{\partial \alpha }}} \right|,\;{\eta _{\alpha mx}} = \left|{\frac{{\partial {x_{m4}}}}{{\partial \alpha }}} \right|,\;{\eta _{\alpha my}} = \left|{\frac{{\partial {y_{m4}}}}{{\partial \alpha }}} \right|.$$
$${\eta _{\alpha a}} = \sqrt {{{\left( {\frac{{\partial {x_{a4}}}}{{\partial \alpha }}} \right)}^2} + {{\left( {\frac{{\partial {y_{a4}}}}{{\partial \alpha }}} \right)}^2}} ,\;{\eta _{\alpha m}} = \sqrt {{{\left( {\frac{{\partial {x_{m4}}}}{{\partial \alpha }}} \right)}^2} + {{\left( {\frac{{\partial {y_{m4}}}}{{\partial \alpha }}} \right)}^2}} .$$
Figure 5 shows that the wedge angle has the same influences on the emergent beam vector and exiting point position in both x and y directions. A system with larger wedge angle is more sensitive to the geometric errors of prisms. This observation also suggests that the wedge angle of each prism should be designed with a relatively low error tolerance, in order to minimize the negative impacts on the imaging tracking accuracy in a near-distance situation.

 figure: Fig. 5.

Fig. 5. The influence of prism wedge angle α on the inverse ray tracing results. (a) The influence on the unit emergent beam vector, including ηαax, ηαay and ηαa. (b) The influence on the exiting point position, including ηαmx, ηαmy and ηαm.

Download Full Size | PDF

Note that the above analysis is performed under the condition that two prisms rotate to the initial orientations, namely θr10 = θr20 = 135°, and the analysis results may become different for other prism orientations. In order to further explore whether the prisms’ orientations will affect the partial derivatives ηfax, ηfay, ηfmx, ηfmy, ηαax, ηαay, ηαmx and ηαmy, the initial orientations of two prisms are taken as independent variables for quantitative analysis. Figure 6 depicts the partial derivatives ηfax, ηfay, ηfmx, ηfmy, ηαax, ηαay, ηαmx and ηαmy as functions of the prisms’ initial orientations with fixed camera focal length f = 16 mm and wedge angle of prisms α = 10°. The conclusion can be drawn that the inverse ray tracing results are not only affected by the structural parameters such as focal length and wedge angle, but also dependent on the prisms’ orientations. Therefore, the orientations of two prisms must be taken into account to analyze the sensitivity of the system to various error sources. With the change of prisms’ orientations, the angle between the principal section of each prism and the XP-axis changes, which causes different trends between the X-component and Y-component of the above partial derivatives. When the orientations of two prisms are set as (θr10, θr20) = (90°, 90°), the principal sections of two prisms are parallel to the YP-axis. In this situation, the partial derivative ηαay reaches its maximum value while ηαax reaches its minimum value. Besides, because the exiting point position is related both to the direction of the incident beam and the position of the incident point, the system parameters have different influences on the emergent beam vector and the exiting point position even if the initial orientations of two prisms are not changed.

 figure: Fig. 6.

Fig. 6. The influence of prisms’ initial orientations on the inverse ray tracing results, including the different partial derivatives as functions of the prisms’ orientations: (a) ηfax, (b) ηfay, (c) ηfmx, (d) ηfmy, (e) ηαax, (f) ηαay, (g) ηαmx and (h) ηαmy.

Download Full Size | PDF

3. Camera boresight adjustment process

3.1 Boresight adjustment principle

In the imaging tracking applications, the emergent beam vector can be derived using the aforementioned inverse ray tracing method. Commonly, for a given emergent beam vector, the two-step method is effective to obtain the required rotation angles of two prisms [22]. But in the near-distance situation, there will be inevitable systematic errors introduced by using the two-step method, in which the variation of exiting point is neglected.

It is assumed that the exiting point locates at M4 (xm4, ym4, zm4), and the distance between M4 and the center of surface 4 is defined as ${d_m} = \sqrt {x_{m4}^2 + y_{m4}^2} .$ The pitch angle ρ of the emergent beam is decomposed into X-component and Y-component, namely ρx and ρy. Figure 7 illustrates the relation between the distance dm and the structural parameters of two prisms. It can be seen from Fig. 7(a) that the distance dm depends on the pitch angle ρ and has nothing to do with the azimuth angle φ. Since dm increases with the increment of ρ, the boresight adjustment deviation induced by the two-step method becomes more significant for a larger pitch angle ρ. Using the quantitative method, it can be seen from Fig. 7(b) that compare with the increase of pitch angle ρ, dm undergoes a considerable increase with the increase of distance D1. When D1 is set as the initial value D1 = 67 mm, the maximum and minimum values of dm are respectively expressed as dm,max and dm,min, as depicted in Figs. 7(c) and 7(d). It can be concluded that dm,max and dm,min both increase with increasing α, while on the contrary, dm,max and dm,min gradually decrease as d0 increases. Inspired by the above analysis, the systematic errors brought by the two-step method in near-distance situation can be effectively reduced by virtue of the optimization of system parameters in practical applications.

 figure: Fig. 7.

Fig. 7. The relation between distance dm and the parameters of two prisms. (a) Distance dm in various directions of the emergent beam. (b) The relation between the distance dm and the pitch angle ρ. (c) dm,max and dm,min as functions of different wedge angle α. (d) dm,max and dm,min as functions of different distance d0.

Download Full Size | PDF

In some specific cases when the target moves in a plane at the certain distance D2, the spatial position of the target can be accurately derived by providing the imaging feedback for the ray tracing model in Section 2.2. Given the target location in space, the iterative refinement method is used to solve the rotation angles of two prisms without sacrificing the real-time performance of the system, since the method requires the calculation time within several milliseconds for the accuracy better than 0.1 µm [19]. By combining the inverse ray tracing process with the iterative refinement method as shown in Fig. 8, it is viable to accomplish direct and high-speed boresight adjustment. It should be noted that the system calibration is needed before the tracking process to obtain the camera parameters and the prism system parameters, which are necessary for the boresight adjustment process. At the first stage, the feature recognition is conducted to obtain the pixel coordinates of the target imaging point from the image information. The second stage requires the inverse ray tracing process starting with the imaging point, and ending with the emergent beam vector and the exiting point from the prism Π2 to the target. At the third stage, the target location in space is determined from the inverse ray tracing results and the known distance D2, which can be provided for the iterative refinement method to calculate the consequent rotation angles of two prisms. Finally the prisms are rotated according to the orientations obtained through iterative calculation, such that the camera boresight can be steered towards the target after just once adjustment. During the calculation process of the iterative refinement method, the two prisms remain stationary until the calculation is completed and the acquired orientations are used to drive the two prisms.

 figure: Fig. 8.

Fig. 8. Flow chart of the boresight adjustment process.

Download Full Size | PDF

Assuming that the target moves in a spiral trajectory, Fig. 9 shows the position of the center of FOV obtained by the two-step method and the iterative refinement method, respectively. It is obvious that in the near-distance tracking situation, the boresight adjustment deviation occurs due to the approximate solution by the two-step method. Meanwhile, when the target movies away from the ZP-axis, the deviation grows with increasing pitch angle of the emergent beam, which is consistent with the analysis in the previous analysis. In this example, it takes less than nine iterations to obtain the prism orientations for any target point in the given trajectory when the allowable threshold is set as 0.1 µm. The boresight adjustment accuracy by this method is decided by the given allowable threshold according to the previous conclusions [19]. Therefore, if a given allowable threshold is set small enough, the exact solutions of prisms’ orientations can be derived to meet the requirement of high-precision target tracking.

 figure: Fig. 9.

Fig. 9. Comparison of the calculation accuracy of two-step method and the iterative refinement method, the first set of solutions is taken as an example. Where trajectory I represents the trajectory of the center of FOV by using the two-step method, while trajectory II represents the one by using the iterative refinement method.

Download Full Size | PDF

3.2 Dynamic characteristics of the boresight adjustment process

The dynamic characteristics of the boresight adjustment should be investigated to design the control strategy for the imaging tracking process. Considering the rotation characteristics of double prisms, the velocity of the target is respectively decomposed into the radial velocity vr and the tangential velocity vt, as depicted in Fig. 10. Since the pitch angle ρ of the emergent beam is determined by the deviation angle between two prisms, the radial velocity vr of the target depends on the relative rotational speed of two prisms, expressed as Δωr= |ωr1 - ωr2|. On the other hand, the azimuth angle φ is achieved by rotating two prisms at the same speed when the deviation angle of two prisms remains constant. Therefore, the tangential velocity vt is only related to the rotational speed when two prisms rotate synchronously, namely ωr = |ωr1| = |ωr2|. On this basis, the ratios Δωr/vr and ωr/vt are discussed to further study the dynamic laws of the boresight adjustment process.

 figure: Fig. 10.

Fig. 10. Schematic of the decomposition of target velocity along radial and tangential directions, where M5 represents the actual target point, v represents the actual velocity of the target, vr and vt respectively represent the radial and tangential velocity of the target.

Download Full Size | PDF

Under the system parameters given in Section 2, Fig. 11(a) describes the change rate of Δωr/vr as a function of pitch angle ρ. It can be seen that the ratio Δωr/vr varies severely with the pitch angle ρ increasing from 0° to ρmax, and the ratio Δωr/vr tends to infinity when ρ reaches ρmax. While the ratio ωr/vt decreases with increasing pitch angle ρ, as shown in Fig. 11(b). When the pitch angle ρ approaches to 0°, the ratio ωr/vt tends to infinity as well. It can be concluded that the required rotational speeds of two prisms increase dramatically when the target moves close to the center and outer edge of the tracking region, which puts forward higher requirements for the performance of the driving motors. Apart from the relation between prism rotation and target motion, the available tracking region can be redefined according to the ratios Δωr/vr and ωr/vt for the limited speed of the driving motors. As depicted in Fig. 12, there are two restricted areas corresponding to the center and outer edge of the tracking region, where the smooth tracking cannot be accomplished.

 figure: Fig. 11.

Fig. 11. Ratios of the rotational velocity of the prisms to the radial velocity of the target Δωr/vr and the tangential velocity of the target ωr/vt, where ρmax = 10.48°. (a) Δωr/vr, (b) ωr/vt.

Download Full Size | PDF

 figure: Fig. 12.

Fig. 12. Diagram illustrates the redefinition of the tracking region, where R1 represents the actually available tracking region for smooth tracking, R2 represents the blind zone, R3 and R4 represent the restricted areas according to ratios ωr/vt and Δωr/vr, respectively.

Download Full Size | PDF

As a case example, it is assumed that the maximum angular velocity of the prism is 50 rpm and the maximum velocity of the target is 200 mm/s. Firstly, suppose that the tangential velocity of the target reaches the maximum, in which case vr = 200 mm/s and vt = 0. When the distance D2 is set to 500 mm and the angular velocity of two prisms reaches the maximum, the ratio Δωr/vr is calculated to be 0.05233 and the corresponding maximum pitch angle for smooth tracking is 9.57°, denoted as ρo1 = 9.57°. It can also be calculated that ρo2 = 10.38° and ρo3 = 10.44° with target distance D2 = 1000 mm and 1500 mm, respectively. Similarly, when the radial velocity of the target reaches the maximum, the minimum pitch angles corresponding to different target distance are obtained, expressed as ρi1 = 4.371°, ρi2 = 1.461° and ρi3 = 0.88°. In fact, the smooth tracking can only be implemented within a certain range of pitch angle ρi and ρo for a given system. Moreover, it can be concluded that the available tracking region extends with increasing target distance D2. Furthermore, motors with higher speed can increase the ratios Δωr/vr and ωr/vt to expand the actual tracking region as well. In addition, other components can be adopted for the extension of the tracking region, such as a third prism [23,24], but it also complicates the design and control of the system.

4. Experiment validation

Figure 13 shows the test platform. The white circular marker attached to the end-effector of the manipulator acts as the tracked target. All functions are integrated into the software, including image information acquisition, boresight adjustment algorithm, and serial communication.

 figure: Fig. 13.

Fig. 13. Experiment platform. (1) CCD camera; (2) camera bracket; (3) rotation double-prism system; (4) 3 degree-of-freedom manipulator; (5) computer; (6) motor control box.

Download Full Size | PDF

In the experiments, the maximum velocity of the end-effector of the manipulator and the maximum angular velocity of the motor are set to 50 mm/s and 150°/s respectively, with the target distance D2 = 1000 mm and the field angle of the camera θc = 23.32°. The structural parameters of each prism are designed to provide sufficient imaging tracking range while facilitating the manufacturing and assembly process, among which Dp = 80 mm, α = 10° and d0 = 5 mm. According to the analysis in Section 3.1, increasing the distance D1 will enlarge the error caused by the change of exiting point position when tracking far-distance target based on the approximate formula of beam refraction. Considering both the minimization of distance D1 to reduce the error in far-distance condition and the convenience of manufacture and assembly, the distance D1 is finally set to 67 mm. However, the distance between the exiting point and the center of surface 4 is about 5 mm as analyzed in Section 3.1, which cannot be ignored under high-accuracy tracking condition, indicating that the iterative refinement method is valuable in this situation. Besides, based on the analysis in Section 3.2, the range of pitch angles for smooth tracking is between 0.37° and 10.46°. The maximum moving distances of the manipulator along x and y directions on the XOY plane are 307 mm and 255 mm, respectively.

4.1 Static tracking experiment

During the experiment process, the camera boresight is altered just one time by rotating two prisms according to the imaging feedback, and then the center of camera FOV is moved towards the target center, as shown in Fig. 14. Using the least square method (LSM), the experimental deviation distribution is described directly in Fig. 15(a), in which the red solid points represent the error distributions and the blue circle represents the experimental error distribution circle with a radius of 6.4097 pixels. As shown in Fig. 15(a), the error range obtained by LSM is plotted as a green circle with the center of (-0.0219, -0.1353) and the radius of 4.0047 pixels. In order to evaluate the accuracy of the boresight adjustment, the total deviation is defined as $\Delta e = \sqrt {\Delta {u^2} + \Delta {v^2}} $, where Δu and Δv represent the pixel deviations of target imaging point from the image center in different directions. Figure 15(b) shows the pixel deviations of 35 test points, and the root mean square error (RMSE) is calculated to be 4.0058 pixels, which is displayed with a green dotted line. The experimental deviation is mainly caused by calibration error, relative position error between camera and prism, geometric parameter errors of two prisms, and errors in motor and encoder. Regardless of all these various error sources, the experimental results still exhibit the feasibility and high-precision of the proposed method.

 figure: Fig. 14.

Fig. 14. Images captured before and after adjustment during the experiment, where (a)–(d) are images captured during the experiment of 4 groups among the total 35 test points.

Download Full Size | PDF

 figure: Fig. 15.

Fig. 15. Boresight adjustment accuracy. (a) Error distribution range. (b) Total pixel errors among 35 test points.

Download Full Size | PDF

4.2 Dynamic tracking experiment

In this experiment, the camera boresight is altered through continuous rotation of two prisms driven by servo motors in order to dynamically track a moving target. During the dynamic tracking process, the end-effect of the manipulator passes through Q1 (-230mm, -70mm), Q2 (-200mm, -20mm), Q3 (-140mm, 100mm) and Q4 (-244mm, -90mm) successively in the manipulator coordinate system. Once the camera boresight is pointed to the target, the imaging point of the target will locate at the center of the image in the ideal case. Thus the pixel deviation between the target imaging point and the image center is utilized as the feedback for the dynamic control of the tracking system. Figure 16 illustrates the closed-loop control block of the tracking system, where (u0, v0) and (u, v) respectively represent the pixel coordinates of the image center and the target imaging point, and (θr1, θr2) are the rotation angles of two prisms obtained by the proposed boresight adjustment principle in Section 3.1. During the experiment process, the target is locked in the central FOV of the camera as much as possible by the adjustment of camera boresight. The total pixel deviation Δe is defined as the pixel deviation between the target imaging point and image center. The root mean square error (RMSE) of the whole dynamic tracking process is used to evaluate the performance of the tracking system in dynamic tracking for near-distance target.

 figure: Fig. 16.

Fig. 16. Control block diagram of the dynamic tracking system.

Download Full Size | PDF

Figure 17 shows the images recorded every four seconds during the tracking process. Figure 18(a) describes the pixel coordinates of the target imaging point during the experiment. The solid lines represent the pixel coordinates of the target imaging point during the tracking process. While the dotted lines represent the coordinates of the image center, which is considered as the expected boresight projection in the image plane. The corresponding relationship between the four key points (Q1Q4) and the experimental time is also shown in Fig. 18(a), from which the motion direction of the manipulator at different times can be obtained. It is obvious that the target imaging point always locates near the image center in the whole dynamic tracking process. At about ten seconds after the beginning of the experiment, there is a sudden variation of the pixel coordinates of the target imaging point as shown in Fig. 18(a), which is caused by the abrupt change of the movement direction of the target. However, it is notable that the system can perfectly adapt to this unexpected disturbance and exhibits strong robustness. Figure 18(b) shows the total pixel deviation during the whole tracking process, which is plotted as a red solid line. The green dotted line plotted in Fig. 18(b) represent the RMSE during the whole tracking process, which is calculated to be 8.487 pixels. The experimental results indicate that the tracking system is feasible and robust in near-distance dynamic target tracking applications.

 figure: Fig. 17.

Fig. 17. Target images captured at different times during the dynamic tracking process, where (a) ∼ (e) are five images captured at an interval of four seconds.

Download Full Size | PDF

 figure: Fig. 18.

Fig. 18. The experimental results of the dynamic tracking process. (a) The pixel coordinates of the target imaging point in the tracking process, and Q1 (-230 mm, -70 mm), Q2 (-200 mm, -20 mm), Q3 (-140 mm, 100 mm) and Q4 (-244 mm, -90 mm) are four points that the manipulator passes through successively at different times. (b) The total deviation of the target imaging point in the dynamic tracking process.

Download Full Size | PDF

5. Conclusion

An effective imaging tracking model from a target to a rotation Risley prism pair embedded with a camera is derived by the beam vector propagation method. A theoretical model of the Risley-prism-based imaging tracking prototype is established to illustrate the law of boresight adjustment and imaging feedback. Considering the system error introduced by using the two-step method to solve the rotation angles of double prisms, a boresight adjustment principle combined the inverse ray tracing and the iterative refinement method is further developed to accomplish direct, fast controlling of prisms for tracking a near-distance target. The analysis of dynamic characteristics of the tracking model reveals the coupling mechanisms between the imaging feedback and the prism rotation, which offers a theoretical foundation for the system optimization and the formulation of control strategy. The experimental results well validate the flexibility and robustness of the tracking prototype, providing a useful guidance for further developments of the Risley-prism-based imaging tracking in broad fields, such as industrial detection and military reconnaissance.

Funding

National Natural Science Foundation of China (61675155, 61975152).

Acknowledgments

The authors gratefully acknowledge Wei Gong, Qiao Li, School of Mechanical Engineering, Tongji University, for assistance with discussion and critical reviews.

Disclosures

The authors declare no conflicts of interests.

References

1. Y. Li, J. Zhu, and S. C. Hoi, “Reliable patch trackers: Robust visual tracking by exploiting reliable patches,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (IEEE, 2015), pp. 353–361.

2. P. Li, D. Wang, L. Wang, and H. Lu, “Deep visual tracking: Review and experimental comparison,” Pattern Recognit. 76, 323–338 (2018). [CrossRef]  

3. D. W. Murray, P. F. McLauchlan, I. D. Reid, and P. M. Sharkey, “Reactions to peripheral image motion using a head/eye platform,” in 1993 (4th) International Conference on Computer Vision (IEEE, 1993), pp. 403–411.

4. A. Hampapur, L. Brown, J. Connell, A. Ekin, N. Haas, M. Lu, H. Merkl, and S. Pankanti, “Smart video surveillance: exploring the concept of multiscale spatiotemporal tracking,” IEEE Signal Process. Mag. 22(2), 38–51 (2005). [CrossRef]  

5. I. W. Park, J. Y. Kim, J. Lee, and J. H. Oh, “Mechanical design of humanoid robot platform KHR-3 (KAIST humanoid robot 3: HUBO),” in 5th IEEE-RAS International Conference on Humanoid Robots (IEEE, 2005), pp. 321–326.

6. C. Cigla, K. Emrecan Sahin, and F. Alim, “GPU based Video Object Tracking on PTZ Cameras,” in Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Workshops (IEEE, 2018), pp. 654–662.

7. A. Kanada, T. Mashimo, T. Minami, and K. Terashima, “High response master-slave control eye robot system using gaze tracking data,” in 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IEEE, 2015), pp. 937–943.

8. S. Shi, Z. Huang, J. Yang, Y. Liu, W. Chen, and K. Uchino, “Development of a compact ring type MDOF piezoelectric ultrasonic motor for humanoid eyeball orientation system,” Sens. Actuators, A 272, 1–10 (2018). [CrossRef]  

9. S. Hu, Y. Matsumoto, T. Takaki, and I. Ishii, “Monocular stereo measurement using high-speed catadioptric tracking,” Sensors 17(8), 1839 (2017). [CrossRef]  

10. Y. Zhou, Y. Lu, M. Hei, G. Liu, and D. Fan, “Motion control of the wedge prisms in Risley-prism-based beam steering system for precise target tracking,” Appl. Opt. 52(12), 2849–2857 (2013). [CrossRef]  

11. V. Lavigne, P. C. Chevrette, B. Ricard, and A. Zaccarin, “Step-stare technique for airborne high-resolution infrared imaging,” Proc. SPIE 5049, 128–138 (2004). [CrossRef]  

12. Z. Wang, J. Cao, Q. Hao, F. Zhang, Y. Cheng, and X. Kong, “Super-resolution imaging and field of view extension using a single camera with Risley prisms,” Rev. Sci. Instrum. 90(3), 033701 (2019). [CrossRef]  

13. X. Tao, H. Cho, and F. Janabi-Sharifi, “Active optical system for variable view imaging of micro objects with emphasis on kinematic analysis,” Appl. Opt. 47(22), 4121–4132 (2008). [CrossRef]  

14. A. A. Wagadarikar, N. P. Pitsianis, X. Sun, and D. J. Brady, “Video rate spectral imaging using a coded aperture snapshot spectral imager,” Opt. Express 17(8), 6368–6388 (2009). [CrossRef]  

15. Y. Zhu, W. Pan, J. Sun, A. Li, and Y. Li, “Compact design of projection lens for 3D profilometry based on interferometric fringes,” Optik 124(3), 209–212 (2013). [CrossRef]  

16. A. Li, S. Zhong, X. Liu, and Y. Zhang, “Double-wedge prism imaging tracking device based on the adaptive boresight adjustment principle,” Rev. Sci. Instrum. 90(2), 025107 (2019). [CrossRef]  

17. G. Roy, X. Cao, R. Bernier, and S. Roy, “Enhanced scanning agility using a double pair of Risley prisms,” Appl. Opt. 54(34), 10213–10226 (2015). [CrossRef]  

18. A. Li, W. Sun, X. Liu, and W. Gong, “Laser coarse-fine coupling tracking by cascaded rotation Risley-prism pairs,” Appl. Opt. 57(14), 3873–3880 (2018). [CrossRef]  

19. A. Li, X. Gao, W. Sun, W. Yi, Y. Bian, H. Liu, and L. Liu, “Inverse solutions for a Risley prism scanner with iterative refinement by a forward solution,” Appl. Opt. 54(33), 9981–9989 (2015). [CrossRef]  

20. A. Li, Double-Prism Multi-mode Scanning: Principles and Technology (Springer, 2018), Chap. 2.

21. R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision (Cambridge University, 2003), Chap. 6.

22. C. T. Amirault and C. A. DiMarzio, “Precision pointing using a dual-wedge scanner,” Appl. Opt. 24(9), 1302–1308 (1985). [CrossRef]  

23. P. J. Bos, H. Garcia, and V. Sergan, “Wide-angle achromatic prism beam steering for infrared countermeasures and imaging applications: solving the singularity problem in the two-prism design,” Opt. Eng. 46(11), 113001 (2007). [CrossRef]  

24. A. Li, X. Liu, and W. Sun, “Forward and inverse solutions for three-element Risley prism beam scanners,” Opt. Express 25(7), 7677–7688 (2017). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (18)

Fig. 1.
Fig. 1. Schematic and coordinate diagram of the imaging tracking system using Risley prisms.
Fig. 2.
Fig. 2. Schematic diagram of a beam passing through the rotation double-prism system.
Fig. 3.
Fig. 3. Schematic diagram illustrating the inverse ray tracing process.
Fig. 4.
Fig. 4. The influence of camera focal length f on the inverse ray tracing results. (a) The influence on the unit emergent beam vector, including ηfax, ηfay and ηfa. (b) The influence on the exiting point position, including ηfmx, ηfmy and ηfm.
Fig. 5.
Fig. 5. The influence of prism wedge angle α on the inverse ray tracing results. (a) The influence on the unit emergent beam vector, including ηαax, ηαay and ηαa. (b) The influence on the exiting point position, including ηαmx, ηαmy and ηαm.
Fig. 6.
Fig. 6. The influence of prisms’ initial orientations on the inverse ray tracing results, including the different partial derivatives as functions of the prisms’ orientations: (a) ηfax, (b) ηfay, (c) ηfmx, (d) ηfmy, (e) ηαax, (f) ηαay, (g) ηαmx and (h) ηαmy.
Fig. 7.
Fig. 7. The relation between distance dm and the parameters of two prisms. (a) Distance dm in various directions of the emergent beam. (b) The relation between the distance dm and the pitch angle ρ. (c) dm,max and dm,min as functions of different wedge angle α. (d) dm,max and dm,min as functions of different distance d0.
Fig. 8.
Fig. 8. Flow chart of the boresight adjustment process.
Fig. 9.
Fig. 9. Comparison of the calculation accuracy of two-step method and the iterative refinement method, the first set of solutions is taken as an example. Where trajectory I represents the trajectory of the center of FOV by using the two-step method, while trajectory II represents the one by using the iterative refinement method.
Fig. 10.
Fig. 10. Schematic of the decomposition of target velocity along radial and tangential directions, where M5 represents the actual target point, v represents the actual velocity of the target, vr and vt respectively represent the radial and tangential velocity of the target.
Fig. 11.
Fig. 11. Ratios of the rotational velocity of the prisms to the radial velocity of the target Δωr/vr and the tangential velocity of the target ωr/vt, where ρmax = 10.48°. (a) Δωr/vr, (b) ωr/vt.
Fig. 12.
Fig. 12. Diagram illustrates the redefinition of the tracking region, where R1 represents the actually available tracking region for smooth tracking, R2 represents the blind zone, R3 and R4 represent the restricted areas according to ratios ωr/vt and Δωr/vr, respectively.
Fig. 13.
Fig. 13. Experiment platform. (1) CCD camera; (2) camera bracket; (3) rotation double-prism system; (4) 3 degree-of-freedom manipulator; (5) computer; (6) motor control box.
Fig. 14.
Fig. 14. Images captured before and after adjustment during the experiment, where (a)–(d) are images captured during the experiment of 4 groups among the total 35 test points.
Fig. 15.
Fig. 15. Boresight adjustment accuracy. (a) Error distribution range. (b) Total pixel errors among 35 test points.
Fig. 16.
Fig. 16. Control block diagram of the dynamic tracking system.
Fig. 17.
Fig. 17. Target images captured at different times during the dynamic tracking process, where (a) ∼ (e) are five images captured at an interval of four seconds.
Fig. 18.
Fig. 18. The experimental results of the dynamic tracking process. (a) The pixel coordinates of the target imaging point in the tracking process, and Q1 (-230 mm, -70 mm), Q2 (-200 mm, -20 mm), Q3 (-140 mm, 100 mm) and Q4 (-244 mm, -90 mm) are four points that the manipulator passes through successively at different times. (b) The total deviation of the target imaging point in the dynamic tracking process.

Equations (10)

Equations on this page are rendered with MathJax. Learn more.

N 1 = [ 0 , 0 , 1 ] T , N 2 = [ cos ( θ r 1 ) sin ( α ) , sin ( θ r 1 ) sin ( α ) , cos ( α ) ] T , N 3 = [ cos ( θ r 2 ) sin ( α ) , sin ( θ r 2 ) sin ( α ) , cos ( α ) ] T , N 4 = [ 0 , 0 , 1 ] T .
A i  =  1 n A i 1 + { 1 ( 1 n ) 2 [ 1 ( A i 1 T N i ) 2 ] 1 n A i 1 T N i } N i  = [ x a i , y a i , z a i ] T , ( i = 1 , 3 ) , A i  =  n A i 1 + { 1 n 2 [ 1 ( A i 1 T N i ) 2 ] n A i 1 T N i } N i  = [ x a i , y a i , z a i ] T , ( i = 2 , 4 ) .
{ x m i = x i 1 t i 1 + x m ( i 1 ) y m i = y i 1 t i 1 + y m ( i 1 ) z m i = z i 1 t i 1 + z m ( i 1 )
A 0 = 1 x m 2 + y m 2 + f 2 [ x m , y m , f ] T .
A 0 P = 1 x m 2 + y m 2 + f 2 [ y m , x m , f ] T .
[ x m 1 , y m 1 , z m 1 ] T = [ y m ( d t f ) / f , x m ( d t f ) / f , 0 ] T .
η f a x = | x a 4 f | , η f r y = | y a 4 f | , η f m x = | x m 4 f | , η f m y = | y m 4 f | .
η f a = ( x a 4 f ) 2 + ( y a 4 f ) 2 , η f m = ( x m 4 f ) 2 + ( y m 4 f ) 2 .
η α a x = | x a 4 α | , η α a y = | y a 4 α | , η α m x = | x m 4 α | , η α m y = | y m 4 α | .
η α a = ( x a 4 α ) 2 + ( y a 4 α ) 2 , η α m = ( x m 4 α ) 2 + ( y m 4 α ) 2 .
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.