Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

3D shape measurement based on projection of triangular patterns of two selected frequencies

Open Access Open Access

Abstract

In this paper, a temporal shift unwrapping technique is presented for solving the problem of shift wrapping associated with spatial shift estimation (SSE)-based fringe pattern profilometry (FPP). Based on this technique, a novel 3D shape measurement method is proposed, where triangular patterns of two different spatial frequencies are projected. The patterns of the higher frequency are used to implement the FPP, and the one with lower frequency is utilized to achieve shift unwrapping. The proposed method is able to solve the shift unwrapping problem associated with the existing multi-step triangular pattern FPP by projection of an additional fringe pattern. The effectiveness of the proposed method is verified by experimental results, where the same accuracy as existing multi-step triangular pattern FPP can be achieved, but enabling the measurement of objects with complex surface shape and high steps.

© 2014 Optical Society of America

1. Introduction

As an enabling technology for noncontact three-dimension (3D) profile measurement, fringe projection profilometry (FPP) [1] has attracted increasing research efforts due to many potential applications. Figure 1 shows the structure of a FPP system, consisting of a digital video projector, a CCD camera and the reference plane. The project generates a set of images with a particular fringe pattern, which are casted respectively onto the reference plane and the surface of the object, and the reflections are captured by the CCD camera. Due to the variance of the height of the object surface, the image patterns reflected by the object surface are a deformed version of the ones from the reference plan. These deformed patterns carry the information of surface shape, and hence can be used to retrieve the profile of the object.

 figure: Fig. 1

Fig. 1 Schematic diagram of FPP system.

Download Full Size | PDF

Many approaches have been developed during the past decades for FPP, e.g., Fourier transform profilometry (FTP) [2], phase shifting profilometry (PSP) [3–5], modulation measurement profilometry (MMP) [6], spatial phase detection (SPD) [7, 8], phase lock loop (PLL) profilometry [9], Moiré technique (MT) [10], laser triangulation measurement [11], color-coded fringe projection [12–14] and other methods [15, 16]. Among these approaches, the most widely used are those based on analysis of the phase maps of the fringe patterns, such as FTP and PSP. With these phase based approaches, fringe patterns projected are sinusoidal or periodic, and the deformed fringe patterns are considered as the result of phase modulation of the original patterns projected. Detection of phase maps from original and deformed fringe patterns enables the retrieval of the 3D shape.

Although phase based approaches have many advantages, they also suffers from a number of weaknesses. A major problem is the influence of nonlinear distortions inherent to digital video projection [17, 18], which makes it difficult for the original fringe patterns to be either sinusoidal or even periodic, which are required by phase based approaches. As an effort to solve the problem, an approach was proposed by Hu, et al. [19–21], referred to as spatial shift estimation (SSE) profilometry. Rather than relying on the analysis of the two phase maps, the technique utilizes the spatial shift of the corresponding pixels on the two fringe patterns, thus leading to a great advantage. The projected fringe patterns are no longer required to be sinusoidal or even periodic, resulting in an increased flexibility for the selection of the fringe patterns. Recently, Wu, et al. [22] introduced a multiple-step triangular-pattern spatial shift estimation algorithm by combining inverse function based shift estimation (IFSE) [21] with the multiple-step triangular-pattern phase shifting algorithm [23]. Compared to the IFSE approach in [21], the measurement accuracy is greatly improved due to the use of multiple patterns. Also compared to the existing intensity ratio based techniques [23], the proposed in [22] does not suffer from the influence of non-linear distortion.

While the multiple-step triangular-pattern spatial shift method is characterized by the above advantages, its performance relies on the result of shift unwrapping. This problem arises as the result of fringe reuse (that is, use of fringes with periodic light intensity variance). In the SSE based approaches, spatial shift between corresponding pixels on the two fringe patterns can only be detected within the range of [0, λ], where λ is the wavelength, or the width of the individual fringe, i.e., number of pixels per fringe stripe. Obviously, shift unwrapping is also required in order to correctly restore the 3D shape of the object surface [24], which is a challenging task particularly in the cases of noisy image patterns and discontinuities of the object surface. To solve the shift unwrapping problem, a multiple-wavelength unwrapping algorithm was proposed in [25], where, in order to unwrap the shift of a pair of fringe patterns with frequency f (i.e., the number of fringes on the image), a set of fringe patterns with their frequencies between 1 and f are also projected. This method is not suitable for fast 3D measurement, as the number of image patterns can be rather large.

The shift unwrapping problem is similar to the phase unwrapping problem in phase based FPP. Phase unwrapping problem arises because the phase can only be detected within the principle value range of [-π, π], but the true phase can exceed the range. In order to retrieve the actual surface shape of the object, phase unwrapping must be carried out to obtain the actual phase maps. Ding, et al. [26, 27] developed an approach to recover the absolute phase maps of two image patterns with selected frequencies. Inspired by the method [26], we introduce a new approach to unwrap the shift maps using two image patterns with selected frequencies. Based on this approach, we propose a novel 3D shape measurement method based on projection of a set of triangular patterns with a higher frequency, and a triangular pattern with a lower frequency. The former are utilized to yield a high frequency shift map by the multi-step triangular pattern profilometry method in [22], and the latter is used to obtain a shift map of lower frequency using the single-step triangular-pattern spatial shift estimation algorithm [28]. With the method to be proposed, the corresponding absolute shift maps can be retrieved, enabling the measurement of the profile of object surface. Compared with the multiple frequency shift estimation algorithm in [25], the number of image patterns is greatly reduced since only one additional image is used.

This paper is organized as follows. In Section 2 we firstly give a brief introduction on the conventional phase based fringe pattern profilometry and the spatial shift estimation based techniques. Then in Section 3 we indicate that the unwrapping problem in spatial shift estimation approach is similar to the phase unwrapping problem in phase based fringe pattern profilometry. The paper then introduces a method for unwrapping spatial shift maps of two fringe patterns with different frequencies. Based on this method, a novel 3D shape measurement approach is proposed. Finally in Section 4 a set of experimental results are presented to demonstrate performance of the proposed method. Section 5 concludes the paper.

2. Principle of fringe pattern profilometry

2.1 Phase based approaches for FPP

FPP is based on the triangulation principle described as follows. Without loss of generality and for the simplicity in expression, we assume that the projector has a fringe structure, where the light intensity varies periodically alone x direction, while keeping constant along y direction, as shown in Fig. 1. We uses(x), d(x) and h(x) to denote the variance of light intensity of the fringe pattern on the reference plane, the object surface and the height distribution along x coordinate respectively. We also assume that the reference plane and the object surface have the same reflective characteristics.

The phase based FPP utilize fringe patterns that are periodic and can be expressed as [29, 30]:

s(x)=k=0+bkcos(2πkf0x+Ψk)
and the deformed fringe pattern acquired from the object surface can also be expressed as:
d(x)=k=0+bkcos(2πkf0x+kϕ(x)+Ψk)
In the above equations, f0 is the spatial frequency of the fringe patterns, and bk is the amplitude of the k-th order harmonic component. ψk is the initial phase of the k-th order harmonic component, and ϕ(x) denotes the phase difference between the fundamental component of these two fringe patterns. If ϕ(x)can be detected, we are able to calculate the height distribution h(x) of the object surface by the following:
h(x)=l0ϕ(x)2πf0d0
where l0 is the distance between the camera and reference plane and d0is the distance between the camera and projector.

2.2 SSE based approaches for FPP

The phase based FPP methods suffer from some limitations. In particular, the fringe patterns used to project must be sinusoidal or periodic in order that the phase maps of s(x) and d(x) exist and can be detected. However, due to many undesired factors inherent to digital projection, such as geometrical distortion and nonlinear intensity distortion, purely sinusoidal fringe patterns are hard to produce. In order to solve these problems, Hu, et al. [19] introduced a method which is based on the SSE, where the 3D shape can be calculated by the following:

h(x)=l0u(x)d0
where u(x)is the spatial distance between a point x on d(x) and the corresponding point on s(x) with the same light intensity, that is:

d(x)=s(xu(x))

The SSE based approach is rather simple and straight forward. Withd(x) and s(x) available, if we are able to obtain u(x) to meet Eq. (5), we then can utilize Eq. (4) to determine h(x).

The spatial shift based approach has a particular advantage. The projected fringe patterns are no longer required to be sinusoidal, thus leading to significant flexibility in the design of the fringe patterns. A number of approaches were proposed to retrieve the u(x) [19–21]. Among these approaches, the one referred to as IFSE [21] is particularly interesting. Base on this approach, a single-step triangular-pattern spatial shift estimation algorithm is proposed in [28]. Wu, et al. [22] also introduced a multiple-step triangular-pattern spatial shift estimation algorithm by combining IFSE with the multiple-step triangular-pattern phase shifting algorithm [23], which greatly improved the accuracy of measurement.

3. The unwrapping problem

With most phase based approaches, ϕ(x) can only be identified within the range of [-π, π]. In other words, ϕ(x)is obtained by modulo 2π operation (called phase wrapping), resulting in discontinuities in its values. In order to recover the absolute phase Φ(x), we have:

Φ(x)=2πm(x)+ϕ(x)
where m(x)is an integer indicating the number of 2π lost due to phase wrapping, and the procedure of retrieving m(x)is referred to as phase unwrapping [26].

The wrapping problem also exists in SSE approaches. Ass(x) has a fringe structure with a periodic fringe of width λ, u(x) can only be detected within the main value of [0,λ], where λ is the width of an individual fringe referred to as the spatial wavelength of the fringe pattern. In other words, u(x) is wrapped into[0,λ]. Thus the true shift function, denoted as U(x), should be continues and recovered from u(x)as follows:

U(x)=λm(x)+u(x)
where m(x)are integers.

Similar to the phase unwrapping associated with PDE, we also must restore U(x)from the wrapped u(x), and the process is referred to as shift unwrapping [24].

3.1 Shift unwrapping of two fringe patterns with different frequencies

In order to unwrap the spatial shift, a simple method was proposed in [24] based on the relationship between the adjacent pixels. However, the method does not work if the intensity difference of neighborhood pixels is greater than λ/2, and hence the method does not work for the objects with high steps on the surface. Recently, a new a multiple-wavelength unwrapping algorithm for spatial shift estimation approach is proposed [25]. In this method, a series of fringe patterns with their frequencies increasing by a constant factor are projected. The first image has only one fringe covering the whole measurement area (i.e., its frequency is 1), and the remaining ones are characterized by their frequency increasing by a constant factor (or their wavelengths decreasing by a constant factor). For example, for unwrapping an image pattern with 16 fringes, at least 5 image patterns are required. The method is not suitable for fast 3D measurement, as the number of image patterns can be rather large. Therefore, the reduction of the number of image patterns while maintaining the accuracy of shift unwrapping is still a challenging problem.

To address this problem, inspired by the method proposed for phase unwrapping in [26], we introduce a method for unwrap the shift based on projection of two fringe patterns with different spatial frequency. We assume that the measurement area has a resolution of W × H, and that the normalized spatial frequencies of the two patterns are f1 and f2 respectively, which are positive integers representing the total number of fringes on the respective patterns. The corresponding wavelengths are λ1=H/f1 and λ2=H/f2. Combining this with Eq. (7) yields the following:

Ui(x)=Hfimi(x)+ui(x),i=1,2
As U(x) only depends on h(x), d0 and l0. The two unwrapped shift function should be same, that is, U2(x)=U1(x)=U(x).
Hf1m1(x)+u1(x)=Hf2m2(x)+u2(x)
or
f1f2H[u1(x)u2(x)]=m2(x)f1m1(x)f2
Using the similar reasoning to that in [26], we are able to derive a unique mapping from U(x)to m1(x)andm2(x), as follows:
m1(x)={f11(f11)Hf1U(x)<H22Hf1U(x)<3Hf11Hf1U(x)<2Hf100U(x)<Hf1
and
m2(x)={f21(f21)Hf2U(x)<H22Hf2U(x)<3Hf21Hf2U(x)<2Hf200U(x)<Hf2
Combing the right hand side of (11) and (12), the whole range [0, H] can be divided into f1+f2 segments. With the same approach as in [26] we can show that each segment corresponds to different values of m2(x)f1m1(x)f2as well as different combinations of m1(x)andm2(x). Hence a table can be generated forming a unique mapping from m2(x)f1m1(x)f2to m1(x)andm2(x).

As f1f2[u1(x)u2(x)]/H=m2(x)f1m1(x)f2, we can uniquely determine m1(x)andm2(x)from the value of f1f2[u1(x)u2(x)]/H. Let us choose f1=5 and f2=8, the mapping relationship is shown in Table 1.

Tables Icon

Table 1. mapping from m2(x)f1m1(x)f2 to m1(x)andm2(x)

From the above results we can reconstruct the absolute shift maps of two fringe patterns by following steps:

  • 1. Select two frequencies (f1,f2) and construct a mapping table, making sure the table provides a unique mapping from m2(x)f1m1(x)f2 to m1(x)andm2(x). It should be noticed that the principle of frequency pair selection is similar to the principle mentioned in [27], where the selection of frequency pair has the restriction that two selected frequencies must be coprime, otherwise the mapping will not be unique.
  • 2. Project two fringe patterns onto the object and acquire the two shift distance maps u1(x)and u2(x) by a spatial estimation algorithm.
  • 3. Calculate f1f2[u1(x)u2(x)]/H by rounding its value to the closest integer, denoted as M. Using the lookup table derived in Step 1, find the row (or entry) whose value of m2(x)f1m1(x)f2 is the closest to M. Record the corresponding m1(x)andm2(x)in the same row.
  • 4. Using m1(x)andm2(x)obtained in Step 3, reconstruct the absolute shift maps U(x) using Eq. (8).

Since the unwrapped spatial shift map is retrieved point by point, compared to algorithms which use neighborhood pixels to do the unwrapping [24], the method we propose here can measure the surface profile with larger steps or multiple separate objects.

3.2 3D shape measurement based on projection of triangular patterns of two selected frequencies

In this section, a novel 3D shape measurement method based on projection of triangular patterns of two selected frequencies is proposed. As both m1(x)andm2(x)can be used to reconstruct the absolute shift maps, we can select the one with lower measurement noise. Usually, the higher the spatial frequency used, the smaller the noise induced. Hence, we will project a set of triangular patterns with a higher frequency, from which a shift map can be obtained by the multiple-step triangular-pattern spatial shift estimation algorithm [22]. Then, a triangular pattern with a lower frequency is projected, and the shift map can be retrieved by the single-step triangular-pattern spatial shift estimation algorithm [28]. The absolute shift map of the higher frequency is recovered by the method described in 3.1, which is then used to determine the 3D shape. The procedure of the 3D shape measurement can be described using the flow chart in Fig. 2.

 figure: Fig. 2

Fig. 2 Flow chart of the proposed method.

Download Full Size | PDF

From the flow chart in Fig. 2, we can find that two different methods are used for the retrieving of the wrapped shift map. For the image pattern with higher frequency, the multiple-step triangular-pattern spatial-shifting algorithm is used. Here N is the number of images captured, which depends on the steps of the algorithm. As mentioned in method [22], at least two steps are required. However, the measurement accuracy can be improved by increasing the number of the spatial-shift steps. On the other side, the single-step triangular-pattern spatial-shifting algorithm is applied for the image pattern with the lower frequency, requiring only a single image pattern.

It can be seen from the flow chart, compared the proposed method to the existing single frequency multiple-step triangular-pattern spatial-shifting algorithm, only one additional image pattern is projected, which is much more efficient in terms of the number of fringe patterns required in contrast to the existing multiple-frequency shift unwrapping method [25].

4. Experiments and results

In order to test the performance of the approach proposed in Section 3, experiments were carried out in our laboratory. The triangular fringe patterns [22] are projected by a HITACHI CP-X260, and a Duncan Tech MS3100 3CCD digital camera is used to capture the fringe patterns. The digital camera is placed on top of the projector with a distance of 350.3 mm. The distance between the camera lens and the reference plane is 1295.1 mm. The resolution of the CCD camera is 1392 × 1039 pixels, and the field of vision for CCD camera is 250mm × 187mm. Hence, the equivalent spatial resolution is 0.1796 mm/pixel. The fringe projection system is calibrated using method in [31].

To verify that the proposed method can measure the surface profile with arbitrary step height, we first chose a flat box of 84mm high as the measured object. Figure 3(a) shows the photograph of the object. Figure 3(b) shows the captured fringe image of object with spatial frequency f = 5. Figures 3(c)–3(f) show the captured 4-step fringe images of object with spatial frequency f = 8.

 figure: Fig. 3

Fig. 3 Captured fringe images of flat box. (a) object; (b) fringe image (f = 5); (c)–(f) fringe images (f = 8).

Download Full Size | PDF

We firstly use single frequency multiple-step triangular-pattern spatial-shifting algorithm to reconstruct this object. The 4-step fringe images with normalized spatial frequency f = 8 are used, where the wrapped u(x)is retrieved using method in [22], and it is unwrapped using method in [24]. In contrast, the proposed method is also used. An additional fringe images with normalized spatial frequency f = 5 is captured, and the wrapped u(x)is retrieved using method in [28]. Figure 4 shows the cross section of the object and Fig. 5 shows the 3D reconstruct results.

 figure: Fig. 4

Fig. 4 Cross Section of flat box. (a) result using single frequency algorithm(f = 8); (b) result using select frequency algorithm(f1 = 5, f2 = 8)

Download Full Size | PDF

 figure: Fig. 5

Fig. 5 3D reconstruct results of flat box. (a) result using single frequency algorithm(f = 8); (b) result using select frequency algorithm(f1 = 5, f2 = 8).

Download Full Size | PDF

Figure 4(a) and Fig. 5(a) show that, if we only use patterns of the higher frequency, the object step height is measured as about 24 mm, which is obviously incorrect in contrast to the true value of 84mm. However, with the proposed method, the step height can be measured successfully, with the result very close to 84mm. This experiment demonstrated that the proposed approach can successfully perform the measurement even when the surface profile has a large step.

As mentioned in Section 3, since the measurement is performed on point-by-point basis, this technique can also be used to measure multiple separate objects or islands. To verify this, we measured two separate objects, a mask model and a plaster hand model. Both objects have complex surface shape. The proposed method is used with two spatial frequencies f1=8 and f2=13 respectively. We performed two experiments by fixing the higher frequency, but using different number of steps (3 and 6-steps) for retrieving the wrapped u(x). Figure 6(a) shows the two objects and Fig. 6(b) shows the captured fringe images of the objects with frequency f1=8. Figure 7 shows the captured images of the objects with 3-step triangular patterns with frequency f2=13, and Fig. 8 depicts the 6-step fringe images of objects with frequency f2=13.

 figure: Fig. 6

Fig. 6 Captured fringe images of separate objects with different wavelengths. (a) objects; (b) fringes image (f1 = 8).

Download Full Size | PDF

 figure: Fig. 7

Fig. 7 Captured 3-step fringe images of objects with f2 = 13. (a)–(c) fringe images in different step

Download Full Size | PDF

 figure: Fig. 8

Fig. 8 Captured 6-step fringe images of objects with f2 = 13. (a)–(f) fringe images in different step

Download Full Size | PDF

Figure 9 and Fig. 10 show the reconstructed 3D surface shape of the object using the proposed unwrapping method, where Fig. 9 shows the result using 3-step fringe images and Fig. 10 gives the result using 6-step fringe images. It can be seen from these figures that two separated objects is reconstructed successfully in both measurements. The height information for those places with high step drops (edge of mask and hand, nose part) is retrieved correctly. Some details, such as mouth and eyes part, are also recovered. This result demonstrated that the proposed method can successfully measure those complex objects with arbitrary step height, even if two objects are completely separate. Besides, it is noticed that the quality of the reconstructed model is also improved with the increase of the number of steps for the higher frequency image.

 figure: Fig. 9

Fig. 9 3D reconstruct results of two separated objects using 3-step fringe images. (a)–(d) results in different angle of view

Download Full Size | PDF

 figure: Fig. 10

Fig. 10 3D reconstruct results of two separated objects using 6-step fringe images. (a)–(d) results in different angle of view

Download Full Size | PDF

5. Conclusion

In this paper, we studied the shift wrapping problem associated with SSE based FPP. Compared to phase based approaches, SSE techniques are advantageous in that non-sinusoidal fringe patterns can also be employed, and that they do not suffer from the nonlinear distortion associated with the digital fringe projection. However, similar to the phase unwrapping problem encountered in phase based approaches, a shift unwrapping problem also exists in SSE based approaches. Hence we presented a temporal shift unwrapping technique to overcome this limitation. Based on this technique, we introduced a novel 3D shape measurement method. This method is an improvement of the existing multiple-step triangular-pattern spatial shift estimation algorithm [22], by the projection of only one additional image with a lower frequency. In order to test the performance of our proposed method, we carried out experiments on three different objects with significant steps and complex surface shapes: a flat box, a mask, and a plaster hand-model. The results show that the proposed method works very well, where complex objects surface with significant step height or multiple separate objects can be measured successfully with the same accuracy as the method in [22]. Compare to the existing multiple-wavelength unwrapping algorithm for spatial shift estimation approach [25], the number of image pattern required is greatly reduced.

It should be pointed out that similar to the method in [27], the selection of frequency pair has the restriction that two selected frequencies must be coprime. In order to meet such a requirement, the total number of pixels perpendicular to the fringe must be an integer multiple of the number of pixels within a fringe. However, such a selection may not be possible in some cases. Taking an example where the measurement area is 800 × 600 pixels, if the selected frequencies are f1=8 andf2=13, the numbers of pixels per fringe period will be about 75 and 46.1538, respectively, which are not integers and thus are not implementable. As a future work we will investigate the possibilities that the selected frequencies does not need to be an integer, thus yielding more flexibility for the design of the patterns.

References and links

1. S. Gorthi and P. Rastogi, “Fringe projection techniques: whither we are?” Opt. Lasers Eng. 48(2), 133–140 (2010). [CrossRef]  

2. X. Su and W. Chen, “Fourier transform profilometry: a review,” Opt. Lasers Eng. 35(5), 263–284 (2001). [CrossRef]  

3. H. Zhang, M. J. Lalor, and D. R. Burton, “Spatiotemporal phase unwrapping for the measurement of discontinuous objects in dynamic fringe-projection phase-shifting profilometry,” Appl. Opt. 38(16), 3534–3541 (1999). [CrossRef]   [PubMed]  

4. M. Halioua and H. C. Liu, “Optical three-dimensional sensing by phase measuring profilometry,” Opt. Lasers Eng. 11(3), 185–215 (1989). [CrossRef]  

5. Y. Fu and Q. Luo, “Fringe projection profilometry based on a novel phase shift method,” Opt. Express 19(22), 21739–21747 (2011). [CrossRef]   [PubMed]  

6. X. Su, L. Su, W. Li, and L. Xiang, “New 3D profilometry based on modulation measurement,” Proc. SPIE 3853, 1–7 (1998). [CrossRef]  

7. S. Toyooka and M. Tominga, “Spatial fringe scanning for optical phase measurement,” Opt. Commun. 51(2), 68–70 (1984). [CrossRef]  

8. S. Toyooka and Y. Iwaasa, “Automatic profilometry of 3-D diffuse objects by spatial phase detection,” Appl. Opt. 25(10), 1630–1633 (1986). [CrossRef]   [PubMed]  

9. R. Rodríguez-Vera and M. Servin, “Phase locked loop profilometry,” Opt. Laser Technol. 26(6), 393–398 (1994). [CrossRef]  

10. D. M. Meadows, W. O. Johnson, and J. B. Allen, “Generation of surface contours by moiré patterns,” Appl. Opt. 9(4), 942–947 (1970). [CrossRef]   [PubMed]  

11. A. Asundi and Z. Wensen, “Unified calibration technique and its applications in optical triangular profilometry,” Appl. Opt. 38(16), 3556–3561 (1999). [CrossRef]   [PubMed]  

12. P. Huang, Q. Ho, F. Jin, and F. Chiang, “Colour-enhanced digital fringe projection technique for high-speed 3-D surface contouring,” Opt. Eng. 38(6), 1065–1071 (1999). [CrossRef]  

13. Z. Zhang, C. E. Towers, and D. P. Towers, “Robust color and shape measurement of full color artifacts by RGB fringe projection,” Opt. Eng. 51(2), 021109 (2012). [CrossRef]  

14. J. Guo, X. Peng, J. Yu, X. Liu, A. Li, and M. Wang, “Real-time 3D imaging by using color structured light based on Hilbert transform,” Proc. SPIE 8856, 885624 (2013). [CrossRef]  

15. A. J. Moore and F. Mendoza-Santoyo, “Phase demodulation in the space domain without a fringe carrier,” Opt. Lasers Eng. 23(5), 319–330 (1995). [CrossRef]  

16. J. Villa, M. Servin, and L. Castillo, “Profilometry for the measurement of 3-D object shapes based on regularized filters,” Opt. Commun. 161(1–3), 13–18 (1999). [CrossRef]  

17. S. Zhang and S. T. Yau, “Generic nonsinusoidal phase error correction for three-dimensional shape measurement using a digital video projector,” Appl. Opt. 46(1), 36–43 (2007). [CrossRef]   [PubMed]  

18. K. Liu, Y. Wang, D. L. Lau, Q. Hao, and L. G. Hassebrook, “Gamma model and its analysis for phase measuring profilometry,” J. Opt. Soc. Am. A 27(3), 553–562 (2010). [CrossRef]   [PubMed]  

19. Y. Hu, J. Xi, E. Li, J. Chicharo, and Z. Yang, “Three-dimensional profilometry based on shift estimation of projected fringe patterns,” Appl. Opt. 45(4), 678–687 (2006). [CrossRef]   [PubMed]  

20. Y. Hu, J. Xi, J. Chicharo, E. Li, and Z. Yang, “Discrete cosine transform-based shift estimation for fringe pattern profilometry using a generalized analysis model,” Appl. Opt. 45(25), 6560–6567 (2006). [CrossRef]   [PubMed]  

21. Y. Hu, J. Xi, J. Chicharo, W. Cheng, and Z. Yang, “Inverse function analysis method for fringe pattern Profilometry,” IEEE Trans. Instrum. Meas. 58(9), 3305–3314 (2009). [CrossRef]  

22. K. Wu, J. Xi, Y. Yu, and Z. Yang, “3D profile measurement based on estimation of spatial shifts between intensity ratios from multiple-step triangular patterns,” Opt. Lasers Eng. 51(4), 440–445 (2013). [CrossRef]  

23. P. Jia, J. Kofman, and C. English, “Multiple-step triangular-pattern phase shifting and the influence of number of steps and pitch on measurement accuracy,” Appl. Opt. 46(16), 3253–3262 (2007). [CrossRef]   [PubMed]  

24. P. Cao, J. Xi, J. Chicharo, and Y. Yu, “A fringe period unwrapping technique for digital fringe profilometry based on spatial shift estimation,” Proc. SPIE 7432, 743208 (2009). [CrossRef]  

25. P. Cao, J. Xi, Y. Yu, and Q. Guo, “Spatial shift unwrapping for digital fringe profilometry based on spatial shift estimation,” J. Electron. Imaging 23(4), 043002 (2014). [CrossRef]  

26. Y. Ding, J. Xi, Y. Yu, and J. Chicharo, “Recovering the absolute phase maps of two fringe patterns with selected frequencies,” Opt. Lett. 36(13), 2518–2520 (2011). [CrossRef]   [PubMed]  

27. Y. Ding, J. Xi, Y. Yu, W. Q. Cheng, S. Wang, and J. F. Chicharo, “Frequency selection in absolute phase maps recovery with two frequency projection fringes,” Opt. Express 20(12), 13238–13251 (2012). [CrossRef]   [PubMed]  

28. P. Cao, J. Xi, Y. Yu, and Q. Guo, “Digital fringe profilometry based on triangular fringe patterns and spatial shift estimation,” Proc. SPIE 9110, 91100C (2014). [CrossRef]  

29. M. Takeda, H. Ina, and S. Kobayashi, “Fourier-transform method of fringe-pattern analysis for computer-based topography and interferometry,” J. Opt. Soc. Am. 72(1), 156–160 (1982). [CrossRef]  

30. M. Takeda and K. Mutoh, “Fourier transform profilometry for the automatic measurement of 3-D object shapes,” Appl. Opt. 22(24), 3977–3982 (1983). [CrossRef]   [PubMed]  

31. Q. Hu, P. S. Huang, Q. Fu, and F. Chiang, “Calibration of a three-dimensional shape measurement system,” Opt. Eng. 42(2), 482–493 (2003). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (10)

Fig. 1
Fig. 1 Schematic diagram of FPP system.
Fig. 2
Fig. 2 Flow chart of the proposed method.
Fig. 3
Fig. 3 Captured fringe images of flat box. (a) object; (b) fringe image (f = 5); (c)–(f) fringe images (f = 8).
Fig. 4
Fig. 4 Cross Section of flat box. (a) result using single frequency algorithm(f = 8); (b) result using select frequency algorithm( f 1 = 5, f 2 = 8)
Fig. 5
Fig. 5 3D reconstruct results of flat box. (a) result using single frequency algorithm(f = 8); (b) result using select frequency algorithm( f 1 = 5, f 2 = 8).
Fig. 6
Fig. 6 Captured fringe images of separate objects with different wavelengths. (a) objects; (b) fringes image ( f 1 = 8).
Fig. 7
Fig. 7 Captured 3-step fringe images of objects with f 2 = 13. (a)–(c) fringe images in different step
Fig. 8
Fig. 8 Captured 6-step fringe images of objects with f 2 = 13. (a)–(f) fringe images in different step
Fig. 9
Fig. 9 3D reconstruct results of two separated objects using 3-step fringe images. (a)–(d) results in different angle of view
Fig. 10
Fig. 10 3D reconstruct results of two separated objects using 6-step fringe images. (a)–(d) results in different angle of view

Tables (1)

Tables Icon

Table 1 mapping from m 2 ( x ) f 1 m 1 ( x ) f 2 to m 1 ( x ) and m 2 ( x )

Equations (12)

Equations on this page are rendered with MathJax. Learn more.

s ( x ) = k = 0 + b k cos ( 2 π k f 0 x + Ψ k )
d ( x ) = k = 0 + b k cos ( 2 π k f 0 x + k ϕ ( x ) + Ψ k )
h ( x ) = l 0 ϕ ( x ) 2 π f 0 d 0
h ( x ) = l 0 u ( x ) d 0
d ( x ) = s ( x u ( x ) )
Φ ( x ) = 2 π m ( x ) + ϕ ( x )
U ( x ) = λ m ( x ) + u ( x )
U i ( x ) = H f i m i ( x ) + u i ( x ) , i = 1 , 2
H f 1 m 1 ( x ) + u 1 ( x ) = H f 2 m 2 ( x ) + u 2 ( x )
f 1 f 2 H [ u 1 ( x ) u 2 ( x ) ] = m 2 ( x ) f 1 m 1 ( x ) f 2
m 1 ( x ) = { f 1 1 ( f 1 1 ) H f 1 U ( x ) < H 2 2 H f 1 U ( x ) < 3 H f 1 1 H f 1 U ( x ) < 2 H f 1 0 0 U ( x ) < H f 1
m 2 ( x ) = { f 2 1 ( f 2 1 ) H f 2 U ( x ) < H 2 2 H f 2 U ( x ) < 3 H f 2 1 H f 2 U ( x ) < 2 H f 2 0 0 U ( x ) < H f 2
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.