Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Deep learning-based fringe modulation-enhancing method for accurate fringe projection profilometry

Open Access Open Access

Abstract

Fringe projection profilometry (i.e., FPP) has been one of the most popular 3-D measurement techniques. The phase error due to system random noise becomes non-ignorable when fringes captured by a camera have a low fringe modulation, which are inevitable for objects’ surface with un-uniform reflectivity. The phase calculated from these low-modulation fringes may have a non-ignorable phase error and generate 3-D measurement error. Traditional methods reduce the phase error with losing details of 3-D shapes or sacrificing the measurement speed. In this paper, a deep learning-based fringe modulation-enhancing method (i.e., FMEM) is proposed, that transforms two low-modulation fringes with different phase shifts into a set of three phase-shifted high-modulation fringes. FMEM enables to calculate the desired phase from the transformed set of high-modulation fringes, and result in accurate 3-D FPP without sacrificing the speed. Experimental analysis verifies its effectiveness and accurateness.

© 2020 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Fringe projection profilometry (i.e., FPP) plays an important role for non-contact, high-resolution and high-precision 3-D measurement [14]. In FPP, the desired phase is calculated by either using a phase-shifting algorithm [5] or a transform-based method [6]. The former achieves higher accuracy but requires at least three fringes [7], and the latter requires less fringes but loses 3-D details [810]. The measurement error mainly comes from system nonlinear response (i.e., gamma distortion), which can be compensated by using a look-up table [11], pre-coding [12], or gamma model-based technique [13], etc.

However, FPP system inevitably contains random noise due to unstable ambient light, camera and projector flicker, camera noise, and quantization error in the frame grabber and the projector [14], etc. The random noise may generate non-ignorable phase error when the captured fringes have a low fringe modulation [15,16], because the random phase error is inverse-proportional to the fringe modulation [17]. The phase error could be larger than 0.06 rad when the fringe modulation is lower than 11, and the 3-D measurement error could be more than 0.4 mm according to the experimental set-up in this study.

In FPP, the fringe modulation is influenced by the projector light reaching to the object’s surface, the surface reflectivity, the aperture and exposure of the camera [18,19], etc. The projector light can be assumed as constant, and the aperture of the camera is pre-determined. The measured object’s surface contains un-uniform reflectivity (i.e., partial surface with relatively low reflectivity and other with high reflectivity) [20]. The camera can use a large camera exposure to capture high-modulation fringes for local surface with a low reflectivity, but may capture saturated fringes for other surface with a high reflectivity. Furthermore, high-speed 3-D measurement requires a small camera exposure [21,22], which also brings challenges to capture high-modulation fringes.

Traditional methods reduce the phase error by temporally increasing fringes [14] or filtering the spectrum of the random noise [23,24], but the former sacrifices the measurement speed and the latter loses 3-D details [25]. Therefore, it has great importance for FPP to reduce the phase error effectively and accurately. Deep learning has been successfully used in computer vision applications [2629]. For low-light images enhancement, deep learning can map low-light images to normal-light images [30]. Recently, deep learning has also been introduced to fringe reduction in FPP [3134] and denoising in optical interferometric techniques [25,35,36], etc. Most deep learning-based FPP applications focus on improving the measurement speed by using a single fringe [31,33], which are difficult to achieve high accuracy similar to traditional phase-shifting algorithms using relatively more fringes.

In this paper, a deep learning-based fringe modulation-enhancing method (i.e., FMEM) is proposed by designing a fringe modulation enhancement convolutional neural network (i.e., FMENet) to transform two low-modulation fringes with different phase shifts into a set of three phase-shifted high-modulation fringes. The desired phase can be calculated from these transformed high-modulation fringes. The phase error is obviously reduced, and accurate 3-D shapes can be obtained even for fringes with a very low modulation.

The rest of the paper is organized as follows. Section 2 analyzes the phase error. Section 3 presents the proposed FMEM. Section 4 gives experiments. Section 5 concludes this paper.

2. Analysis of the phase error due to low fringe modulation

In FPP, a set of phase-shifted sinusoidal fringes are first projected by a projector, and then captured by a camera as [3740]

$$\begin{aligned} I_n(x,y)=a(x,y)+\Delta a_n(x,y)+b(x,y)\cos \left[\varphi(x,y)-\delta_n \right], n=1,2,3,\ldots,N, \end{aligned}$$
where $(x,y)$ denotes the camera coordinate, $\varphi$ denotes the desired phase, $\delta _n=2\pi (n-1)/N$ is the phase shift amount, $N$ is the number of phase steps, $\Delta a_n$ denotes the random noise, $a$ denotes the fringe background, and $b$ denotes the fringe modulation as described by [18,19,41]
$$b(x,y)=st_e\alpha(x,y)b^p(x,y),$$
where $s$ is the camera sensitivity, $t_e$ is the camera exposure, $\alpha$ is the surface reflectivity, and $b^p$ is the fringe modulation of projection fringes. The objects’ surface reflectivity is unknown and the camera sensitivity and fringe modulation of projection fringes are constant. Therefore, the fringe modulation can be adjusted by using different camera exposures. In real, the fringe modulation is calculated by [21]
$$b=\frac{2}{N}\sqrt{\left(\sum_{n=1}^{N}I_n\sin\delta_n\right)^2+\left(\sum_{n=1}^{N}I_n\cos\delta_n\right)^2}.$$
For simplicity, the notation of $(x,y)$ is omitted hereafter. The actual phase is calculated by using a least-squares algorithm [42]
$$\varphi=\tan^{-1}\frac{\sum_{n=1}^{N}I_n\sin\delta_n}{\sum_{n=1}^{N}I_n\cos\delta_n}.$$
The phase error caused by the random noise can be described by [39]
$$\Delta\varphi=\frac{2}{Nb}\sum_{n=1}^{N}\sin \left( \delta_n-\varphi \right) \Delta a_n,$$
which is proportional to the random noise and inverse-proportional to the fringe modulation. Because the noise level in a FPP system can be assumed as constant in a short camera exposure time [43,44], the phase error can be reduced by enhancing the fringe modulation according to Eq. (5). A large camera exposure directly enhances the fringe modulation, but reduces the measurement speed and also causes motion-induced problems [45].

For clarity, a simulation is provided to illustrate the phase error by selecting the commonly used three-step FPP. A Gaussian distribution with a zero-mean and a standard derivation of 1 is used to generate the random noise [16]. The phase error of the three-step FPP for different modulations is shown in Fig. 1(a). When the fringe modulation is smaller than 30, the phase error becomes non-ignorable as larger than 0.02 rad [46]. When a small fringe modulation of 10 is selected, the phase error of FPP with different phase steps is shown in Fig. 1(b). The number of phase steps is at least 25 to reduce the phase error to less than 0.02 rad.

 figure: Fig. 1.

Fig. 1. The phase error: (a) for different fringe modulations, and (b) phase steps.

Download Full Size | PDF

3. Proposed FMEM

The proposed FMEM designs a fringe modulation enhancement convolutional neural network (i.e., FMENet) to transform two low-modulation fringes with different phase shifts into a set of three phase-shifted high-modulation fringes, which includes two steps of training and testing. A three-step phase-shifting algorithm is used for FMEM due to its high-speed [47]. Two low-modulation fringes with respective phase shift amounts 0 and $2/3\pi$ are captured under a small camera exposure and selected as the input. A set of three phase-shifted high-modulation is captured under a large camera exposure (i.e., fringes have the highest fringe modulation and without intensity saturation) and selected as the ground-truth. In the training step, FMENet is trained to minimize the loss function (i.e., the difference between the output and the ground-truth) [26]. In the testing step, FMENet outputs a set of three phase-shifted high-modulation fringes with the input of two low-modulation fringes according to the trained model.

Figure 2 illustrates the schematic of FMEM. The input, output and ground-truth fringes are denoted by $I_n, n=1,2$, $I_n^{out}, n=1,2,3$ and $I_n^{gt},n=1,2,3$, respectively. By substituting $I_n$ with $I_n^{out}$ in Eq. (4), the desired phase of $\varphi ^{out}$ is calculated. The desired phase is wrapped in a range of $(-\pi ,\pi ]$, and the gray code-based method is used to retrieve the absolute phase [48,49]. Because of the projector defocusing and the camera discrete sampling, the random noise generates unwrapping errors along $2\pi$ phase jumps, which can be effectively removed by introducing a median filter, even for the captured fringes with a low fringe modulation [5]. The 3-D shape can be reconstructed by combining the absolute phase with calibrated parameters [50]. The loss function of FMENet is expressed as

$$Loss(\theta_1)=\frac{1}{m}\sum_{n=1}^{3}\left\|I_n^{out}-I_n^{gt}\right\|^2,$$
where $\theta _1$ denotes the parameter space in the network that includes the weights, bias and convolutional kernels [32], $m$ denotes the number of pixels and $\left \| \cdot \right \|^2$ is $\ell _2$-norm.

 figure: Fig. 2.

Fig. 2. Schematic diagram of the proposed FMEM.

Download Full Size | PDF

The structure of FMENet is illustrated in Fig. 3, where the fringe resolution of 512$\times$512 is selected. FMENet consists of a U-shaped structure [51,52] and a residual structure [26,53], which increase the computing efficiency of the network and achieve better restoration of image details. The U-shaped structure consists of operations: convolution (Conv) [51] and batch normalization (BN) [54]; Conv, BN and Dropout [55] by three times; Conv, BN and rectified linear unit (Relu) [56]. The input fringes are downsampled and upsampled spatially, which achieve the extraction of high-level features and the recovery of detail features. Meanwhile, the skip connection is used to integrate low-level information with high-level information to retrieve missing feature information, which realizes fully mining of pixel-level information and semantic-level features [52]. The feature information is fed into the residual structure, which includes the operations of Conv, three residual block [53] and Conv. The residual structure alleviates the problem of gradient disappearance caused by increasing depth in deep neural networks and further restores local details in the outputted fringes. FMENet is trained by the adaptive moment estimation (ADAM) optimizor [57]. We set a batch size of 8 and start with a learning rate of 5$e^{-4}$, which is divided by a factor of 2 each time (usually every 200 epochs) when the training error becomes stagnant.

 figure: Fig. 3.

Fig. 3. Network structure of the designed FMENet.

Download Full Size | PDF

4. Experiment analysis

We first verify the designed FMENet, and then compare the proposed FMEM with traditional methods. A high-speed three-step defocusing FPP is used [46,58], and the defocusing level is properly selected to remove high-order harmonics [59]. The FPP system includes a projector (DLP6500, Texas Instruments) with a resolution of $1920\times 1080$, a CMOS camera (Basler acA800-510um) with a resolution of $400\times 400$ and a lens of 12 mm focal length. The distance between our system and the measured objects is about 1.5 m. The fringes are designed with a fringe period of 30 pixels and a fringe frequency of 64, according to the projector’s resolution of 1920. FMENet is implemented by using Python and the framework of Pytorch on a PC with Intel Core i9-7900X CPU (3.30 GHz), 32 GB of RAM, and the GeForce GTX Titan RTX (NVIDIA).

The training set, the validation set and the testing set of FMENet contain 100 scenes, 20 scenes, and 20 scenes, respectively. For simplicity, fringes with a low modulation are captured by using a small camera exposure. For each scenes, seven sets of low-modulation phase-shifted fringes are captured under seven relatively small camera exposures (i.e., 250 us, 350 us, 450 us, 550 us, 750 us, 950 us, and 1,150 us), and a set of high-modulation fringes are captured under a large camera exposure (i.e., 3,500 us). FMENet uses two low-modulation fringes with respective phase shift amounts 0 and 2/3$\pi$ as the input and a set of three phase-shifted high-modulation fringes as the ground-truth. In addition, we emphasize that all the experimental results are obtained from the testing set.

In our datasets, the collected samples are independent and identically distributed, and similar results can be obtained in the training set, validation set and testing set, which can prove that our datasets contain sufficient samples and have extensively covered the sample space in our experiment environment. It should be noted that, the proposed FMEM requires additional samples for different measurement systems due to its data dependent.

4.1 Verification of FMENet

By randomly selecting one scene from the testing set, the experimental result is obtained and shown in Fig. 4. The fringes, captured under seven relatively small camera exposures, outputted by FMENet, and captured under the large camera exposure of 3,500 us are shown in Figs. 4(a)–4(c), respectively. In details, the first phase-shifted fringe of each set is selected, and the intensity of the 158th row is plotted in Fig. 4(d). FMENet outputted fringes are clear and bright as fringes captured under the large exposure. Fringe modulations of the inputted seven sets are provided in Table 1, where the modulations of inputted fringes and outputted fringes are denoted by $b^{in}$ and $b^{out}$, respectively. The mean fringe modulation of the whole image size is calculated to describe the fringe modulation, because different pixels in the captured fringes may have different fringe modulations. The fringe modulations of inputted seven sets range from 3.53 to 17.59, all the fringe modulations of outputted seven sets are above 53, and the fringe modulation of the ground-truth is 54.86. FMENet enhances the fringe modulation to the value close to the desired modulation of the ground-truth and performs consistently well for other scenes in testing set.

 figure: Fig. 4.

Fig. 4. Experimental results of FMENet. (a) The inputted fringes by FMENet. (b) The outputted fringes by FMENet. (c) The ground-truth. (d) The intensity of the 158th row of first phase-shifted fringes.

Download Full Size | PDF

Tables Icon

Table 1. Fringe modulations under seven relatively small camera exposures.

4.2 Comparison between the proposed FMEM and traditional methods

The proposed FMEM is compared with the three-step FPP (i.e., 3-step), windowed Fourier filtering (i.e., WFF) [24] and fifteen-step FPP (i.e., 15-step) [60]. The ground-truth phase is calculated from fifteen phase-shifted high-modulation fringes. By subtracting the actual phase from the ground-truth phase, the mean absolute value of the phase error is provided in Table 2, where $\Delta \Phi ^{3-step}$, $\Delta \Phi ^{FMEM}$, $\Delta \Phi ^{WFF}$ and $\Delta \Phi ^{15-step}$ denote the phase errors of 3-step, FMEM, WFF and 15-step, respectively. FMEM performs slightly better than WFF and obviously better than 15-step when the fringe modulation is very low (i.e., ranges from 3.53 to 8.16), and reduces the phase error to around 0.02 rad for different fringe modulations. Although 15-step performs similar to FMEM when the fringe modulation is higher than 8.16, it requires fifteen fringes, which obviously reduces the measurement speed compared with FMEM using two fringes.

The phase error distributions of two scenes (i.e., 250 us and 750 us) are shown in Fig. 5(a) and Fig. 6(a), respectively. For clarity, two areas are enlarged and shown in Fig. 5(b) and Fig. 6(b), respectively. As shown in Fig. 5, the phase error of FMEM is obviously smaller than that of 15-step, and also smaller than that of WFF in areas with relatively complex surface. The phase error shown in Fig. 6 is generally smaller than that in Fig. 5, because a larger camera exposure results in a smaller phase error.

 figure: Fig. 5.

Fig. 5. Phase error distributions of the scene captured under the camera exposure of 250 us. (a) The phase error. (b) The details of the object’s eyes and mouth.

Download Full Size | PDF

 figure: Fig. 6.

Fig. 6. Phase error distributions of the scene captured under the camera exposure of 750 us. (a) The phase error. (b) The details of the object’s eyes and mouth.

Download Full Size | PDF

Tables Icon

Table 2. Phase error of different methods.

The absolute phases of the above two scenes are retrieved from the obtained phases with the assistance of gray code-based patterns [61], and then combined with system calibrated parameters to reconstruct the 3-D shapes [62], as shown in Fig. 7(a) and Fig. 8(a), respectively, where the first, second, third, forth and fifth columns represent the 3-D shapes calculated from the phases obtained by the ground-truth, 3-step, FMEM, WFF and 15-step, respectively. For clarity, two areas are enlarged and shown in Figs. 7(b)–7(c) and Figs. 8(b)–8(c), respectively. The measurement errors of the 126th row of the four methods are selected and plotted in Fig. 7(d) and Fig. 8(d), respectively. For clarity, a dynamic measurement scene (i.e., fringes captured under different exposures and their resulting 3-D shapes) is provided in Visualization 1.

 figure: Fig. 7.

Fig. 7. 3-D shapes of the scene captured under the camera exposure of 250 us. (a) The 3-D shapes. (b) The details of the object’s eyes. (c) The details of the object’s mouth. (d) The measurement error of the 126th row.

Download Full Size | PDF

 figure: Fig. 8.

Fig. 8. 3-D shapes of the scene captured under the camera exposure of 750 us. (a) The 3-D shapes. (b) The details of the object’s eyes. (c) The details of the object’s mouth. (d) The measurement error of the 126th row.

Download Full Size | PDF

As shown in Fig. 7, the resulting 3-D shapes of FMEM and WFF are smoother than the shape of 15-step, which contains speckles caused by the larger phase error. FMEM preserves more 3-D details than WFF around the outline of the eyes, the underside of the nose, the mouth and the whiskers, etc, and achieves accurate 3-D shapes similar to the 3-D shape of the ground-truth. The measurement error of 3-step gives a mean value of 0.93 mm, which can be reduced to 0.23 mm, 0.26 mm and 0.40 mm by the proposed FMEM, WFF, and 15-step, respectively. As shown in Fig. 8, FMEM also performs well by obtaining the smooth 3-D shape similar to the 3-D shape of the ground-truth. Due to a larger camera exposure, WFF preserves more 3-D details and 15-step generates less speckles compared with Fig. 7. The measurement error of 3-step is 0.41 mm, which can be reduced to about 0.17 mm by using FMEM, WFF and 15-step.

The calculation times of 3-step, FMEM, WFF and 15-step are respective 0.9 ms, 79.5 ms, 231.1 ms and 1.8 ms when these methods are run on GPU. FMEM is obviously more time consuming than 3-step and 15-step, but less time consuming than WFF. However, FMEM achieves similar accuracy as 15-step by only using two fringes, and the 3-D shape can be reconstructed offline, which is important for measuring dynamic objects.

4.3 3-D reconstruction for scenes with un-uniform reflectivity

One scene containing two objects with un-uniform reflectivity are reconstructed and shown in Fig. 9. In Fig. 9(a), the first, second and third columns are fringes captured under the camera exposure of 4,000 us and 50,000 us, and fringes outputted by FMENet, respectively. It is difficult to capture high-modulation fringes for the two objects due to their un-uniform reflectivity. A small camera exposure causes low-modulation fringes of the right object as shown in the first column, and a large camera exposure causes saturated fringes of the left object as shown in the second column. FMEM outputs the desired high-modulation fringes as shown in the third column. The fringe modulations of the left and right objects in the first column are respective 47.99 and 8.77, and they are respective 13.35 and 77.11 in the second column. FMEM enhances the fringe modulations for the two objects to 47.35 and 80.97, respectively. Figure 9(b) is the resulting 3-D shapes. The low-modulation fringes and saturated fringes introduce speckles for the right object and large holes for the left object in 3-D shape, respectively. FMEM obtains the accurate 3-D shape of the scene with un-uniform reflectivity.

 figure: Fig. 9.

Fig. 9. 3-D reconstruction for the scene with un-uniform reflectivity. (a)The fringes. (b) The 3-D shapes.

Download Full Size | PDF

5. Conclusion

In this paper, a fringe modulation enhancement method (i.e., FMEM) is proposed by designing a fringe modulation enhancement convolutional neural network (i.e., FMENet). Two low-modulation fringes with different phase shifts are transformed into a set of three phase-shifted high-modulation fringes by using FMENet. The desired results can be obtained on sufficient samples, which are reliable and repeatable. FMEM is verified on scenes captured under different low camera exposures and scenes containing un-uniform reflectivity, which enables to reconstruct 3-D shapes from only two low-modulation fringes and achieves high accuracy similar to traditional phase-shifting algorithm with a large number of phase steps (e.g. 15-step).

Funding

Jiangsu Provincial Key Research and Development Program (BE2018126); Fundamental Research Funds for the Central Universities (30920031101); National Natural Science Foundation of China (61727802, 61971227).

Disclosures

The authors declare no conflicts of interest.

References

1. S. S. Gorthi and P. Rastogi, “Fringe projection techniques: Whither we are?” Opt. Lasers Eng. 48(2), 133–140 (2010). [CrossRef]  

2. X. Liu, X. Peng, H. Chen, D. He, and B. Z. Gao, “Strategy for automatic and complete three-dimensional optical digitization,” Opt. Lett. 37(15), 3126–3128 (2012). [CrossRef]  

3. Z. Wu, C. Zuo, W. Guo, T. Tao, and Q. Zhang, “High-speed three-dimensional shape measurement based on cyclic complementary gray-code light,” Opt. Express 27(2), 1283–1297 (2019). [CrossRef]  

4. F. Chen, G. M. Brown, and M. Song, “Overview of 3-D shape measurement using optical methods,” Opt. Eng. 39(1), 10–22 (2000). [CrossRef]  

5. D. Zheng, F. Da, Q. Kemao, and H. S. Seah, “Phase-shifting profilometry combined with gray-code patterns projection: unwrapping error removal by an adaptive median filter,” Opt. Express 25(5), 4700–4713 (2017). [CrossRef]  

6. X. Su and Q. Zhang, “Dynamic 3-d shape measurement method: A review,” Opt. Lasers Eng. 48(2), 191–204 (2010). [CrossRef]  

7. Y. An, J.-S. Hyun, and S. Zhang, “Pixel-wise absolute phase unwrapping using geometric constraints of structured light system,” Opt. Express 24(16), 18445–18459 (2016). [CrossRef]  

8. J. Zhong and J. Weng, “Spatial carrier-fringe pattern analysis by means of wavelet transform: wavelet transform profilometry,” Appl. Opt. 43(26), 4993–4998 (2004). [CrossRef]  

9. X. Su and W. Chen, “Fourier transform profilometry:: a review,” Opt. Lasers Eng. 35(5), 263–284 (2001). [CrossRef]  

10. Q. Zhang and X. Su, “High-speed optical measurement for the drumhead vibration,” Opt. Express 13(8), 3110–3116 (2005). [CrossRef]  

11. S. Zhang and P. S. Huang, “Phase error compensation for a 3-D shape measurement system based on the phase-shifting method,” Opt. Eng. 46, 60000E (2007). [CrossRef]  

12. H. Guo, H. He, and M. Chen, “Gamma correction for digital fringe projection profilometry,” Appl. Opt. 43(14), 2906–2914 (2004). [CrossRef]  

13. F. Lü, S. Xing, and H. Guo, “Self-correction of projector nonlinearity in phase-shifting fringe projection profilometry,” Appl. Opt. 56(25), 7204–7216 (2017). [CrossRef]  

14. C. Zuo, S. Feng, L. Huang, T. Tao, W. Yin, and Q. Chen, “Phase shifting algorithms for fringe projection profilometry: A review,” Opt. Lasers Eng. 109, 23–59 (2018). [CrossRef]  

15. Y. Wu, H. Yue, J. Yi, M. Li, and Y. Liu, “Phase error analysis and reduction in phase measuring deflectometry,” Opt. Eng. 54(6), 064103 (2015). [CrossRef]  

16. C. Rathjen, “Statistical properties of phase-shift algorithms,” J. Opt. Soc. Am. A 12(9), 1997–2008 (1995). [CrossRef]  

17. H. Jiang, H. Zhao, and X. Li, “High dynamic range fringe acquisition: A novel 3-d scanning technique for high-reflective surfaces,” Opt. Lasers Eng. 50(10), 1484–1493 (2012). [CrossRef]  

18. S. Feng, Y. Zhang, Q. Chen, C. Zuo, R. Li, and G. Shen, “General solution for high dynamic range three-dimensional shape measurement using the fringe projection technique,” Opt. Lasers Eng. 59, 56–71 (2014). [CrossRef]  

19. T. Yang, G. Zhang, H. Li, Z. Zhang, and X. Zhou, “Theoretical proof of parameter optimization for sinusoidal fringe projection profilometry,” Opt. Lasers Eng. 123, 37–44 (2019). [CrossRef]  

20. S. Feng, L. Zhang, C. Zuo, T. Tao, Q. Chen, and G. Gu, “High dynamic range 3d measurements with fringe projection profilometry: a review,” Meas. Sci. Technol. 29(12), 122001 (2018). [CrossRef]  

21. W. Yin, S. Feng, T. Tao, L. Huang, M. Trusiak, Q. Chen, and C. Zuo, “High-speed 3d shape measurement using the optimized composite fringe patterns and stereo-assisted structured light system,” Opt. Express 27(3), 2411–2431 (2019). [CrossRef]  

22. S. Zhang, “High-speed 3d shape measurement with structured light methods: A review,” Opt. Lasers Eng. 106, 119–131 (2018). [CrossRef]  

23. Q. Kemao, “Windowed fourier transform for fringe pattern analysis,” Appl. Opt. 43(13), 2695–2702 (2004). [CrossRef]  

24. Q. Kemao, H. Wang, and W. Gao, “Windowed fourier transform for fringe pattern analysis: theoretical analyses,” Appl. Opt. 47(29), 5408–5419 (2008). [CrossRef]  

25. B. Lin, S. Fu, C. Zhang, F. Wang, and Y. Li, “Optical fringe patterns filtering based on multi-stage convolution neural network,” Opt. Lasers Eng. 126, 105853 (2020). [CrossRef]  

26. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2016), pp. 770–778.

27. C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution using deep convolutional networks,” IEEE Trans. Pattern Anal. Mach. Intell. 38(2), 295–307 (2016). [CrossRef]  

28. F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2015), pp 815–823.

29. D. Ciregan, U. Meier, and J. Schmidhuber, “Multi-column deep neural networks for image classification,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, (2012), pp. 3642–3649.

30. K. G. Lore, A. Akintayo, and S. Sarkar, “Llnet: A deep autoencoder approach to natural low-light image enhancement,” Pattern Recognit. 61, 650–662 (2017). [CrossRef]  

31. H. Yu, X. Chen, Z. Zhang, C. Zuo, Y. Zhang, D. Zheng, and J. Han, “Dynamic 3-d measurement based on fringe-to-fringe transformation using deep learning,” Opt. Express 28(7), 9405–9418 (2020). [CrossRef]  

32. S. Feng, C. Zuo, W. Yin, G. Gu, and Q. Chen, “Micro deep learning profilometry for high-speed 3d surface imaging,” Opt. Lasers Eng. 121, 416–427 (2019). [CrossRef]  

33. S. Feng, Q. Chen, G. Gu, T. Tao, L. Zhang, Y. Hu, W. Yin, and C. Zuo, “Fringe pattern analysis using deep learning,” Adv. Photonics 1(02), 1–7 (2019). [CrossRef]  

34. H. Nguyen, N. Dunne, H. Li, Y. Wang, and Z. Wang, “Real-time 3d shape measurement using 3lcd projection and deep machine learning,” Appl. Opt. 58(26), 7100–7109 (2019). [CrossRef]  

35. K. Yan, Y. Yu, C. Huang, L. Sui, K. Qian, and A. Asundi, “Fringe pattern denoising based on deep learning,” Opt. Commun. 437, 148–152 (2019). [CrossRef]  

36. K. Yan, Y. Yu, T. Sun, A. Asundi, and Q. Kemao, “Wrapped phase denoising using convolutional neural networks,” Opt. Lasers Eng. 128, 105999 (2020). [CrossRef]  

37. S. Zhang, D. V. D. Weide, and J. Oliver, “Superfast phase-shifting method for 3-d shape measurement,” Opt. Express 18(9), 9684–9689 (2010). [CrossRef]  

38. S. Zhang, “Recent progresses on real-time 3d shape measurement using digital fringe projection techniques,” Opt. Lasers Eng. 48(2), 149–158 (2010). [CrossRef]  

39. J. Li, L. G. Hassebrook, and C. Guan, “Optimized two-frequency phase-measuring-profilometry light-sensor temporal-noise sensitivity,” J. Opt. Soc. Am. A 20(1), 106–115 (2003). [CrossRef]  

40. S. Zhang and S.-T. Yau, “High dynamic range scanning technique,” Opt. Eng. 48, 70660A (2009). [CrossRef]  

41. D. Zheng, Q. Kemao, J. Han, J. Wang, H. Yu, and L. Bai, “High-speed phase-shifting profilometry under fluorescent light,” Opt. Lasers Eng. 128, 106033 (2020). [CrossRef]  

42. C. Zuo, Q. Chen, G. Gu, S. Feng, and F. Feng, “High-speed three-dimensional profilometry for multiple objects with complex shapes,” Opt. Express 20(17), 19493–19510 (2012). [CrossRef]  

43. C. Chen, Y. Wan, and Y. Cao, “Instability of projection light source and real-time phase error correction method for phase-shifting profilometry,” Opt. Express 26(4), 4258–4270 (2018). [CrossRef]  

44. Cai Rongtai, Wu Qingxiang, Shi Wenzao, Sun Honghai, Wu Yuanhao, and Wang Zichen, “Ccd performance model and noise control,” in 2011 International Conference on Image Analysis and Signal Processing, (2011), pp. 389–394.

45. O. Kupyn, V. Budzan, M. Mykhailych, D. Mishkin, and J. Matas, “Deblurgan: Blind motion deblurring using conditional adversarial networks,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2018), pp. 8183–8192.

46. D. Zheng, F. Da, Q. Kemao, and H. S. Seah, “Phase error analysis and compensation for phase shifting profilometry with projector defocusing,” Appl. Opt. 55(21), 5721–5728 (2016). [CrossRef]  

47. W. Yin, C. Zuo, S. Feng, T. Tao, Y. Hu, L. Huang, J. Ma, and Q. Chen, “High-speed three-dimensional shape measurement using geometry-constraint-based number-theoretical phase unwrapping,” Opt. Lasers Eng. 115, 21–31 (2019). [CrossRef]  

48. S. Zhang, “Absolute phase retrieval methods for digital fringe projection profilometry: A review,” Opt. Lasers Eng. 107, 28–37 (2018). [CrossRef]  

49. Z. Wu, W. Guo, Y. Li, Y. Liu, and Q. Zhang, “High-speed and high-efficiency three-dimensional shape measurement based on gray-coded light,” Photonics Res. 8(6), 819–829 (2020). [CrossRef]  

50. K. Liu, Y. Wang, D. L. Lau, Q. Hao, and L. G. Hassebrook, “Dual-frequency pattern scheme for high-speed 3-d shape measurement,” Opt. Express 18(5), 5229–5244 (2010). [CrossRef]  

51. E. Romera, J. M. Álvarez, L. M. Bergasa, and R. Arroyo, “Erfnet: Efficient residual factorized convnet for real-time semantic segmentation,” IEEE Trans. Intell. Transport. Syst. 19(1), 263–272 (2018). [CrossRef]  

52. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, eds. (Springer International Publishing, Cham, 2015), pp. 234–241.

53. J. Yu, Y. Fan, J. Yang, N. Xu, Z. Wang, X. Wang, and T. S. Huang, “Wide activation for efficient and accurate image super-resolution,” arXiv preprint arXiv:1808.08718 (2018).

54. S. Ioffe and C. Szegedy, “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in Proceedings of The 32nd International Conference on Machine Learning, (2015), 448–456.

55. N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A simple way to prevent neural networks from overfitting,” J. Mach. Learn. Res. 15, 1929–1958 (2014). [CrossRef]  

56. V. Nair and G. E. Hinton, “Rectified linear units improve restricted boltzmann machines,” in Proceedings of the 27th International Conference on International Conference on Machine Learning (Omnipress, Madison, WI, USA, 2010), ICML’10, pp. 807–814.

57. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980 (2014).

58. S. Lei and S. Zhang, “Flexible 3-d shape measurement using projector defocusing,” Opt. Lett. 34(20), 3080–3082 (2009). [CrossRef]  

59. S. Lei and S. Zhang, “Digital sinusoidal fringe pattern generation: Defocusing binary patterns vs focusing sinusoidal patterns,” Opt. Lasers Eng. 48(5), 561–569 (2010). [CrossRef]  

60. V. Srinivasan, H. C. Liu, and M. Halioua, “Automated phase-measuring profilometry of 3-d diffuse objects,” Appl. Opt. 23(18), 3105–3108 (1984). [CrossRef]  

61. S. Yu, J. Zhang, X. Yu, X. Sun, and H. Wu, “Unequal-period combination approach of gray code and phase-shifting for 3-d visual measurement,” Opt. Commun. 374, 97–106 (2016). [CrossRef]  

62. X. He, D. Zheng, Q. Kemao, and G. Christopoulos, “Quaternary gray-code phase unwrapping for binary fringe projection profilometry,” Opt. Lasers Eng. 121, 358–368 (2019). [CrossRef]  

Supplementary Material (1)

NameDescription
Visualization 1       A dynamic measurement scene (i.e., fringes captured under different exposures and their resulting 3-D shapes) is provided in Visualization 1.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (9)

Fig. 1.
Fig. 1. The phase error: (a) for different fringe modulations, and (b) phase steps.
Fig. 2.
Fig. 2. Schematic diagram of the proposed FMEM.
Fig. 3.
Fig. 3. Network structure of the designed FMENet.
Fig. 4.
Fig. 4. Experimental results of FMENet. (a) The inputted fringes by FMENet. (b) The outputted fringes by FMENet. (c) The ground-truth. (d) The intensity of the 158th row of first phase-shifted fringes.
Fig. 5.
Fig. 5. Phase error distributions of the scene captured under the camera exposure of 250 us. (a) The phase error. (b) The details of the object’s eyes and mouth.
Fig. 6.
Fig. 6. Phase error distributions of the scene captured under the camera exposure of 750 us. (a) The phase error. (b) The details of the object’s eyes and mouth.
Fig. 7.
Fig. 7. 3-D shapes of the scene captured under the camera exposure of 250 us. (a) The 3-D shapes. (b) The details of the object’s eyes. (c) The details of the object’s mouth. (d) The measurement error of the 126th row.
Fig. 8.
Fig. 8. 3-D shapes of the scene captured under the camera exposure of 750 us. (a) The 3-D shapes. (b) The details of the object’s eyes. (c) The details of the object’s mouth. (d) The measurement error of the 126th row.
Fig. 9.
Fig. 9. 3-D reconstruction for the scene with un-uniform reflectivity. (a)The fringes. (b) The 3-D shapes.

Tables (2)

Tables Icon

Table 1. Fringe modulations under seven relatively small camera exposures.

Tables Icon

Table 2. Phase error of different methods.

Equations (6)

Equations on this page are rendered with MathJax. Learn more.

I n ( x , y ) = a ( x , y ) + Δ a n ( x , y ) + b ( x , y ) cos [ φ ( x , y ) δ n ] , n = 1 , 2 , 3 , , N ,
b ( x , y ) = s t e α ( x , y ) b p ( x , y ) ,
b = 2 N ( n = 1 N I n sin δ n ) 2 + ( n = 1 N I n cos δ n ) 2 .
φ = tan 1 n = 1 N I n sin δ n n = 1 N I n cos δ n .
Δ φ = 2 N b n = 1 N sin ( δ n φ ) Δ a n ,
L o s s ( θ 1 ) = 1 m n = 1 3 I n o u t I n g t 2 ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.