Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Enhancement of underwater optical images based on background light estimation and improved adaptive transmission fusion

Open Access Open Access

Abstract

Underwater optical imaging technology plays a vital role in humans’ underwater activities. However, the serious quality degradation of underwater optical images hinders further development of such technology. This phenomenon is mainly caused by the absorption and scattering of light in the underwater medium. The blurred image formation model is widely used in the field of optical images and depends on two optical parameters: background light (BL) and the transmission map (TM). Therefore, we propose an underwater optical image enhancement method in the context of underwater optical image restoration and color correction. First, BL estimation based on the gray close operation, which can avoid the influence of white objects while accurately calculating BL, is proposed. Then, an improved adaptive transmission fusion (IATF) method is proposed, and the adjusted reversed saturation map (ARSM) method is applied to compensate for and refine the estimated TMs to obtain the final TMs. This paper also proposes a new underwater light attenuation prior (NULAP) method. Finally, to enhance color saturation and edge details, a statistical colorless slant correction fusion smoothing filter method is proposed. Experimental results demonstrate that the proposed method outperforms state-of-the-art methods for dehazing, color and detail enhancement, and (uneven) light intensity.

© 2021 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Underwater optical imaging technology plays an important role in underwater optics and ocean optics. It is an important tool for humans to understand the ocean, develop and utilize the ocean, and protect the ocean. This technology has been widely used in the underwater target recognition, underwater archaeology, submarine resource exploration, and other fields [1]. However, due to the internal optical characteristics of pure water and the optical attenuation caused by the medium (the scattering and absorption of solute molecules and suspended particles) in the water, the quality of underwater optical images is severely reduced. The poor visibility and color cast caused by the effects of underwater optical imaging reduce the ability to extract valuable features from underwater images for further processing [24]. Therefore, to obtain high-quality underwater optical images and improve research in the field of underwater optics, it is necessary to restore and enhance degraded underwater optical images. At present, the restoration and enhancement of underwater optical images have been studied from several perspectives, including the use of dedicated hardware, polarization filtering technology, and scene model estimation. An important approach is the methods using specialized hardware. For example, the divergent underwater lidar imaging (UWLI) system can accurately capture nanosecond-level fast gated images in highly turbid water [5], thereby significantly reducing the noise related to water scattering in underwater images. Unfortunately, these complex acquisition systems are very expensive and consume power.

A second approach consists of polarization-based methods. These approaches use several images of the same scene captured with different degrees of polarization, as obtained by rotating a polarizing filter fixed to the camera. For example, to suppress noise in pixels of distant objects with low medium transmittance, an adaptive filtering method has been proposed [6]. Although polarization filtering technology can be used to restore distant areas, it is not suitable for dynamic acquisition. Schechner and Karpel [7] proposed and successfully demonstrated a computer vision-based method to eliminate degradation factors in the underwater optical imaging process.

The third approach employs multiple images [8] or an approximation of the scene model [9]. Narasimhan and Nayar [8] presented a physics model that describes the appearances of scenes in uniform bad weather. A system for browsing casual outdoor photos combined with the existing georeferenced digital terrain model has been proposed [9]. Although it has achieved sufficient dehazing and view synthesis functions, the enhancement effect of the image is not ideal.

The fourth type of method takes advantage of the similarity of light propagation in haze and underwater areas. In this fusion method, results are obtained from two original blurred image inputs by applying white balance and contrast enhancement programs [10]. Ju et al. [11] proposed an improved atmospheric scattering model (IASM) to overcome the inherent limitations of the traditional atmospheric scattering model and combined it with the acceleration framework based on the Gauss-Laplace pyramid to increase the calculation speed. However, underwater imaging is even more challenging because the extinction resulting from scattering depends on the wavelength of light, i.e., on the color component.

Recently, researchers have proposed other algorithms. Peng and Cosman [12] proposed a depth estimation method for underwater scenes based on image blurriness and light absorption, which is employed to enhance underwater images. An image restoration method based on a dark channel prior (DCP) is proposed [13]. It assumes that the radiance of objects in a natural scene is small in at least one color component. Therefore, the area with low transmittance is defined as the area with the smallest color value. Liu et al. [14] developed wavefront engineering technology based on internal guide stars, which increased the speed of optical focusing inside the scattering medium by two orders of magnitude and focused the scattered light inside the dynamic scattering medium. By imaging the target, we demonstrated the first focusing of scattered light inside a dynamic scattering medium containing living tissue. Li et al. [15] proposed a real-time color correction deep learning method based on an underwater unsupervised generative confrontation network, called WaterGAN. Similarly, Li et al. [16] proposed a method of image scattering and classification based on deep neural networks. Although the scattering is eliminated and the color is enhanced, the edge details are not well maintained. Zeng et al. [17] proposed a quaternion principal component analysis network (QPCANet) that uses quaternion theory to extend PCANet for color image classification. In contrast to PCANet, QPCANet considers the spatial distribution information of the RGB channels in the color image and uses the quaternion field to represent the color image to ensure greater intraclass invariance. [18] and [19] proposed a fusion-based underwater optical image enhancement method. The fusion framework obtains input and weight metrics from the degraded version of the image and implements an effective edge-preserving noise reduction strategy to support the temporal coherence between adjacent frames. Khan and Akmeliawati [20] developed a neural network-based dynamic integral sliding mode control (NNDISMC) with an output differential observer for higher derivative estimation and neural networks to estimate nonlinear functions assumed to be unknown. This method has good robustness. The deep convolutional neural network (CNN) method is used to perform super-resolution analysis of underwater optical images of different bands [2124]. Li et al. [21] proposed an underwater image enhancement convolutional neural network (UWGAN) model based on the underwater optical scene prior and reconstructed a clear underwater image through the underwater scene of synthetic training data. Fabbri et al. [23] proposed a novel underwater optical image enhancement method based on a generative adversarial network (UGAN), whose purpose is to further improve the input of vision-driven behavior downstream of the custom pipeline. Zhu et al. [24] developed unpaired image-to-image translation using cycle-consistent generative adversarial networks (CycleGAN). Their goal was to use training on a set of image pairs to learn the mapping between input and output images.

For underwater optical images, it is worth mentioning that there are also some special enhancement methods. Song et al. [25] proposed an image enhancement method consisting of underwater optical image restoration and color correction. The method provided a reliable background light (BL) estimation statistical model and estimated the transmission map (TM) based on the new underwater dark channel prior (NUDCP), which dehazed the image and gave it a natural appearance. From the perspective of style level and feature level adaptation, Ye et al. [26] proposed an unsupervised style adaptive network (SAN) to address the problem of common learning. Although the problem of color distortion and low contrast is solved, the problem of the loss of edge details is not yet sufficiently resolved. Inspired by the morphology and function of the bony fish retina, Gao et al. [27] proposed an underwater optical image enhancement mechanism, which greatly improved image quality. From the perspective of deep learning networks, Guo et al. [28] proposed a multiscale densely generated confrontation network (GAN), which effectively preserves the edge details of optical images. Fu et al. [29] proposed a single underwater optical image enhancement method based on retinex, which effectively solved the problems of underexposure. Zhou et al. [30] proposed an Airy spiral phase filter to enhance the edges of isotropic and anisotropic images and verified the effectiveness of the enhancement method by using amplitude contrast and phase contrast objects. Li et al. [31] proposed an enhancement method for optical wavefront shaping, which achieves the final effect by obtaining hologram and edge information inversion decoding after the scattering medium. Xia et al. [32] attributed image degradation to the wideband loss of high spatial frequency information caused by the high refractive index and proposed a high spatial spectrum enhancement reconstruction method, which effectively enhanced the edge details of the image. Chen et al. [33] proposed an accelerated processing method based on a fully convolutional network to improve the dehazing effect and realism. Peng et al. [34] proposed a generalized dark channel prior (GDCP), which incorporated adaptive color correction into an image formation model. A dehazing method based on the principle of minimum information loss and a contrast enhancement algorithm based on a priori histogram distribution, which effectively improves the visual effect, has been proposed [35].

The above methods will improve underwater optical images to a certain extent in one or several of the following aspects, yet underwater image restoration is still challenging due to the unpredictable variation in vision properties in undersea environments. There is no method that can achieve these effects more comprehensively than current standards: dehazing, denoising, color correction, detail enhancement, and uneven lighting brightness processing, in particular, white objects in the image, floating particles in the foreground, artificial light in the background region, or dim background light. Many state-of-the-art methods cannot solve these problems well. To compensate for the problems of existing methods, this paper proposes an underwater optical image enhancement method based on background light estimation and improved adaptive transmission fusion. First, this paper proposes a BL estimation method based on the gray close operation. Then, an improved adaptive transmission fusion (IATF) method is proposed, and TMs based on the dark channel prior (DCP-W) and TMs based on the new underwater optical attenuation prior (NULAP) are calculated separately. Next, fusion obtains the rough TMs. Underwater light attenuation prior (ULAP) and adjusted reverse saturation map (ARSM) principles are further used to compensate for and refine TMs. Finally, this paper proposes a statistical colorless slant correction fusion smoothing filter method, which greatly improves the quality of underwater optical images.

The contributions of this paper are summarized as follows:

  • 1) This paper adopts a BL estimation method based on the gray close operation, which can not only estimate BL more accurately but also avoid the interference of white objects in the scene.
  • 2) The IATF algorithm is proposed, which applies the principles of ULAP and ARSM to make the estimated TMs more accurate and eliminate the influence of artificial light and spots.
  • 3) To improve the clarity of underwater optical images, a statistical colorless slant correction fusion smoothing filter method is proposed. Experiments have proven that this method not only achieves color correction but also enhances the edge details so that the experimental results have a good visual effect.
  • 4) In addition, in the IATF algorithm, we also proposed a NULAP method, which can effectively identify the distance from the camera to the target object in the scene and the distant background. Furthermore, the white (or light-colored) target in the foreground can be avoided as the background scene, which is conducive to a more accurate estimation of TMs. The underwater light propagation scene from [36] is shown in Fig. 1.

 figure: Fig. 1.

Fig. 1. Underwater light propagation scene. Entry into the water from the atmosphere results in the gradual attenuation of light. Point x represents the scene point closest to the camera.

Download Full Size | PDF

Reflected light reaching the camera through an object is shown in Fig. 1. Scattering is a phenomenon caused by light interaction with the medium in water and is divided into backscattering and forward scattering. Backscattering describes the scattering of ambient light according to the line of sight before finally propagating to the image plane, resulting in large reductions in scene contrast. Forward scattering occurs when part of the reflected light is transmitted at a small angle and easily causes image blur.

2. Related work

2.1 Background light estimation

A simplified image formation model (IFM), often used to approximate the propagation equation of underwater scattering ([36,37]), is described as

$${I^c}(x )= {J^c}(x ){t^c}(x )+ ({1 - {t^c}(x )} ){B^c},\,c \in \{{r,g,b} \},$$
$${J^c}(x )= {\rho ^c}(x )\cdot {B^c},$$

Substituting Eq. (2) into Eq. (1), we get:

$${I^c}(x )= {B^c}{\rho ^c}(x ){t^c}(x )+ ({1 - {t^c}(x )} ){B^c},$$
where c represents one of red, green, or blue color channels; ${I^c}$ is the observed image in color channel c at pixel x, ${B^c}$ represents the intensity of the BL; ${J^c}$ is restored radiance of the pixel point x in one channel c; ${t^c}(x )$ ${\in} ({0,1} )$, ${t^c}(x )$ represents the remaining light after the scene radiation reaches the camera (that is, after scattering or absorption); ${\rho ^c}(x )$ represents the reflectivity of light. The ${B^c}\; $ is selected as the brightest pixel or the average value of the top 0.1% brightest pixels in $DC{P_{rgb}}$[13].

The values of ${t^c}(x )$ for an entire image can form a TM, which describes the portion of the scene radiance that is not scattered or absorbed and reaches the camera. When the haze is uniform, ${t^c}(x )$ can be expressed as:

$${t^c}(x )= {e^{ - {\beta ^c}d(x )}} = {e^{ - {\beta ^c}{d_0}}} \cdot {e^{ - {\beta ^c}{d_n}}} = {K_c} \cdot t_n^c(x ),\,c \in \{{r,g,b} \},$$
where ${d_0}$ represents the distance from the camera to the nearest scene point, ${d_n}$ represents the distance from the nearest scene point to the farthest scene point, $d(x )$ is the scene depth, ${\beta ^c}$ represents the spectral volume attenuation coefficient, ${K_c}$ is a constant attenuation factor, and $t_n^c(x )$ describes relative transmission.
$${B^c} = {I^c}(argmax|\mathop {\max }\limits_{y\epsilon \varOmega } {I^r}(y )- \mathop {\max }\limits_{y\epsilon \varOmega } {I^{c^{\prime}}}(y )|),\; $$
where $c^{\prime}$ denotes one of the G-B channels, $\varOmega $ denotes a local patch, whose size is 9×9 pixels, unless specified otherwise.

2.2 Transmission map estimation

Many factors impact the image in the underwater environment, including dust-like particles in the medium that cause changes in the wavelength-dependent attenuation coefficient. He et al. [13] proposed the discrete cosine transform. It is believed that in most nonsky patches, at least one pixel RGB channel in the local patch has a very low intensity in the statistical prior of outdoor haze-free images:

$$J_{dark}^{rgb}(x )= \mathop {\min }\limits_{y \in {\varOmega ^{\min }}} \left\{ {\mathop {\min }\limits_c {J^c}(y )} \right\} = 0,\; $$

Applying the smallest filter to Eq. (1) and dividing it by ${B^c}$, we can get:

$$\mathop {\min }\limits_{y\epsilon \varOmega } \left\{ {\mathop {\min }\limits_c \frac{{{I^c}(y )}}{{{B^c}}}} \right\} = \mathop {\min}\limits_{y\epsilon \varOmega } \left\{ {\mathop {\min }\limits_c {J^c}(y )} \right\} + 1 - {t_{DCP}}(x ),\; $$

Putting (6) into (7), ${t_{DCP}}(x )$ can be calculated:

$${t_{DCP}}(x )= 1 - \mathop {\min }\limits_{y\epsilon \varOmega } \left\{ {\mathop {\min }\limits_c \frac{{{I^c}(y )}}{{{B^c}}}} \right\},\; $$

In underwater scenes, the red attenuation is the fastest, so we only consider defining the $DC{P_{gb}}$ of the G and B channels as underwater DCP (UDCP) [12,34], and then using ${I^c}(y )$ instead of ${I^{c^{\prime}}}(y )$, ${t_{UDCP}}(x )$ can be obtained. The TM is estimated by calculating the difference between the maximum intensities of the R channel and the G-B channel. The formula is as follows:

$$\left\{ {\begin{array}{l} {{D_{mip}}(x )= {\textrm{ma}}{\textrm{x}_{y \in \varOmega }}{I^r}(y )- {\textrm{ma}}{\textrm{x}_{y \in \varOmega }}{I^{r{\prime}}}(y )}\\ {{t_{MIP}}(x )= {D_{mip}}(x )+ 1 - max({{D_{mip}}(x )} )} \end{array}} \right.,\; $$

Unlike the above method, the scene depth is estimated by combining three depth maps, which include the maximum filter, maximum intensity, and image blur of the red channel:

$${d_n} = {\theta _b}[{{\theta_a}{d_D} + ({1 - {\theta_a}} ){d_R}} ]+ ({1 - {\theta_b}} ){d_B},\; $$
$${\theta _a} = s({av{g_c}({{B^c}} ),\,0.5} ),$$
$${\theta _b} = s({av{g_c}({{I^r}} ),\,0.2} ),$$
where ${\theta _a}$ and ${\theta _b}$ are determined by the sigmoid function. After the relative depth map is refined by the guided filtering method [38], the final scene depth is obtained by converting the relative distance into the actual distance. The TM of the red channel is calculated as follows:
$${t^r}(x )= {e^{ - {\beta ^r}{d_a}}},\; $$
where ${\beta ^r} \in \left( {\frac{1}{8},\frac{1}{5}} \right)$ following [39,40]. Taking into account the attenuation ratio of the G-B channel to the R channel, the TMs of the G-B channel are obtained [41].

3. Methods in this article

A flowchart of the methods used in this paper is shown in Fig. 2.

 figure: Fig. 2.

Fig. 2. Flowchart for the proposed method.

Download Full Size | PDF

3.1 Background light estimation based on the gray close operation

The background light ${B^c}$ is an important factor in determining the color and brightness of the underwater image. When the value of ${B^c}$ is too large, overexposure will occur, and if the value of ${B^c}$ is too small, the light will be too dark. We propose a background light estimation method based on gray close operation and can eliminate the interference of white objects. In Eq. (1), when ${\rho ^c}(x )\to 0$ and ${\rho ^c}(x )\to \infty $, ${B^c}$ can be calculated. The analysis of absorption or scattering characteristics shows that the color of the image is produced by absorbing light of a specific frequency, which comes from vertical light. This shows that for lighter or darker objects, at least one of the three channels has a very small reflection coefficient and is close to zero. Therefore, we take the minimum of the three channels on both sides of Eq. (3) at the same time:

$$\begin{aligned} {I_{min}^c}& = {\mathop {\min }\limits_{c \in \{{R,\; G,\; B} \}} {I^c}(x )}\\ {}& = {\mathop {\min }\limits_{c \in \{{R,\,G,\,B} \}} ({{B^c}{\rho^c}(x ){t^c}(x )+ ({1 - {t^c}(x )} ){B^c}} )}\\ {}& = {B_\textrm {min}^c\rho _\textrm {min}^c(x )t_\textrm {min}^c(x )+ B_\textrm {min}^c({1 - t_\textrm {min}^c(x )} ),\,} \end{aligned}$$

Among all the color channels, the red channel has the lowest residual value and BL. This shows $t_{\min}^c(x )= {t^r}(x )$, $B_{\min}^c(x )= {B^r}$. Replace $t_{\min}^c(x )$ and $B_{\min}^c(x )$ in Eq. (14) with ${t^r}(x )$ and ${B^r}$ :

$$I_{\min}^c = {B^r}\rho _{\min}^c(x ){t^r}(x )+ {B^r}({1 - {t^r}(x )} ),$$

After the above analysis, ${B^r}$ can be calculated when ${\rho ^c}(x )\to 0$ and ${\rho ^c}(x )\to \infty $. For scenes with white objects, there will be the opposite trend. This is because the minimum reflection coefficient of white objects $\rho _{\min}^{whitec}(x )\to 1$, which makes the assumption invalid. To obtain ${B^c}$ robustly, this paper adopts the gray-scale operation method to avoid the interference of white objects:

$$\hat{I}_{\min}^c(x )= {B^r}\hat{\rho }_{\min}^c(x ){t^r}(x )+ {B^r}({1 - {t^r}(x )} ),\,$$
where $\hat{I}_{\min}^c(x )$ and $\hat{\rho }_{\min}^c(x )$ are the result of the gray close operation of $I_{\min}^c$ and $\rho _{\min}^c(x )$. Combined with the DCP prior, if the size of the structural element appears to be larger than the size of the brightness, $\hat{\rho }_{\min}^c(x )\to 0$ will appear. Therefore, ${B^r}$ is expressed as:
$${B^{r{{\prime}}}} = \frac{{\hat{I}_{\min}^c}}{{1 - {t^r}}},$$

Taking into account the attenuation and correction of the three channels, the BL estimates of the other two channels are [12]:

$$\frac{{{B^k}}}{{{B^r}}} = \frac{{{\beta ^r}({m{\lambda^k} + i} )}}{{{\beta ^k}({m{\lambda^r} + i} )}},\,k \in \{{g,b} \}$$
where ${\lambda ^c}$, $c \in \{{r,g,b} \}$ represents the wavelength of the red, green, and blue channels, $m ={-} 0.00113$, and $i = 1.62517$.

Among them, ${\beta ^c}$ is expressed as [37]:

$${\beta ^c} = \left\{ {\begin{array}{c} {0.2\; c = r}\\ {\; 0.05\; c = b,\; }\\ {0.07\; c = g} \end{array}} \right.$$

Here, the wavelength values of the three channels are defined as 620 nm, 540 nm, and 450 nm.

3.2 Transmission map fusion estimation

In this section, we propose an IATF method. First, the dark channel prior-water (DCP-W) method [42] is used to estimate TMs1, a NULAP-based method is proposed to obtain TMs2, and finally, the fusion operation is performed. In this method, to avoid the erroneous estimation of the near-field light and the far-field light and to reduce the interference of artificial light, we use the ULAP and ARSM methods to estimate more accurate TMs.

3.2.1 Transmission estimation based on DCP-W

Here, we use the DCP-W method to standardize Eq. (1) and perform the minimization operation:

$$\mathop {\max }\limits_{\lambda \in \{{R,G,B} \}} ({\tilde{t}({x,\lambda } )} )= 1 - \min\left( {\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({1 - I({y,R} )} )}}{{{B^r}}},\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({I({y,G} )} )}}{{{B^g}}},\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({I({y,B} )} )}}{{{B^b}}}} \right)\,,$$
where $\tilde{t}({x,\lambda } )$ represents the assumption of the transmission coefficient graph under the constant local patch. Since the blue attenuation coefficient is the smallest, the left side of the above formula represents the transmission coefficient. The blue projection image of the first viewing angle is estimated as follows:
$$\widetilde {{t_1}}({x,B} )= 1 - {\min}\left( {\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({1 - I({y,R} )} )}}{{{B^r}}},\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({I({y,G} )} )}}{{{B^g}}},\frac{{\mathop {\min }\limits_{y \in \varOmega (x )} ({I({y,B} )} )}}{{{B^b}}}} \right),$$

The transmission coefficients of the green and red channels [42] are estimated as follows:

$$\widetilde {{t_1}}({x,G} )= {({{e^{ - \beta (B )d(x )}}} )^{\frac{{\beta (G )}}{{\beta (B )}}}} = \widetilde {{t_1}}{({x,B} )^{\frac{{\beta (G )}}{{\beta (B )}}}},$$
$$\widetilde {{t_1}}({x,R} )= {({{e^{ - \beta (B )d(x )}}} )^{\frac{{\beta (R )}}{{\beta (B )}}}} = \widetilde {{t_1}}{({x,B} )^{\frac{{\beta (R )}}{{\beta (B )}}}},$$

3.2.2 Transmission estimation based on NULAP

We found that the ULAP [18] method has obvious inaccuracy when processing and recognizing the close-range distance (the shortest distance from the camera to the object) and the long-range distance (the distance from the camera to the distant background) in the scene. Furthermore, when the color of the object is white, it is easy to regard the white object in the foreground as the background scene. Therefore, we propose a NULAP method. We know that the difference between the maximum value of G-B intensity (MVGB) and the value of R intensity (VR) has a strong relationship with the depth of the scene. We build a scene depth model to estimate the TM.

$$d(x )= {\omega _0} + {\omega _1}mvgb(x )+ {\omega _2}vr(x ),$$
where $d(x )$ represents the depth value of the scene at the pixel point x, $mvgb(x )$ represents MVGB, $vr(x )$ represents VR.

In order to learn the coefficients ${\omega _0}$, ${\omega _1}$ and ${\omega _2}$ accurately, we need relatively correct training data. We collected a large number of underwater images from the dataset in this paper, calculate the scene depth maps, and then manually select the most accurate 100 data sets as the final training data. A guided filter [38] is further used to refine the scene depth map and give the edge details a smoother effect.

To train the model, we take the ratio of training and testing dataset as 7:3 and use 10- fold cross-validation. The best learning result is ${\omega _0}$=0.53214831, ${\omega _1}$=0.51309828 and ${\omega _2}$=-0.91066193.

The scene depth map estimated above is the relative distance. Next, the absolute distance between the camera and the nearest scene point needs to be estimated. From the estimated ${B^c}$ and ${I^c}(x )$, the absolute distance ${d_0}$ is calculated:

$${d_0} = 1 - \mathop {\max }\limits_{x,c} \left( {\frac{{|{{B^c} - {I^c}(x )} |}}{{\overline {{B^c}} }}} \right),$$
where $\overline {{B^c}} = \textrm{max}({1 - {B^c},{B^c}} )$, and ${d_0} \in [{0,1} ]$. The larger the value is, the smaller the distance between the point and the camera. Therefore, combining Eq. (1), (24), and (25), we further obtain the actual scene depth ${d_a}(x )$:
$${d_a}(x )= {D_\infty } \times ({d(x )+ {d_0}} ),$$
where ${D_\infty }$ represents the constant of proportionality. The function is to convert the relative distance into the actual distance, which we set to 10. Through Eq. (13) and (26), the TM of the red channel can be obtained:
$$\widetilde {{t_2}}({x,R} )= {e^{ - \beta (R ){d_a}(x )}},$$

For the green and blue channels, we can also calculate:

$$\widetilde {{t_2}}({x,G} )= {({{e^{ - \beta (R ){d_a}(x )}}} )^{\frac{{\beta (G )}}{{\beta (R )}}}},$$
$$\widetilde {{t_2}}({x,B} )= {({{e^{ - \beta (R ){d_a}(x )}}} )^{\frac{{\beta (B )}}{{\beta (R )}}}},$$

3.2.3 Transmission fusion estimation

Image fusion refers to the process of creating a single composite image from multiple input images and further integrating the relevant information of two (or more) images of the same scene into one output image. Research on the existing image fusion methods and their applications show that the image output by the fusion method is of higher quality than any single input image. The benefits of image fusion are reduced uncertainty, improved reliability, expanded space and time coverage, and compacted representation of information. Therefore, the purpose of the IATF method proposed in this paper is to reduce the impact of unpredictable factors in the underwater environment on underwater optical images.

To reduce the influence of factors such as potential artificial light and speckles while maintaining better scene details during the image fusion process, we conducted a saliency process analysis and ARSM calculation.

$$\widetilde {t_i^s}({x,\lambda } )= \vartheta [{({\widetilde {{t_i}}({x,\lambda } )} )} ]- \textrm{mean}[{\widetilde {{t_i}}({x,\lambda } )} ],\,({i = 1,\,2} )\,,$$
where $\vartheta $ is a bilateral Filter [19,38], $\widetilde {{t_i}}({x,\lambda } )$ represents the transmission saliency map. Finally, the two TMs obtained above are fused into one TM:
$$\tilde{t}({x,\lambda } )= \mathop \sum \nolimits_{i = 1}^2 {w_i}({x,\lambda } )\widetilde {{t_i}}({x,\lambda } ),$$
where ${w_i}({x,\lambda } )$ represents the weight based on the transmission saliency map.
$${w_i}({x,\lambda } )= \frac{{\widetilde {t_i^s}({x,\lambda } )}}{{\mathop \sum \nolimits_{j = 1}^2 \widetilde {t_j^s}({x,\lambda } )}},$$

Next, we further solve the exposure problem through ARSM [24], including the influence of artificial light and uneven light.

Because the area with higher red channel intensity (AL) in the underwater blue-green image has lower saturation in the color space, we defined the chromaticity purity of the pixel [24] as follows:

$$\begin{aligned}Sat({{I^c}(x )} )&= \mathop \sum \nolimits_{i = 1}^M \mathop \sum \nolimits_{j = 1}^N [1 - \frac{{\min({{I^c}(x )} )}}{{\max({{I^c}(x )} )}}]\\Sat({{I^c}(x )} )&= 1,\,if\,\max ({{I^c}(x )} )= 0,\end{aligned}$$

A color loses its saturation upon the addition of white light, which contains energy at all wavelengths. Therefore, saturation reduction in a certain area of an image can be achieved by adding white light to the area. In an underwater image, the saturation of a scene without AL will be much greater than that of the artificially illuminated area. This phenomenon can be represented by the following RSM, where a high RSM value usually denotes the AL-affected area.

$$Sa{t^{rev}}(x )= 1 - Sat({{I^c}(x )} ),\; $$
where $Sa{t^{rev}}(x )$ is the RSM function, which optimizes the estimated refined transmission to reduce the influence of uneven light and artificial light. We introduced a fitting parameter $\mathrm{\lambda } \in [{0,\,1} ]$ as an effective scalar multiplier.
$$Sat_{adj}^{rev}(x )= \mathrm{\lambda }{\ast }Sa{t^{rev}}(x ),\; $$
where $Sat_{adj}^{rev}(x )$ represents the ARSM. To effectively improve light uniformity, $\mathrm{\lambda }$ was set to 0.7.

As shown in Fig. 3, transmission estimation is inaccurate when the light is uneven or artificial light is applied. Moreover, the comparison of RSM and ARSM processing of the original image revealed that the overall brightness of the image was improved by ARSM.

$$t_n^c(x )= \mathop \sum \nolimits_{i = 1}^M \mathop \sum \nolimits_{j = 1}^N \max({t_n^r({i,j} ),Sat_{adj}^{rev}({i,j} )} ),\; $$

According to ARSM and Eq. (36), transmission estimation was further modified to reduce the intensity of the light in certain areas while maintaining the intensity in other areas of the image.

 figure: Fig. 3.

Fig. 3. ARSM. (a) Original image. (b) Inaccurate transmission estimation. (c) RSM. (d) ARSM, setting λ = 0.7.

Download Full Size | PDF

3.3 Scene light recovery

We assumed that when ${J^c}({x = \alpha } )= 1$ or 0, there exists a brightest or darkest pixel, α, that satisfies the following formula:

$$q = \left\{ {\begin{array}{c} {{B^c},\; {B^c} > 1 - {B^c}}\\ {1 - {B^c},{B^c} < 1 - {B^c}} \end{array}} \right.,\; $$
$$\mathop {\max }\limits_{c \in \{{r,g,b} \}} \frac{{|{{J^c}(\alpha )- {B^c}} |}}{q} = 1,$$
${K_c}|{{J^c}(\alpha )- {B^c}} |$ for all pixels in the image can be estimated by $\textrm{max}|{{I^c}(x )- {B^c}} |$. q is a variable that represents the maximum value in ${B^c}$ and $1 - {B^c}$. Simplify the three attenuation factors into one factor (namely, ${K_c}\; \,$$K$) and substituting $K|{{J^c}(\alpha )- {B^c}} |= \textrm{max}|{{I^c}(x )- {B^c}} |$ into Eq. (33) returns the following [36]:
$$K = \mathop {\max }\limits_{c \in \{{r,g,b} \}} \frac{{\max|{{I^c}(x )- {B^c}} |}}{q},\; $$

The reliability of the attenuation ratio depends on accurate estimation of the ${B^c}$, whereas the transmission of other channels continues. We defined an adaptive attenuation ratio, $\delta _c^{\prime}$, to adjust the transmission using Eq. (1) and Eq. (4) to express the resulting saturation constraint of the restored image as follows:

$$0 \le \frac{{{I^c}(x )- {B^c}}}{{K \cdot \hat{t}_n^r{{(x )}^{\delta _c^{\prime}}}}} + {B^c} \le 1,\; $$

Leading to

$$\left\{ {\begin{array}{l} {\delta_c^{\prime} \ge \ln \left( {\frac{{{I^c}(x )- {B^c}}}{{K({1 - {B^c}} )}}} \right)/\ln ({\hat{t}_n^r(x )} )+ {\varepsilon_c},\; if\; {I^c}(x )> {B^c}\; }\\ {\delta_c^{\prime} \ge \ln \left( {\frac{{{B^c} - {I^c}(x )}}{{K \cdot {B^c}}}} \right)/\ln ({\hat{t}_n^r(x )} )+ {\varepsilon_c},\; \textrm{if}\; {I^c}(x )< {B^c}\; } \end{array}} \right.,$$
where $\delta _c^{\prime} \in [{{\delta_c},1} ]$. To improve the estimation accuracy, we used ${\varepsilon _c}$ to improve image contrast, ${\varepsilon _c}$ represents tolerance.

After obtaining BL and final transmission estimates, the scene brightness, $\,J$, is calculated as follows:

$${J^c}(x )= \frac{{{I^c}(x )- {B^c}}}{{\textrm{min}({\max ({K \cdot \hat{t}_n^r{{(x )}^{\delta_c^{\prime}}},{t_1}} ),\,{t_2}} )}} + {t_0}{B^c},\; c \in \{{r,g,b} \},\; $$
where ${t_0}$, ${t_1}$, and ${t_2}$ are constant. As previously described [36,43], we applied a ${t_0}$ value of $\frac{{2e}}{5}$, the value of ${t_1}$ to 0.2, and the value of ${t_2}$ to 0.9, which improved image contrast. Figure 4 shows the different results of the step-by-step implementation of the method in this paper.

 figure: Fig. 4.

Fig. 4. Examples of enhancing underwater images. (a) Original image. (b) The corresponding TM. (c) Restored image and (d) Enhanced image.

Download Full Size | PDF

Figure 4 fully shows the comparison of results between different stages of the method in this paper. First, the BL of the original image is obtained from Section 3.1. The BL of the three images from top to bottom are: (0.03, 0.02, 0.02), (0.04, 0.61, 0.89) and (0.06, 0.87, 1). The TMs are further calculated (in section 3.2) according to the BL obtained in Section 3.1, and the result is shown in Fig. 4(b). Then, the result of scene light restoration is obtained according to Section 3.3, as shown in Fig. 4(c). Finally, the result of underwater image enhancement is obtained through Section 3.4, as shown in Fig. 4(d).

3.4 Image enhancement

The image enhancement method proposed in this paper is a fusion smoothing filter method based on statistical colorless slant correction [28].

$$J_{max}^c = J_{mean}^c + zJ_{var}^c,$$
$$J_{min}^c = J_{mean}^c - zJ_{var}^c\,,$$
where $J_{mean}^c$ and $J_{var}^c\; $ are the mean value and square error in channel c, respectively. z is a parameter, aims to control the image dynamically, and the empirical range of z is between 2 and 5. The color-corrected image $J_{CR}^c$ is
$$J_{CR}^c = ({{J^c} - J_{min}^c} )/({J_{max}^c - J_{min}^c} ),$$

After the image color is corrected and enhanced, an edge preservation smoothing filter is applied. Inspired by the WLS filter [38] and global guided image filtering [44,45,46], the edge preservation smoothing filter formula is as follows:

$$\mathop {\min }\limits_\varphi \mathop \sum \nolimits_x [ {{{({\varphi (x )- {O^{\ast }}(x )} )}^2} + \vartheta ( {\frac{{{{( {\frac{{\partial \varphi (x )}}{{\partial x}}} )}^2}}}{{{{|{{V^h}(x )} |}^\theta } + \epsilon }} + \frac{{{{( {\frac{{\partial \varphi (x )}}{{\partial y}}} )}^2}}}{{{{|{{V^v}(x )} |}^\theta } + \epsilon }}} )} ],$$
where $\vartheta $, $\theta $, and $\epsilon $ are all constants. ${O^\ast }$ represents the output image, and defines $\textrm{V} = ({{V^h},{V^v}} )$ as the guiding vector field.

The edge preservation smoothing filter is an image that is smoothed and a vector field:

$${V^h}(x )= \frac{{\partial {O^{\ast }}(x )}}{{\partial x}};{V^v}(x )= \frac{{\partial {O^{\ast }}(x )}}{{\partial y}}\,,$$

Similarly, the matrix representation is as follows:

$${({\varphi - {O^{\ast }}} )^T}({\varphi - {O^{\ast }}} )+ \vartheta ({{\varphi^T}D_x^T{B_x}{D_x}\varphi + {\varphi^T}D_y^T{B_y}{D_y}\varphi } ),$$

The matrices ${D_x}$ and ${D_y}$ represent discrete differential operators and the matrices ${B_x}$ and ${B_y}$ are described as:

$${B_x} = \textrm{diag}\left\{ {\frac{1}{{{{|{{V^h}(x )} |}^\theta } + \epsilon }}} \right\},\,{B_y} = \textrm{diag}\left\{ {\frac{1}{{{{|{{V^v}(x )} |}^\theta } + \epsilon }}} \right\},$$
$$({I + \vartheta ({D_x^T{B_x}{D_x} + D_y^T{B_y}{D_y}} )} )\varphi = {O^{\ast }},$$

By using the fast separation method in [26,27], O* can be solved quickly. Through the above operations, underwater images can well achieve image dehazing, denoising, color correction, and edge detail enhancement, thereby achieving ideal visual effects.

4. Results and evaluation

To concretely prove the effectiveness of this method, we conduct experimental comparisons with several state-of-the-art methods, including the new underwater dark channel prior (NUDCP) [25], a fusion-based method [19], the histogram prior [35], CycleGAN [24], UGAN [23], and UWCNN [21]. The first three are nondeep learning methods, and the last three are deep learning methods. The experimental dataset contains 3000 underwater images with different scenes, and the sample dataset includes 100 underwater images (Fig. 5).

 figure: Fig. 5.

Fig. 5. Sample test image. From left to right and then top to bottom: “Image-1” through “Image-100”.

Download Full Size | PDF

In order to make the method in this paper more representative and efficient, this requires our data set to contain underwater images of various depths, various underwater environments, extreme conditions (uneven light, artificial light, etc.), and various entities. The experimental data set of this paper mainly comes from UIEBD [47], RUIE [48], DIV2K [49], Fish4Knowlege, and Google Images. Then, some of these images are divided into the following categories: dim images, hazy images, images with different scattering and viewing angles (low backscatter, high backscatter, forward-looking, and downward-looking), greenish and bluish images, color distortion images, images with uneven lighting, and images with artificial light. Therefore, the data set in this paper can represent the general distribution of underwater images. The results of the different methods and the corresponding reference images are shown in Figs. 712. We conduct qualitative and quantitative comparisons and complexity analyses of different datasets.

 figure: Fig. 6.

Fig. 6. Image comparison before and after enhancement using the proposed method. (a) Original image and the image (b) before and (c) after enhancement.

Download Full Size | PDF

 figure: Fig. 7.

Fig. 7. Comparison results for dim underwater images. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

 figure: Fig. 8.

Fig. 8. Comparison results for a hazy underwater image. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

 figure: Fig. 9.

Fig. 9. Comparison results of underwater images with different scattering and viewing angles. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

 figure: Fig. 10.

Fig. 10. Comparisons of bluish and greenish underwater images. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

 figure: Fig. 11.

Fig. 11. Comparison of the effects of color enhancement and detail preservation. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

 figure: Fig. 12.

Fig. 12. Examples of underwater image enhancement in extreme scenes with uneven lighting conditions and artificial light environments. From left to right are (a) the original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Download Full Size | PDF

4.1 Qualitative estimation

We show the comparison of the results of the enhancement method and the restoration method in this article in Fig. 6.

To further prove the enhancement effect of the statistical colorless slant correction fusion smoothing filter method proposed in this paper, we performed an ablation experiment to show the contrast effect before and after the proposed enhancement method. Figure 6(b) and Fig. 6(c) depict the experimental results before and after the enhancement method, respectively. Through the comparison of experimental results, it can be seen that before using the enhancement method proposed in this article, the color saturation of the image and the sharpness of edge details need to be improved. The effect of the enhancement method is more natural, and the brightness of the image (including the background light in the distance) has been significantly improved. Therefore, the statistical colorless slant correction fusion smoothing filter method proposed in this paper has good effects on color enhancement, edge detail preservation, and image brightness adjustment.

Next, we compared the experimental results in different scenes with those of other existing methods, as shown in Figs. 712. The scenes are dim light, haze, different scattering and observation angles, bluish and greenish, color distortion and edge detail loss, extreme scenes with uneven lighting conditions, and artificial light environments.

It can be seen from Fig. 7 that NUDCP [25] can improve the dim environment and the color saturation to a certain extent, but the effect of maintaining edge details and image clarity needs to be further improved. Fusion-based methods [19] can enhance the overall color of the image, making the picture more textured, but cannot present white objects in the scene. The histogram prior [35] can improve the overall brightness of the picture but does not achieve good results in terms of edge detail and overall clarity of the image. CycleGAN [24] achieves good results in terms of color saturation of the image, but the overall brightness of the image is low. The effect achieved by UGAN [23] is remarkable, achieving good results in image brightness, color enhancement, and clarity. UWCNN [21] improves the problems of dim light and color saturation, but the edge details of the image are not well preserved. The method proposed in this paper has great improvements in image contrast, color saturation, and sharpness.

Figure 8 represents a hazy underwater optical scene. It can be seen from the experimental results that the above methods can achieve the dehazing effect, but considering other factors, the fusion-based method [19] does not achieve good results in terms of color saturation, the histogram prior [35] does not achieve a good processing effect on the foreground and background light, and CycleGAN [24] has a good effect in terms of color saturation, but the visual effect of the image is still not ideal. UGAN [23] and UWCNN [21] achieve poor results in edge detail and sharpness.

The four images in Fig. 9 are low-backscatter, high-backscatter, forward-looking, and downward-looking images. NUDCP [25] achieves a good effect in terms of improving the scattering. Although the fusion-based method [19] and histogram prior [35] can eliminate the effects of scattering, they cannot improve the effect of scene images containing greenish hues. CycleGAN [24] makes the overall brightness darker while improving the degree of scattering. UGAN [23] and UWCNN [21] can enhance the saturation of colors but do not improve the clarity of the image very well. It is not difficult to see that the method proposed in this paper achieves the best visual effects in low-backscatter, high-backscatter, forward-looking, and downward-looking.

Figure 10 represents bluish and greenish underwater optical scenes. NUDCP [25] can solve the bluish and greenish problems very well but does not improve the clarity of the image very well when dealing with the blueness of the image. The fusion-based method [19] can significantly improve the clarity of the image, which is significantly better than the effect achieved by NUDCP [25]. The histogram prior [35] can improve the sharpness of the image, but the brightness is darker. CycleGAN [24] and UGAN [23] are not very effective in dealing with the problem of bluishness and greenishness, but they improve the edge detail and sharpness of the image to a certain extent. UWCNN [21] can well eliminate bluish and greenish problems and improve the clarity of the scene target but is prone to artifacts when dealing with distant background light.

Figure 11 represents an image of a scene with color distortion and loss of edge detail. The above methods can enhance the color and edge texture details to a certain extent. NUDCP [25] and the fusion-based method [19] do not achieve natural visual effects for objects appearing in dark areas. The histogram prior [35] does not achieve better edge detail preservation. CycleGAN [24], UGAN [23], and UWCNN [21] can enhance the color saturation and edge details, but the enhancement effect is not good for the local scene area under dim light.

Figure 12 represents extreme scenes with uneven lighting conditions and artificial light environments. NUDCP [25] can solve the interference of uneven light and artificial lighting but will also cause the loss of edge details and will produce an unclear image due to the interference of white objects in the scene. The fusion-based method [19] can eliminate the vignetting caused by artificial lighting, but objects in dark areas cannot be presented. The histogram prior [35], CycleGAN, and UGAN can solve the problem of unclearness in an uneven lighting environment but cannot improve the effect very well for scenes with artificial lighting. UWCNN [21] can effectively reduce the influences of uneven illumination and artificial light. However, when dealing with uneven light scenes, objects in dimly lit areas in the distance are prone to artifacts. In scenes with artificial light, it is not possible to achieve an enhancement effect in dark areas.

It can be seen from Figs. 712 that NUDCP [25] can improve the dim underwater light environment very well but cannot preserve the edge texture details of the image well. The fusion-based method [19], the histogram prior [35], and CycleGAN [24] can enhance the color and eliminate bluish and greenish phenomena but cannot solve the overall brightness problem well. UGAN [23] and UWCNN [21] can effectively maintain edge details and improve image clarity, but they cannot improve the visual effect of unevenly illuminated scenes. The underwater image enhancement method based on background light estimation and improved adaptive transmission fusion proposed in this paper can achieve better results in image dehazing, color enhancement, edge detail preservation, and uneven illumination processing.

4.2 Quantitative analysis

We use entropy (Entropy), a gradient average (GA), an underwater image quality metric (UIQM) [21,36], an underwater color image quality evaluation metric (UCIQE) [25,50], and a natural image quality evaluator (NIQE) [40] for quantitative analysis. The entropy of the image can be used to represent the characteristics of image statistics according to the average signal; the size of the entropy value intuitively defines the quality and clarity of an image. The AVG not only reflects the degree of image clarity but also indicates changes in edge texture details; larger AVG values result in decreased blur. The UCIQE metric is a linear combination of the chroma, saturation, and contrast of underwater images. The NIQE uses space domain natural scene statistics, where a small value represents better quality. The UCIQE and UIQM metrics are used for underwater image assessment. The UCIQE metric is designed specifically to quantify the nonuniform color cast, blurring, and low contrast that characterize underwater images. The UIQM is a linear combination of color, clarity, and contrast and comprises the following processes: underwater image chromaticity measurement (UICM), underwater image sharpness measurement (UISM) [27], and underwater image contrast measurement (UIConM) [12]. A larger UIQM value suggests better image enhancement performance. The UIQM is calculated as follows:

$$\textrm{UIQM}\, = \,\textrm{a}1\ast \textrm{UICM}\, + \,\textrm{a}2\ast \textrm{UISM}\, + \,\textrm{a}3\ast \textrm{UIConM},$$
where UICM is the colorfulness, UISM is the sharpness, and UIConM is the contrast measure. The parameters $\textrm{a}1$, $\textrm{a}2$, and $\textrm{a}3$ are weights, whose values are application dependent. We extract 50 randomly selected underwater images from the training data in this paper. These images are underwater images from different equipment, different depths, and angles. After calculation and experimental analysis [50,51], $\textrm{a}1\,$ is set to 0.0282, $\textrm{a}2$ is set to 0.2953, and $\textrm{a}3$ is set to 3.5753.

Twenty-four sample images (including all the images in this article, numbered 1, 2… 24) and 3000 underwater images are used to confirm the effectiveness of the proposed method. We randomly divide the 3000 underwater images into three groups (G1, G2, and G3) (Table 1). In Table 2, Avg1 represents the average value obtained from 24 images. Avg2 represents the average value of the corresponding G1, G2, and G3. Most of the results show that the proposed method returns better evaluations for underwater images, indicating its superiority in image dehazing, deblurring, edge detail preservation, and color enhancement.

Tables Icon

Table 1. UIQM COMPARISON BETWEEN METHODS

Tables Icon

Table 2. Average UIQM, UCIQE, and NIQE values of the original images in Fig. 5 and their enhanced versions from all methods

The experimental analysis results are shown in Table 1. In general, the three deep learning methods achieve better results than the three non-deep learning methods. Among the three non-deep learning methods, the evaluation index of NUDCP [25] is the best, indicating that the achieved effect is the best. Among the three deep learning methods, the evaluation index of the UWCNN [21] is the best, indicating that the achieved effect is the best. The evaluation index of the method proposed in this paper is better than those of all the comparison methods above, which further illustrates the effectiveness of the method. Table 2 shows the average UIQM, UCIQE, and NIQE values of the original images in Fig. 5 and their various enhanced versions. Our method outperforms the other methods.

4.3 Complexity analysis

The proposed method estimates each curve in a pixel-by-pixel manner. For a pixel with m × n dimensions, the complexity of the initial transmission estimation is O (m × n). According to previous descriptions of evaluation and analysis [35], the computational complexity of refined transmission is also O (m × n).

Here, we used an Intel Xeon E5-1630 (v.3.0) 3.7 GHz CPU and MATLAB R2015b for our analyses. The calculation times for NUDCP [25], the fusion-based method [19], the histogram prior [35], CycleGAN [24], UGAN [23], and UWCNN [21] were all significantly slower than that of the proposed method when processing underwater optical images at various resolutions. The comparison values are presented in Table 3.

Tables Icon

Table 3. Comparison of the Computation Times Required by the Methods (in Seconds)

Table 3 shows the running time of different comparison methods. As the image pixels increase, the running time will also increase. On the whole, the method proposed in this paper has the fastest overall operating efficiency as the value of pixels increases.

5. Discussion

This paper proposes an underwater optical image enhancement method based on background light estimation and improved adaptive transmission fusion. It has obvious effects on dehazing, color enhancement, edge detail preservation, and light processing in extreme environments (uneven light and artificial light). First, this paper proposes BL estimation based on the gray close operation. The IFM minimization operation combines the DCP prior theory and the size of different wavelengths to obtain the BL values of the three channels, which not only accurately estimates the BL value but also avoids the interference of white objects in the scene. When the minimum reflection coefficient of white objects $\rho _{min}^{whitec}(x )\to 1$, the method will be affected, so after the gray close operation, $\hat{\rho }_{min}^c(x )\to 0$ can result, thereby eliminating the interference of these factors.

Then, we propose the IATF method, which makes TM estimation more accurate than that with other existing methods. Our estimation method is compensated for and improved based on the existing methods [18,19], and [36]. The influence of some other interference factors is eliminated to achieve more accurate TM estimation. Specifically, our method is divided into the following steps: 1) Calculate $\widetilde {{t_1}}$ by adopting the method based on DCP-W. The IFM is normalized, and the blue TM is estimated by combining the sizes of the attenuation coefficients of the three colors. Then, the TM of the green and red channels is calculated through the transmission coefficient map [42]. 2) $\widetilde {{t_2}}$ is calculated by the NULAP method proposed in this paper. The scene depth model is established, and the relative distance is estimated from the scene depth map according to the attenuation ratio of the channel [39,40]. Then, the guided filter [38] is used to further refine the scene depth map to obtain the distance between the camera and the scene point. The actual distance $\; \widetilde {{t_2}}$ can be obtained by formulas (26), (27), and (13). 3) TM fusion is performed by formulas (31) and (32) to obtain rough TMs and is further compensated for by the ARSM method to eliminate the influence of artificial light and speckles to obtain accurate TMs after fusion. Finally, the underwater optical scene is restored based on the obtained ${B^c}$ and $t_n^c(x )$.

Finally, this paper proposes a statistical colorless slant correction fusion smoothing filter method, which aims to enhance the color, details, and brightness of restored underwater optical images. First, based on the statistical colorless slant correction [27], the color-corrected image $J_{CR}^c$ is calculated. Then, inspired by the WLS filter [38] and global guided image filtering [44], we introduced an edge-preserving smoothing filter formula and a vector matrix and generated an enhanced underwater optical scene image. Qualitative and quantitative analyses showed that our proposed method based on background light estimation and improved adaptive transmission fusion has obvious advantages in underwater optical image dehazing, edge detail preservation, color enhancement, and uneven light processing (including artificial light). The result is better than that of other existing methods.

6. Conclusions

Currently, humans are engaged in an increasing number of underwater activities. Underwater optics and imaging technology play an increasingly important role, but the absorption and scattering of light in water have seriously affected the development of this field. This paper proposes a method based on background light estimation and improved adaptive transmission fusion. First, BL estimation based on the gray close operation is proposed. The proposed method avoids the interference of white objects in underwater optical scenes. Then, we propose an image restoration method, IATF. The TM values estimated by the DCP-W and NULAP methods are fused to obtain a rough TM value, and then the ARSM method is further compensated for and improved to obtain an accurate TM value. In this way, the underwater optical image is restored. Finally, we propose a statistical colorless slant correction fusion smoothing filter method to enhance color and edge details. Qualitative and quantitative analyses indicate that the proposed method shows better color enhancement and uneven underwater light processing (including artificial light) than other existing methods.

Funding

National Natural Science Foundation of China (91746100); National Key Research and Development Program of China (2017YFC0804406).

Acknowledgments

We thank some teachers for their help in writing the paper; it was with their encouragement and guidance that we finally finished this paper. We also thank the anonymous reviewers for their critical comments on the manuscript.

Disclosures

The authors declare that there are no conflicts of interest related to this article.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. X. D. Qiu, F. S. Li, W. H. Zhang, Z. H. Zhu, and L. X. Chen, “Spiral phase contrast imaging in nonlinear optics: seeing phase objects using invisible illumination,” Optica 5, 208–212 (2018). [CrossRef]  

2. N. Vasilis, “Going deeper than microscopy: the optical imaging frontier in biology,” Nat. Methods 7, 603–614 (2010). [CrossRef]  

3. A. M. C. Aguirre, D. B. Conkey, J. D. Dove, H. Y. Ju, T. W. Murray, and R. Piestun, “High contrast three-dimensional photoacoustic imaging through scattering media by localized optical fluence enhancement,” Opt. Express 21, 26671–26676 (2013). [CrossRef]  

4. W. H. Liu, Z. Y. Xu, and L. Q. Yang, “SIMO detection schemes for underwater optical wireless communication under turbulence,” Photon. Res. 3(3), 48–53 (2015). [CrossRef]  

5. D.-M. He and G. G. L. Seet, “Divergent-beam LiDAR imaging in turbid water,” Opt. Lasers Eng. 41, 217–231 (2004). [CrossRef]  

6. Y. Y. Schechner and Y. Averbuch, “Regularized image recovery in scattering media,” IEEE Trans. Pattern Anal. Machine Intell 19(9), 1655–1660 (2007). [CrossRef]  

7. Y. Y. Schechner and N. Karpel, “Recovery of underwater visibility and structure by polarization analysis,” IEEE J. Ocean. Eng. 30(3), 570–587 (2005). [CrossRef]  

8. S. G. Narasimhan and S. K. Nayar, “Contrast restoration of weather degraded images,” IEEE Trans. Pattern Anal. Mach. Learn. 25(6), 713–724 (2003). [CrossRef]  

9. J. Kopf, B. Neubert, B. Chen, M. Cohen, D. Cohen-Or, O. Deussen, M. Uyttendaele, and D. Lischinski, “Deep photo: Model-based photograph enhancement and viewing,” ACM Trans. Graph. 27, 1557–7368 (2008). [CrossRef]  

10. C. O. Ancuti and C. Ancuti, “Single image dehazing by multi-scale fusion,” IEEE Trans. Image Process. 22(8), 3271–3282 (2013). [CrossRef]  

11. M. Y. Ju, Z. F. Gu, and D. Y. Zhang, “Single image haze removal based on the improved atmospheric scattering model,” Neurocomputing 260, 180–191 (2017). [CrossRef]  

12. Y.-T. Peng and P. C. Cosman, “Underwater image restoration based on image blurriness and light absorption,” IEEE Trans. Image Process. 26(4), 1579–1594 (2017). [CrossRef]  

13. K. He, J. Sun, and X. Tang, “Single image haze removal using dark channel prior,” IEEE Trans. Pattern Anal. Mach. Intell. 33(12), 2341–2353 (2011). [CrossRef]  

14. Y. Liu, P. Lai, C. Ma, X. Xu, A. A. Grabar, and L. V. Wang, “Optical focusing deep inside dynamic scattering media with near-infrared time-reversed ultrasonically encoded (TRUE) light,” Nat. Commun. 6, 5904 (2015). [CrossRef]  

15. J. Li, K. A. Skinner, R. M. Eustice, and M. Johnson-Roberson, “WaterGAN: Unsupervised Generative Network to Enable Real-time Color Correction of Monocular Underwater Images,” IEEE Rob. Automat. Lett. 3(1), 387–394 (2018). [CrossRef]  

16. Y. J. Li, H. M. Lu, J. R. Li, X. Li, Y. Li, and S. Serikawa, “Underwater image de-scattering and classification by deep neural network,” Computers & Electrical Engineering 54, 68–77 (2016). [CrossRef]  

17. R. Zeng, J. S. Wu, Z. H. Shao, Y. Chen, B. J. Chen, L. Senhadji, and H. Z. Shu, “Color image classification via quaternion principal component analysis network,” Neurocomputing 216, 416–428 (2016). [CrossRef]  

18. W. Song, Y. Wang, D. Huang, and D. Tjondronegoro, “A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration,” in Advances in Multimedia Information Processing—PCM, 678–688 (2018).

19. C. Ancuti, C. O. Ancuti, T. Haber, and P. Bekaert, “Enhancing underwater images and videos by fusion,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 81–88, Jun. 2012.

20. Q. Khan and R. Akmeliawati, “Neuro-adaptive dynamic integral sliding mode control design with output differentiation observer for uncertain higher order MIMO nonlinear systems,” Neurocomputing 226, 126–134 (2017). [CrossRef]  

21. C. Y. Li, S. Anwar, and F. Porikli, “Underwater scene prior inspired deep underwater image and video enhancement,” Pattern Recogn. 98, 107038 (2020). [CrossRef]  

22. X. Liu and M. Wang, “Super-Resolution of VIIRS-Measured Ocean Color Products Using Deep Convolutional Neural Network,” IEEE Trans. Geosci. Remote Sensing 59(1), 114–127 (2021). [CrossRef]  

23. C. Fabbri, M. J. Islam, and J. Sattar, “Enhancing Underwater Imagery Using Generative Adversarial Networks,” in 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 7159–7165, 2018.

24. J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2242–2251, Oct. 2017.

25. W. Song, Y. Wang, D. Huang, A. Liotta, and C. Perra, “Enhancement of Underwater Images with Statistical Model of Background Light and Optimization of Transmission Map,” IEEE Trans. Broadcast. 66(1), 153–169 (2020). [CrossRef]  

26. X. Ye, Z. Li, B. Sun, Z. Wang, R. Xu, and H. Fan, “Deep Joint Depth Estimation and Color Correction from Monocular Underwater Images Based on Unsupervised Adaptation Networks,” IEEE Trans. Circuits Syst. Video Technol. 30(11), 3995–4008 (2020). [CrossRef]  

27. S. -B. Gao, M. Zhang, Q. Zhao, X. -S. Zhang, and Y. -J. Li, “Underwater Image Enhancement Using Adaptive Retinal Mechanisms,” IEEE Trans. Image Process. 28(11), 5580–5595 (2019). [CrossRef]  

28. Y. Guo, H. Li, and P. Zhuang, “Underwater Image Enhancement Using a Multiscale Dense Generative Adversarial Network,” IEEE J. Oceanic Eng. 45(3), 862–870 (2020). [CrossRef]  

29. X. Fu, P. Zhuang, Y. Huang, Y. Liao, X.-P. Zhang, and X. Ding, “A retinex-based enhancing approach for single underwater image,” in Proc. IEEE Int. Conf. Image Process. (ICIP), 4572–4576, Oct. 2014.

30. Y. Zhou, S. Feng, S. Nie, J. Ma, and C. Yuan, “Image edge enhancement using Airy spiral phase filter,” Opt. Express 24, 25258–25268 (2016). [CrossRef]  

31. Z. H. Li, Z. P. Yu, H. Hui, H. H. Li, T. T. Zhong, H. L. Liu, and P. X. Lai, “Edge enhancement through scattering media enabled by optical wavefront shaping,” Photon. Res. 8, 954–962 (2020). [CrossRef]  

32. H. J. Xia, Y. Q. Wu, L. Zhang, Y. Y. Sun, Z. Y. Wang, and R. Z. Tai, “Great enhancement of image details with high fidelity in a scintillator imager using an optical coding method,” Photon. Res. 8, 1079–1085 (2020). [CrossRef]  

33. Q. Chen, J. Xu, and V. Koltun, “Fast image processing with fully convolutional networks,” in Proc. IEEE Int. Conf. Comput. Vis., 2516–2525, Oct. 2017.

34. Y. Peng, K. Cao, and P. C. Cosman, “Generalization of the Dark Channel Prior for Single Image Restoration,” IEEE Trans. Image Process. 27(6), 2856–2868 (2018). [CrossRef]  

35. C.-Y. Li, J.-C. Guo, R.-M. Cong, Y.-W. Pang, and B. Wang, “Underwater image enhancement by dehazing with minimum information loss and histogram distribution prior,” IEEE Trans. Image Process. 25(12), 5664–5677 (2016). [CrossRef]  

36. Y. Wang, H. Liu, and L.-P. Chau, “Single underwater image restoration using attenuation-curve prior,” Proc. IEEE Trans. Circuits Syst. 65(3), 992–1002 (2018). [CrossRef]  

37. Z. Liang, Y. Y. Wang, X. Y. Ding, Z. T. Mi, and X. P. Fu, “Single underwater image enhancement by attenuation map guided color correction and detail preserved dehazing,” Neurocomputing 425(15), 160–172 (2021). [CrossRef]  

38. K. He, J. Sun, and X. Tang, “Guided image filtering,” IEEE Trans. Pattern Anal. Mach. Intell. 35(6), 1397–1409 (2013). [CrossRef]  

39. J. Y. Chiang and Y.-C. Chen, “Underwater image enhancement by wavelength compensation and dehazing,” IEEE Trans. Image Process. 21(4), 1756–1769 (2012). [CrossRef]  

40. W. Chen, K. Gu, W. Lin, F. Yuan, and E. Cheng, “Statistical and Structural Information Backed Full-Reference Quality Measure of Compressed Sonar Images,” IEEE Transactions on Circuits and Systems for Video Technology 30(2), 334–348 (2020). [CrossRef]  

41. X. Zhao, T. Jin, and S. Qu, “Deriving inherent optical properties from background color and underwater image enhancement,” Ocean Eng. 94, 163–172 (2015). [CrossRef]  

42. H.-H. Chang, C.-Y. Cheng, and C.-C. Sung, “Single underwater image restoration based on depth estimation and transmission compensation,” IEEE J. Ocean. Eng. 44(1), 1130–1149 (2019). [CrossRef]  

43. Z. Yu, T. Li, S. Horng, Y. Pan, H. Wang, and Y. Jing, “An Iterative Locally Auto-Weighted Least Squares Method for Microarray Missing Value Estimation,” IEEE Trans. NanoBiosci. 16(1), 21–33 (2017). [CrossRef]  

44. Z. Li and J. Zheng, “Single Image De-Hazing Using Globally Guided Image Filtering,” IEEE Trans. Image Process. 27(1), 442–450 (2018). [CrossRef]  

45. H. Y. Lu, S. Q. Li, Q. G. Liu, and M. H. Zhang, “MF-LRTC: Multi-filters guided low-rank tensor coding for image restoration,” Neurocomputing 303, 88–102 (2018). [CrossRef]  

46. K. Liu and Y. Liang, “Underwater image enhancement method based on adaptive attenuation-curve prior,” Opt. Express 29(7), 10321–10345 (2021). [CrossRef]  

47. C. Li, C. Guo, W. Ren, R. Cong, J. Hou, S. Kwong, and D. Tao, “An Underwater Image Enhancement Benchmark Dataset and Beyond,” IEEE Trans. Image Process. 29, 4376–4389 (2020). [CrossRef]  

48. R. Liu, X. Fan, M. Zhu, M. Hou, and Z. Luo, “Real-World Underwater Enhancement: Challenges, Benchmarks, and Solutions Under Natural Light,” IEEE Transactions on Circuits and Systems for Video Technology 30(12), 4861–4875 (2020). [CrossRef]  

49. E. Agustsson and R. Timofte, “NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 1122–1131, 2017.

50. M. Yang and A. Sowmya, “An underwater color image quality evaluation metric,” IEEE Trans. Image Process. 24(12), 6062–6071 (2015). [CrossRef]  

51. K. Panetta, C. Gao, and S. Agaian, “Human-visual-system-inspired underwater image quality measures,” IEEE J. Ocean. Eng. 41(3), 541–551 (2016). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (12)

Fig. 1.
Fig. 1. Underwater light propagation scene. Entry into the water from the atmosphere results in the gradual attenuation of light. Point x represents the scene point closest to the camera.
Fig. 2.
Fig. 2. Flowchart for the proposed method.
Fig. 3.
Fig. 3. ARSM. (a) Original image. (b) Inaccurate transmission estimation. (c) RSM. (d) ARSM, setting λ = 0.7.
Fig. 4.
Fig. 4. Examples of enhancing underwater images. (a) Original image. (b) The corresponding TM. (c) Restored image and (d) Enhanced image.
Fig. 5.
Fig. 5. Sample test image. From left to right and then top to bottom: “Image-1” through “Image-100”.
Fig. 6.
Fig. 6. Image comparison before and after enhancement using the proposed method. (a) Original image and the image (b) before and (c) after enhancement.
Fig. 7.
Fig. 7. Comparison results for dim underwater images. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.
Fig. 8.
Fig. 8. Comparison results for a hazy underwater image. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.
Fig. 9.
Fig. 9. Comparison results of underwater images with different scattering and viewing angles. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.
Fig. 10.
Fig. 10. Comparisons of bluish and greenish underwater images. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.
Fig. 11.
Fig. 11. Comparison of the effects of color enhancement and detail preservation. From left to right are (a) original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.
Fig. 12.
Fig. 12. Examples of underwater image enhancement in extreme scenes with uneven lighting conditions and artificial light environments. From left to right are (a) the original images and the results generated by (b) NUDCP [25], (c) the fusion-based method [19], (d) the histogram prior [35], (e) CycleGAN [24], (f) UGAN [23], (g) UWCNN [21], and (h) the proposed method.

Tables (3)

Tables Icon

Table 1. UIQM COMPARISON BETWEEN METHODS

Tables Icon

Table 2. Average UIQM, UCIQE, and NIQE values of the original images in Fig. 5 and their enhanced versions from all methods

Tables Icon

Table 3. Comparison of the Computation Times Required by the Methods (in Seconds)

Equations (51)

Equations on this page are rendered with MathJax. Learn more.

I c ( x ) = J c ( x ) t c ( x ) + ( 1 t c ( x ) ) B c , c { r , g , b } ,
J c ( x ) = ρ c ( x ) B c ,
I c ( x ) = B c ρ c ( x ) t c ( x ) + ( 1 t c ( x ) ) B c ,
t c ( x ) = e β c d ( x ) = e β c d 0 e β c d n = K c t n c ( x ) , c { r , g , b } ,
B c = I c ( a r g m a x | max y ϵ Ω I r ( y ) max y ϵ Ω I c ( y ) | ) ,
J d a r k r g b ( x ) = min y Ω min { min c J c ( y ) } = 0 ,
min y ϵ Ω { min c I c ( y ) B c } = min y ϵ Ω { min c J c ( y ) } + 1 t D C P ( x ) ,
t D C P ( x ) = 1 min y ϵ Ω { min c I c ( y ) B c } ,
{ D m i p ( x ) = ma x y Ω I r ( y ) ma x y Ω I r ( y ) t M I P ( x ) = D m i p ( x ) + 1 m a x ( D m i p ( x ) ) ,
d n = θ b [ θ a d D + ( 1 θ a ) d R ] + ( 1 θ b ) d B ,
θ a = s ( a v g c ( B c ) , 0.5 ) ,
θ b = s ( a v g c ( I r ) , 0.2 ) ,
t r ( x ) = e β r d a ,
I m i n c = min c { R , G , B } I c ( x ) = min c { R , G , B } ( B c ρ c ( x ) t c ( x ) + ( 1 t c ( x ) ) B c ) = B min c ρ min c ( x ) t min c ( x ) + B min c ( 1 t min c ( x ) ) ,
I min c = B r ρ min c ( x ) t r ( x ) + B r ( 1 t r ( x ) ) ,
I ^ min c ( x ) = B r ρ ^ min c ( x ) t r ( x ) + B r ( 1 t r ( x ) ) ,
B r = I ^ min c 1 t r ,
B k B r = β r ( m λ k + i ) β k ( m λ r + i ) , k { g , b }
β c = { 0.2 c = r 0.05 c = b , 0.07 c = g
max λ { R , G , B } ( t ~ ( x , λ ) ) = 1 min ( min y Ω ( x ) ( 1 I ( y , R ) ) B r , min y Ω ( x ) ( I ( y , G ) ) B g , min y Ω ( x ) ( I ( y , B ) ) B b ) ,
t 1 ~ ( x , B ) = 1 min ( min y Ω ( x ) ( 1 I ( y , R ) ) B r , min y Ω ( x ) ( I ( y , G ) ) B g , min y Ω ( x ) ( I ( y , B ) ) B b ) ,
t 1 ~ ( x , G ) = ( e β ( B ) d ( x ) ) β ( G ) β ( B ) = t 1 ~ ( x , B ) β ( G ) β ( B ) ,
t 1 ~ ( x , R ) = ( e β ( B ) d ( x ) ) β ( R ) β ( B ) = t 1 ~ ( x , B ) β ( R ) β ( B ) ,
d ( x ) = ω 0 + ω 1 m v g b ( x ) + ω 2 v r ( x ) ,
d 0 = 1 max x , c ( | B c I c ( x ) | B c ¯ ) ,
d a ( x ) = D × ( d ( x ) + d 0 ) ,
t 2 ~ ( x , R ) = e β ( R ) d a ( x ) ,
t 2 ~ ( x , G ) = ( e β ( R ) d a ( x ) ) β ( G ) β ( R ) ,
t 2 ~ ( x , B ) = ( e β ( R ) d a ( x ) ) β ( B ) β ( R ) ,
t i s ~ ( x , λ ) = ϑ [ ( t i ~ ( x , λ ) ) ] mean [ t i ~ ( x , λ ) ] , ( i = 1 , 2 ) ,
t ~ ( x , λ ) = i = 1 2 w i ( x , λ ) t i ~ ( x , λ ) ,
w i ( x , λ ) = t i s ~ ( x , λ ) j = 1 2 t j s ~ ( x , λ ) ,
S a t ( I c ( x ) ) = i = 1 M j = 1 N [ 1 min ( I c ( x ) ) max ( I c ( x ) ) ] S a t ( I c ( x ) ) = 1 , i f max ( I c ( x ) ) = 0 ,
S a t r e v ( x ) = 1 S a t ( I c ( x ) ) ,
S a t a d j r e v ( x ) = λ S a t r e v ( x ) ,
t n c ( x ) = i = 1 M j = 1 N max ( t n r ( i , j ) , S a t a d j r e v ( i , j ) ) ,
q = { B c , B c > 1 B c 1 B c , B c < 1 B c ,
max c { r , g , b } | J c ( α ) B c | q = 1 ,
K = max c { r , g , b } max | I c ( x ) B c | q ,
0 I c ( x ) B c K t ^ n r ( x ) δ c + B c 1 ,
{ δ c ln ( I c ( x ) B c K ( 1 B c ) ) / ln ( t ^ n r ( x ) ) + ε c , i f I c ( x ) > B c δ c ln ( B c I c ( x ) K B c ) / ln ( t ^ n r ( x ) ) + ε c , if I c ( x ) < B c ,
J c ( x ) = I c ( x ) B c min ( max ( K t ^ n r ( x ) δ c , t 1 ) , t 2 ) + t 0 B c , c { r , g , b } ,
J m a x c = J m e a n c + z J v a r c ,
J m i n c = J m e a n c z J v a r c ,
J C R c = ( J c J m i n c ) / ( J m a x c J m i n c ) ,
min φ x [ ( φ ( x ) O ( x ) ) 2 + ϑ ( ( φ ( x ) x ) 2 | V h ( x ) | θ + ϵ + ( φ ( x ) y ) 2 | V v ( x ) | θ + ϵ ) ] ,
V h ( x ) = O ( x ) x ; V v ( x ) = O ( x ) y ,
( φ O ) T ( φ O ) + ϑ ( φ T D x T B x D x φ + φ T D y T B y D y φ ) ,
B x = diag { 1 | V h ( x ) | θ + ϵ } , B y = diag { 1 | V v ( x ) | θ + ϵ } ,
( I + ϑ ( D x T B x D x + D y T B y D y ) ) φ = O ,
UIQM = a 1 UICM + a 2 UISM + a 3 UIConM ,
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.