Journal of Advanced Transportation

Journal of Advanced Transportation / 2021 / Article
Special Issue

Computer Vision Techniques in Intelligent Transportation Systems

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 5598390 | https://doi.org/10.1155/2021/5598390

Yuxu Lu, Yu Guo, Maohan Liang, "CNN-Enabled Visibility Enhancement Framework for Vessel Detection under Haze Environment", Journal of Advanced Transportation, vol. 2021, Article ID 5598390, 14 pages, 2021. https://doi.org/10.1155/2021/5598390

CNN-Enabled Visibility Enhancement Framework for Vessel Detection under Haze Environment

Academic Editor: Yi-Sheng Lv
Received16 Jan 2021
Revised05 Mar 2021
Accepted24 Apr 2021
Published20 May 2021

Abstract

Maritime images captured under haze environment often have a terrible visual effect, making it easy to overlook important information. To avoid the failure of vessel detection caused by fog, it is necessary to preprocess the collected hazy images for recovering vital information. In this paper, a novel CNN-enabled visibility dehazing framework is proposed, consisting of two subnetworks, that is, Coarse Feature Extraction Module (C-FEM) and Fine Feature Fusion Module (F-FFM). Specifically, C-FEM is a multiscale haze feature extraction network, which can learn information from three scales. Correspondingly, F-FFM is an improved encoder-decoder network to fuse multiscale information obtained by C-FEM and enhance the visual effect of the final output. Meanwhile, a hybrid loss function is designed for monitoring the multiscale output of C-FEM and the final result of F-FFM simultaneously. It is worth mentioning that massive maritime images are considered the training dataset to further adapt the vessel detection task under haze environment. Comprehensive experiments on synthetic and realistic images have verified the superior effectiveness and robustness of our CNN-enabled visibility dehazing framework compared to several state-of-the-art methods. Our method preprocesses images before vessel detection to demonstrate our framework has the capacity of promoting maritime video surveillance.

1. Introduction

1.1. Background and Related Work

It is well known that the maritime surveillance system is an indispensable part of vessel traffic services [1]. As an efficient, convenient, and intuitive monitoring method, Closed Circuit Television (CCTV) is thus widely applied to critical regions, for example, ports and waterways. As shown in Figure 1, significant information in the images, however, is easily buried under the haze. Therefore, it is difficult for maritime regulatory authority to effectively extract detailed information (e.g., monitoring targets and water traffic conditions) from degraded images, which seriously affects maritime supervision efficiency. Besides, the low-quality images collected under haze environment have also brought severe challenges to intelligent surveillance methods based on vessel detection, recognition, and tracking [25]. To improve the maritime safety surveillance capability under haze environment, it is necessary to restore images under CCTV monitoring. In current literature, dehazing methods can be categorized into image enhancement-based methods, physical model-based methods, and deep learning-based methods.

1.1.1. Image Enhancement-Based Methods

Early research mainly enhanced the contrast of hazy images to highlight the scene characteristics of the interest region. Histogram Equalization (HE) [6] is a classic enhancement method devoted to enhancing the contrast by stretching the dynamic range of image pixel values. In current literature, HE-based methods can be divided into two categories, that is, global and local histogram equalization. Since the global histogram equalization can enhance the entire image by single mapping, it has the characteristics of simple principle and fast calculation. However, these methods often ignore the local information, resulting in the haze-free images having poor performance. To solve this problem, Stark et al. [7] proposed an adaptive local histogram equalization method. Subsequently, Kim et al. [8] proposed a nonoverlapping subblock histogram equalization method to reduce the blocky effect and computational complexity. Retinex theory-based image dehazing method is devoted to separating the illumination and reflection from the hazy image and enhancing the image by reducing the illumination impact. Jobson et al. [9] first used the Gaussian filter to obtain a smooth illumination according to the Retinex theory and thus proposed a single-scale Retinex (SSR). To avoid color distortion, Rahman et al. [10] proposed a multiscale Retinex algorithm with color restoration (MSRCR) by introducing a color compensation factor. To sum up, the image generated by these methods has higher contrast and color fidelity, but the halo often appears on the edge of the interest object.

1.1.2. Physical Model-Based Methods

These methods are proposed based on a certain physical model that describes the process of image degradation under haze weather. Because these methods use mathematical methods to describe the haze formation process based on light scattering, the final restored target is clear and natural. Physical model-based methods include the following categories, that is, depth-based method and prior-based method. The depth-based methods mainly obtain depth information through a specific method and then get stable model parameters. Finally, the potentially clear image can be obtained by the atmospheric scattering model. For instance, Oakley et al. [11] first used radar and other types of equipment to measure the shooting scene depth. Hautiere et al. [12] proposed an image dehazing algorithm based on the 3D geographic model for vehicle vision systems. Although these methods have an excellent dehazing effect, they heavily rely on distance measuring equipment. Therefore, Liu et al. [13] proposed a dehazing method to estimate the depth map through a second-order variational framework. In contrast, the prior-based method mainly analyzes haze formation and relies on specific prior information to achieve image dehazing. Dark channel prior (DCP) [14] and its improvements [1517] have an excellent performance in the image dehazing task. Through numerous statistics on outdoor haze-free images, He et al. proposed DCP based on the assumption that most local color blocks contain some pixels with very low intensity in at least one color channel. Zhu et al. proposed a novel linear color attenuation prior [18], based on the difference between the brightness and the saturation of pixels within the hazy image. Subsequently, a nonlocal prior dehazing method [19] is employed to obtain the nonlocal transmission map from the haze-line property. To reduce halo and unnatural artifacts, a low-complexity color ellipsoid prior [20] is designed to accurately and swiftly estimate the transmission map. In current literature, several variational model-based transmission estimation methods [15, 21, 22] are also proposed. Although prior-based methods have verified excellent dehazing performance, they may cause a loss in color fidelity under certain circumstances and fail to obtain pleasing visual effects on maritime images.

1.1.3. Deep Learning-Based Method

Affected by the excellent results of machine learning technology in computer vision, the CNN-enabled dehazing method has gradually become a new research direction. For instance, Tang et al. [23] proposed a learning-based approach to systematically investigate different haze-relevant features and identify the best feature combination for image dehazing. Zhu et al. [18] designed the color attenuation prior model to estimate the scene depth. Meanwhile, a supervised learning method is used to obtain the scene depth and the atmospheric scattering model parameters. However, this method fails to estimate the scene depth in the white environment. To further improve the deep learning-based dehazing methods performance, Cai et al. [24] first constructed a convolutional neural network (DehazeNet) to learn the mapping relationship between hazy images and transmission. DehazeNet employed artificially synthesized hazy images of different concentrations as the dataset. The trained DehazeNet can directly estimate the corresponding transmission and restore the potentially clear image according to the traditional atmospheric light scattering model. Ren et al. [25] proposed a Multiscale Convolutional Neural Network (MSCNN). MSCNN used the New York University indoor image depth database [26] to synthesize different hazy images as the dataset, making the hazy image more realistic. Subsequently, Zhao et al. [27] used outdoor scenes synthetic images and proposed a fully convolutional neural network model to estimate the transmission. Compared with DehazeNet and MSCNN, this network has better visual performance. However, it produces more parameters and calculations. To simplify the calculation, Li et al. [28] designed an end-to-end light-weight convolutional neural network (AOD-Net) that effectively balances calculation speed and visual effects. Inspired by image denoising, Du et al. [29] proposed a Deep Residual Learning (DRL) network to reconstruct the potential image. Besides, Chen et al. [30] proposed an end-to-end gated context aggregation network to directly restore the final haze-free image. It is worth noting that if the training datasets do not contain the geometric features presented in the haze-free target, it is usually difficult to produce satisfactory image quality. Therefore, it is necessary to design a CNN-enabled visibility enhancement framework for vessel detection under haze environment to further improve maritime video surveillance efficiency.

1.2. Contributions

This paper presents a CNN-enabled framework for practically solving vessel detection problem under haze environment. The main contribution of our method differs from others in the following aspects:(i)A CNN-enabled visibility dehazing framework is proposed to improve the visibility of maritime images. Specifically, this framework includes a Coarse Feature Extraction Module (C-FEM) for capturing multiscale features and Fine Feature Fusion Module (F-FFM) for information fusion and enhancement.(ii)To improve the generalization of the proposed network, we design a novel hybrid loss function to supervise the multiscale outputs of C-FEM and the final output of F-FFM simultaneously.(iii)Image dehazing and vessel detection experiments under haze conditions are conducted to verify our superior performance compared to several state-of-the-art methods.

1.3. Construction

The remainder of this paper is divided into the following sections. Section 2 mainly describes the problem formulation related to the imaging model. In Section 3, a CNN-enabled visibility enhancement framework is proposed to improve the visual effect of hazy images. Implementation details and experiments are implemented in Section 4. Finally, we conclude our main contributions in Section 5.

2. Problem Formulation

2.1. Atmospheric Scattering Model

Video images collected by maritime video surveillance system under haze conditions often have poor visual quality. As shown in Figure 2, Narasimhan et al. [31] proposed the atmospheric scattering model to divide the light irradiance into the incident light attenuation part and the atmospheric light imaging part . The incident light attenuation model considers that the reflected light by the vessel surface is scattered and attenuated by particulate impurities in the air, reducing the intensity of light reaching the imaging system. Note that as the propagation distance increases, the reflected light intensity decays exponentially. On the contrary, the atmospheric light imaging model believes that light intensity scattered by natural light enters the imaging system to participate in imaging. As the propagation distance increases, the scattered light intensity will gradually increase. Finally, the images collected by the imaging system under haze environments exhibit degradation phenomena such as low contrast, blurred images, and color distortion under the combined action of these two models. Mathematically, the atmospheric light scattering model can be expressed aswhere and , respectively, denote the hazy image and haze-free image, and represent the atmospheric light value and scattering coefficient, is the image pixel index, and is the distance between the scene point and the imaging system, that is, field depth. When we set , equation (1) can thus be rewritten as follows:with being the transmission. According to equation (2), the restoration haze-free image can be easily obtained by

2.2. Transformed Formula

According to equation (3), we can obtain a satisfactory haze-free image by accurately estimating and . However, it is intractable to estimate two parameters simultaneously. For the sake of better performance of the end-to-end network, Li et al. [28] proposed the transformed atmospheric scattering model, which can be given bywhere is a particular parameter to integrate and ; that is, . It is worth noting that hazy maritime images usually contain background (i.e., sky and water regions). Many statistical features-based methods, for example, DCP and maximum local contrast, often fail to obtain ideal transmission maps. Deep learning-based methods do not rely on these statistical features and can learn the mapping of hazy and haze-free images. Therefore, we will propose the CNN-enabled visibility enhancement network to effectively improve the quality of hazy maritime images and improve vessel detection accuracy.

3. CNN-Enabled Visibility Enhancement Framework

In this section, a CNN-enabled visibility enhancement framework is proposed to process hazy maritime images shown in Figure 3. This framework consists of two subnetworks, that is, Coarse Feature Extraction Module (C-FEM) and Fine Feature Fusion Module (F-FFM). In this work, C-FEM is introduced to learn multiscale hazy features. Meanwhile, F-FFM, an improved encoder-decoder network, is proposed to fuse and enhance the hazy image and the multiscale output obtained by C-FEM. Once our method gets the sharp image, it can easily detect the vessel containing the image by any target detection method.

3.1. C-FEM

C-FEM is a module for initial extracting the features of the hazy image. In particular, C-FEM can perform mapping learning on three scales (i.e., 1, 1/4, and 1/16) to obtain coarse feature information with different resolutions simultaneously. Figure 4 shows the network architecture of C-FEM under one resolution, which is only composed of six convolutions. In this work, dilated convolution is embedded to increase the reception field of C-FEM. According to our research, dilated convolution can reduce the loss of spatial features without reducing the receptive field. However, the use of dilated convolution may increase the risk of spatially continuous information loss, destroying image feature information (especially edges). To alleviate the interference caused by dilated convolution, we combine standard convolution (Conv) [32] and dilated convolution (DConv) with improving the detailed information extraction ability. DConv can effectively solve this difficult problem with different receptive fields by adjusting the dilation rate value. Formally, standard convolution and dilated convolution are, respectively, defined as follows:where is the discrete signal, is convolution kernel, subscript (·) is the position of a discrete signal, is the dilation factor, and is dilated convolutions with a factor . The only difference between standard convolution and dilated convolution is the influence of the dilation factor on the multiplication position of and . Dilated convolution benefits from are no longer limited to a fixed receptive field, and the dilation factor d can be adjusted to have a larger receptive field. In this work, the method of fusing Conv and DConv can reduce the loss of spatial information caused by the excessive dilated rate and fully consider long- and short-distance information to present a better visual effect. Furthermore, Instance Normalization (IN) [33] and Rectified Linear Unit (ReLU) [34] are deployed after each Conv layer. Meanwhile, the feature map channels of the first five convolution outputs are set to 32.

According to our research, most deep learning-based dehazing methods rely on more complex network models to obtain better visual effects. When the model is relatively simple, it is usually hard to learn the fog feature, causing information damage to potential images. In contrast, the imaging model reduces the algorithm complexity, making it easier for the network to extract information. The introduction of this model makes it possible for a simple network model to extract potential multiscale features from the original image. It is worth noting that the output of C-FEM only introduced to provide a prior is not used as the final result. Simultaneously, it has a faster calculation speed and can satisfy the needs of real-time processing.

3.2. F-FFM

Coarse feature maps of three resolutions (i.e., 1, 1/4, and 1/16) have been obtained by C-FEM, which contains most of the potentially clear image information. The feature map information obtained by a standard encoder-decoder CNN is usually found irregularly. When the prior information obtained by C-FEM is introduced to the encoder, we believe that F-FFM can obtain better parameters and accelerate the convergence speed. When the feature maps are fused at different scales, the deep network can further extract edge detail information. Table 1 shows that the architecture of Fine Feature Fusion Modul (F-FFM) is a special encoder-decoder structure. Specifically, F-FFM only performs two downsampling operations and merges with the corresponding output of C-FEM. Both the encoder and the decoder consist of the same module, that is, a convolution filter (Conv) [32], Instance Normalization (IN) [33], and Rectified Linear Unit (ReLU) [34]. Maximum pooling and bilinear interpolation are exploited to perform down- and upsampling operations on the feature map, respectively. Different from traditional encoder-decoder structures, our F-FFM encoder integrates the output of C-FEM. This strategy can guide F-FEM to learn the mapping of hazy images and haze-free targets. To better preserve the boundary details of the input, we adopt a global skip connection strategy to further ensure the details of the output image. In other words, the output of the last convolution and the input image is directly added as the output of F-FFM, and we find that it can significantly improve the dehazing effect through comparative experiments.


InputLayersTypeChannelsFilterStrideOutputSize

1Convolutional6411
13Convolutional6412
2Max pooling6423
31Convolutional12814
43Convolutional12815
5Max pooling12826
61Convolutional25617
75Convolutional25618
8Bilinear interpolation9
94Skip connection128110
101Convolutional128111
113Convolutional128112
12Bilinear interpolation13
131Skip connection64114
141Convolutional64115
153Convolutional64116
16Global residual31

3.3. Loss Function

To robustly learn the multiscale mapping relationship between hazy image and haze-free image, a specific loss function is proposed. As shown in Figure 3, C-FEM has three scale outputs (i.e., , , and ). These three images sequentially have 1, 1/4, and 1/16 of the original image size. Subsequently, the maximum pooling operation is used to obtain clear images with three scales named , , and , which, respectively, correspond to the scale of , , and . In this work, Mean Square Error (MSE) loss function is employed to constrain each scale output of C-FEM; that is,where , , , and are trade-off parameters of corresponding loss functions. To further preserve the high-frequency details of the potential haze-free image while eliminating boundary artifacts, a hybrid loss function is introduced to limit the ground truth and the predicted restored image ; that is,with , , , and being the penalty weights. Multiscale structural similarity (MS-SSIM) [35] is firstly employed to constrain the structure, brightness, and contrast of the image. The MS-SSIM loss function can be defined as follows:with being the calculation operation of the multiscale structural similarity index between two images. The hazy image inevitably has a low contrast phenomenon in local regions, resulting in color distortion. To solve this problem, the Mean Absolute Error (MAE) loss function is introduced as a part of , which can reduce the color distortion problem to a certain extent. In particular, is defined as

The high-frequency detail information is easily destroyed in the process of image dehazing. To further improve the fidelity and authenticity of details, we propose an additional edge loss function [36] to limit the high-frequency components, for example, edge and texture. can be written aswhere and represent edge maps extracted from and via the Laplacian operator, respectively. The penalty coefficient is empirically set to . In addition, the Total Variation (TV) loss function [37] is exploited to suppress the pixel-jump problem, which can be given bywhere and represent the operators of the horizontal and vertical gradients, respectively. We refer interested readers to [3537] for more details on calculations of MS-SSIM, edge loss, and TV. To sum up, the total loss function can be written as follows:where and are the penalty coefficient of and . By comparative experiment, we manually selected the optimal weight of all loss functions; that is, , , , , , , , , and .

4. Experimental Results and Analysis

This section will describe all the implementation details of network training, including dataset construction and network parameter settings. We will compare our method with several state-of-the-art dehazing methods on both synthetic and realistic hazy maritime images. To prove that our method can improve detection accuracy, our proposed framework will be employed in vessel detection tasks under haze environment.

4.1. Comparison Methods and Evaluation Indicators

Our method will be compared with four handcrafted prior-based methods and three deep learning-based methods. For the sake of fair comparison, the parameters of other competing dehazing methods are provided by the authors’ code.(1)DCP: Dark Channel Prior-Based Method [14]. Through numerous statistics on outdoor haze-free images, DCP is proposed based on the assumption that most local color blocks contain some pixels with very low intensity in at least one color channel. According to this statistic prior and the haze imaging model, a high-quality haze-free image can be directly obtained.(2)GRM: Gradient Residual Minimization-Based Method [16]. This method first proposes the depth-edge-aware smoothing algorithm to refine the transmission map generated by local priors. Meanwhile, Gradient Residual Minimization (GRM) is introduced during the image recovery process. By comparison, the GRM-based method can jointly recover the haze-free image and explicitly minimize possible visual artifacts in it.(3)HL: Haze-Lines-Based Method [38]. This method finds that the pixel values of a hazy image can be modeled as lines intersecting at the air light. Based on this prior condition, a novel haze-lines-based method is proposed to restore the hazy image better. It is worth noting that the complexity of HL is linear in the number of pixels, having higher computational efficiency.(4)F-LDCP: Fusion of Luminance and Dark Channel Prior-Based Method [39]. To make the sky region more natural in long-shot images, a Fusion of Luminance and Dark Channel Prior (F-LDCP) method is proposed. The transmission maps estimated by the brightness model and the DCP model are fused through a soft segmentation.(5)MSCNN: Multiscale Convolutional Neural Networks [25]. To learn the practical features of a hazy image, a multiscale deep network (MSCNN) is designed to address the image dehazing problem. MSCNN can be divided into the coarse-scale network and fine-scale network. The coarse-scale network can learn a holistic estimation of the scene transmission, and the fine-scale network is used to optimize the obtained transmission. Finally, the haze-free image can be obtained by the atmospheric scattering model.(6)AOD-Net: All-in-One Dehazing Network [28]. AOD-Net, a light-weight CNN, is designed according to the reformulated atmospheric scattering model. This network replaces the atmospheric light value and transmission with one parameter. It is worth mentioning that AOD-Net has been embedded in other deeper models (e.g., Faster R-CNN) to improve the advanced tasks of hazy images.(7)GCA-Net: Gated Context Aggregation Network [30]. GCA-Net is an end-to-end Gated Context Aggregation Network. In particular, the latest smoothed dilation technology is designed to eliminate gridding artifacts caused by the extensive-used dilated convolution with negligible additional parameters.

In synthetic and realistic experiments, we will compare these methods with our proposed method. In addition, three full-reference indicators, that is, Peak-Signal-to-Noise Ratio (PSNR) [40], SSIM [41], and Feature Similarity (FSIM) [42], are introduced to evaluate the dehazing performance in the synthetic experiment. Meanwhile, one popular no-reference image quality assessment method, that is, Natural Image Quality Evaluator (NIQE) [43], is also exploited to perform dehazing quality evaluation in the real experiment. Theoretically, higher values of PSNR, SSIM, FSIM, and lower values of NIQE indicate better visual performance.

4.2. Experimental Datasets and Settings

To guarantee high-quality dehazing results, we tend to select 7000 haze-free maritime images as the dataset and randomly cropped these images into 34000 patches of size . In this work, our network is trained for 80 epochs. The learning rate of the first 40 epochs is and the learning rate of the last 40 epochs is to increase the convergence rate. In each epoch, the hazy synthetic versions are obtained by equation (2), that is, atmospheric scattering model. In particular, the transmission and atmospheric light value are random constants ranging between (0.2, 0.6) and (0.8, 0.9). All numerical experiments and model training are conducted in Python 3.7 and Matlab2019a environment running on a PC with Intel(R) Core (TM) i7-9750H CPUa 2.60 GHz and a Nvidia GeForce GTX 2080Ti GPU. It takes about 10 hours to train our network with the Pytorch package [44]. The Python source code is available at https://github.com/LouisYuxuLu/JAT_Dehazing.

4.3. Experiments on Synthetic Maritime Datasets

This subsection is devoted to comparing our proposed method with seven popular dehazing methods, that is, DCP [14], GRM [16], HL [38], F-LDCP [39], MSCNN [25], AOD-Net [28], and GCA-Net [30]. In this work, six clear maritime images are exploited to conduct synthetic experiments shown in Figure 5. In particular, we tend to produce 36 degraded images by setting and according to equation (2). To quantitatively evaluate the dehazing performance, three full-reference metrics (i.e., PSNR, SSIM, and FSIM) are employed in this experiment.

For the sake of better visual comparisons, the dehazing versions of hazy images with different degrees obtained by various methods are shown in Figure 6. It can be clearly observed that DCP and HL often make the color unnatural. Meanwhile, due to the incomplete dehazing, the results obtained by DCP easily suffer from the interference of boundary artifacts around the object. Although GRM can get satisfactory visual effects, it requires complex calculations and has the risk of excessive smoothness. F-LDCP can excellently solve the blocking artifacts and halo problems in the sky regions, but the color fidelity in the water regions needs improvement. MSCNN and AOD-Net can handle the low-concentration hazy image. However, the restored versions of the high-concentration hazy images (i.e., hazy images with ) usually have a poor visual effect. GCA-Net fails in the synthetic experiment, resulting in a nonuniform distribution of fog remaining in the results. By comparison, our method can not only make the restored image visually more natural but also ensure the color reproduction of the sky and water regions.

To further confirm the superiority of our method, the quantitative results of PSNR, SSIM, and FSIM are shown in Figure 7 and Table 2. PSNR, SSIM, and FSIM values are illustrated using box-plot in Figure 7. It can be seen that our method has higher index values in most cases. Particularly for high-concentration hazy images, our method can stably obtain high-quality restored versions. Besides, Table 2 shows three metrics value comparisons of various image enhancement methods on 36 hazy images. In particular, we display the best result of each metric in bold. Due to the highest values of PSNR, SSIM, and FSIM, our method has the best dehazing performance. Meanwhile, the standard deviation calculated by the SSIM and FSIM is the smallest, which verifies that our method has excellent robustness.


MethodsPSNRSSIMFSIM

Hazy13.17 ± 2.900.727 ± 0.1380.830 ± 0.105
DCP15.23 ± 2.930.789 ± 0.0860.933 ± 0.022
GRM18.96 ± 3.210.845 ± 0.0700.912 ± 0.046
HL19.46 ± 1.710.874 ± 0.0420.937 ± 0.027
F-LDCP20.58 ± 3.000.933 ± 0.0320.972 ± 0.011
MSCNN19.71 ± 5.190.890 ± 0.0910.940 ± 0.055
AOD-net18.70 ± 3.300.848 ± 0.1130.889 ± 0.079
GCA-net21.26 ± 1.380.899 ± 0.0270.947 ± 0.012
Ours25.80 ± 3.010.954 ± 0.0140.982 ± 0.003

4.4. Experiments on Realistic Maritime Datasets

This subsection will verify the reliability of several methods in realistic hazy maritime images due to the distinctness between synthetic and realistic versions. Meanwhile, NIQE is introduced to describe the naturalness of visual effects quantitatively, and our proposed method will be compared with seven dehazing methods, that is, DCP [14], GRM [16], HL [38], F-LDCP [39], MSCNN [25], AOD-Net [28], and GCA-Net [30]. Figure 8 shows several dehazing results to reflect the imaging performance more intuitively.

From the visual comparisons, DCP and HL have serious color distortion problems and blocking effects in the sky regions. Recovery results obtained by GRM have the risk of low contrast, especially in the recovery task of Image 9. F-LDCP and AOD-Net fail to correct the color of the image. GCA-Net not only has the problem of overexposure in the sky region but also has nonuniform fog remaining in the image. Although MSCNN has better visual effects than other methods, our method has pleasing color and can remove fog more fully. Our superior performance can be further confirmed by the quantitative results NIQE shown in Table 3.


DCPGRMHLF-LDCPMSCNNAOD-NetGCA-NetOurs

Image 75.4485.7805.2965.9326.2247.2205.0805.057
Image 85.52510.2974.9615.8685.8265.7165.6945.122
Image 94.8565.2815.2045.3305.2555.4094.6265.103
Average5.276 ± 0.2997.119 ± 2.2995.154 ± 0.1415.710 ± 0.2705.768 ± 0.3986.115 ± 0.7925.133 ± 0.4385.094 ± 0.027

4.5. Experiments on Vessel Detection under Haze Environment

In the maritime imaging system, the harsh imaging environment severely restricts the regular operation of the visible light imaging sensor, reduces vessel detection accuracy, and leads to incorrect identification. To prove this phenomenon, we, respectively, used YOLOv4 [45] and Faster-RCNN [46] to detect vessels in haze and haze-free images. As shown in Figure 9, it is easily found that the haze image has low contrast and massive useful information is obscured, which leads to problems, for example, identification errors or missing identification during the target detection process. After dehazing, the vessel target is effectively captured and recognized, and the recognition accuracy is significantly increased. Therefore, dehazing the degraded hazy image by our method can improve vessel detection performance. The computer and the related workers can make correct decisions in time.

5. Conclusion

In this paper, a novel CNN-enabled visibility dehazing framework was proposed, which could significantly improve the visual effect of images captured by the maritime camera under haze environment. In particular, this framework is composed of two subnetwork named Coarse Feature Extraction Module (C-FEM) and Fine Feature Fusion Module (F-FFM). C-FEM is an initial multiscale feature extraction network containing three simple six-layer convolutional networks, that is, Single C-FEM. C-FEM can obtain coarse feature maps from 1, 1/4, and 1/16 of the original image pixel size. F-FFM is a special encoder-decoder structure used to fuse and enhance the multiscale information obtained by C-FEM and original hazy image. To further improve the network performance, a corresponding loss function is proposed to simultaneously supervise the multiscale output of C-FEM and the final result of F-FFM. Furthermore, our dataset contains massive maritime images to complete the vessel detection task under haze environment successfully. Both qualitative and quantitative experiments have illustrated the effectiveness of our proposed framework.

Data Availability

The image data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Authors’ Contributions

Yuxu Lu and Yu Guo are co-first authors.

Acknowledgments

This work was supported by the National Key R&D Program of China (No. 2018YFC0309602).

References

  1. R. W. Liu, J. Nie, S. Garg, Z. Xiong, Y. Zhang, and M. S. Hossain, “Data-driven trajectory quality improvement for promoting intelligent vessel traffic services in 6G-enabled maritime IoT systems,” IEEE Internet of Things Journal, vol. 8, pp. 1–12, 2021. View at: Google Scholar
  2. X. Chen, L. Qi, Y. Yang et al., “Video-based detection infrastructure enhancement for automated ship recognition and behavior analysis,” Journal of Advanced Transportation, vol. 2020, Article ID 7194342, 12 pages, 2020. View at: Publisher Site | Google Scholar
  3. X. Chen, X. Xu, Y. Yang, H. Wu, J. Tang, and J. Zhao, “Augmented ship tracking under occlusion conditions from maritime surveillance videos,” IEEE Access, vol. 8, pp. 42884–42897, 2020. View at: Publisher Site | Google Scholar
  4. X. Chen, S. Wang, C. Shi, H. Wu, J. Zhao, and J. Fu, “Robust ship tracking via multi-view learning and sparse representation,” Journal of Navigation, vol. 72, no. 1, pp. 176–192, 2019. View at: Publisher Site | Google Scholar
  5. X. Chen, X. Xu, Y. Yang, Y. Huang, J. Chen, and Y. Yan, “Visual ship tracking via a hybrid kernelized correlation filter and anomaly cleansing framework,” Applied Ocean Research, vol. 106, Article ID 102455, 2020. View at: Google Scholar
  6. E. D. Pisano, S. Zong, B. M. Hemminger et al., “Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms,” Journal of Digital Imaging, vol. 11, no. 4, pp. 193–200, 1998. View at: Publisher Site | Google Scholar
  7. J. A. Stark, “Adaptive image contrast enhancement using generalizations of histogram equalization,” IEEE Transactions on Image Processing, vol. 9, no. 5, pp. 889–896, 2000. View at: Publisher Site | Google Scholar
  8. T. K. Kim, J. K. Paik, and B. S. Kang, “Contrast enhancement system using spatially adaptive histogram equalization with temporal filtering,” IEEE Transactions on Consumer Electronics, vol. 44, no. 1, pp. 82–87, 1998. View at: Google Scholar
  9. D. J. Jobson, Z. Rahman, and G. A. Woodell, “Properties and performance of a center/surround retinex,” IEEE Transactions on Image Processing, vol. 6, no. 3, pp. 451–462, 1997. View at: Publisher Site | Google Scholar
  10. Z. Rahman, D. J. Jobson, and G. A. Woodell, “Multi-scale retinex for color image enhancement,” in Proceedings of the IEEE International Conference on Image Processing, vol. 3, pp. 1003–1006, Lausanne, Switzerland, September 1996. View at: Google Scholar
  11. J. P. Oakley and B. L. Satherley, “Improving image quality in poor visibility conditions using a physical model for contrast degradation,” IEEE Transactions on Image Processing, vol. 7, no. 2, pp. 167–179, 1998. View at: Publisher Site | Google Scholar
  12. N. Hautiere, J. P. Tarel, and D. Aubert, “Towards fog-free in-vehicle vision systems through contrast restoration,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8, Minneapolis, MN, USA, June 2007. View at: Google Scholar
  13. R. W. Liu, S. Xiong, and H. Wu, “A second-order variational framework for joint depth map estimation and image dehazing,” in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 1433–1437, Calgary, Canada, September 2018. View at: Google Scholar
  14. K. He, J. Sun, and X. Tang, “Single image haze removal using dark channel prior,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 33, no. 12, pp. 2341–2353, 2011. View at: Publisher Site | Google Scholar
  15. Q. Shu, C. Wu, R. W. Liu, K. T. Chui, and S. Xiong, “Two-phase transmission map estimation for robust image dehazing,” in Proceedings of the International Conference on Neural Information Processing, pp. 529–541, Siem Reap, Cambodia, December 2018. View at: Publisher Site | Google Scholar
  16. C. Chen, M. N. Do, and J. Wang, “Robust image and video dehazing with visual artifact suppression via gradient residual minimization,” in Proceedings of the European Conference on Computer Vision, pp. 576–591, Amsterdam, The Netherlands, October 2016. View at: Publisher Site | Google Scholar
  17. Q. Liu, X. Gao, L. He, and W. Lu, “Single image dehazing with depth-aware non-local total variation regularization,” IEEE Transactions on Image Processing, vol. 27, no. 10, pp. 5178–5191, 2018. View at: Publisher Site | Google Scholar
  18. Q. Zhu, J. Mai, and L. Shao, “A fast single image haze removal algorithm using color attenuation prior,” IEEE Transactions on Image Processing: A Publication of the IEEE Signal Processing Society, vol. 24, no. 11, pp. 3522–3533, 2015. View at: Publisher Site | Google Scholar
  19. D. Berman and S. Avidan, “Non-local image dehazing,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1674–1682, Las Vegas, NV, USA, June 2016. View at: Google Scholar
  20. T. M. Bui and W. Kim, “Single image dehazing using color ellipsoid prior,” IEEE Transactions on Image Processing, vol. 27, no. 2, pp. 999–1009, 2018. View at: Publisher Site | Google Scholar
  21. Q. Shu, C. Wu, Z. Xiao, and R. W. Liu, “Variational regularized transmission refinement for image dehazing,” in Proceedings of the IEEE International Conference on Image Processing, pp. 2781–2785, Taipei, Taiwan, August 2019. View at: Google Scholar
  22. Q. Shu, C. Wu, Q. Zhong, and R. W. Liu, “Alternating minimization algorithm for hybrid regularized variational image dehazing,” Optik, vol. 185, pp. 943–956, 2019. View at: Publisher Site | Google Scholar
  23. K. Tang, J. Yang, and J. Wang, “Investigating haze-relevant features in a learning framework for image dehazing,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2995–3000, Columbus, GA, USA, September 2014. View at: Google Scholar
  24. B. Cai, X. Xu, K. Jia, C. Qing, and D. Tao, “Dehazenet: an end-to-end system for single image haze removal,” IEEE Transactions on Image Processing, vol. 25, no. 11, pp. 5187–5198, 2016. View at: Publisher Site | Google Scholar
  25. W. Ren, S. Liu, H. Zhang, J. Pan, X. Cao, and M.-H. Yang, “Single image dehazing via multi-scale convolutional neural networks,” in Proceedings of the European Conference on Computer Vision, pp. 154–169, Cham, Amsterdam, October 2016. View at: Publisher Site | Google Scholar
  26. N. Silberman, D. Hoiem, P. Kohli, and R. Fergus, “Indoor segmentation and support inference from RGBD images,” in Proceedings of the European Conference on Computer Vision, pp. 746–760, Berlin, Germany, October 2012. View at: Publisher Site | Google Scholar
  27. X. Zhao, K. Wang, Y. Li, and J. Li, “Deep fully convolutional regression networks for single image haze removal,” in Proceedings of the IEEE Visual Communications and Image Processing, pp. 1–4, St. Petersburg, FL, USA, December 2017. View at: Google Scholar
  28. B. Li, X. Peng, Z. Wang, J. Xu, and D. Feng, “AOD-net: all-in-one dehazing network,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 4770–4778, Venice, Italy, October 2017. View at: Google Scholar
  29. Y. Du and X. Li, “Recursive deep residual learning for single image dehazing,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 730–737, Salt Lake City, UT, USA, June 2018. View at: Google Scholar
  30. D. Chen, M. He, Q. Fan et al., “Gated context aggregation network for image dehazing and deraining,” in Proceedings of the IEEE Winter Conference on Applications of Computer Vision, pp. 1375–1383, Waikoloa Village, HI, USA, January 2019. View at: Google Scholar
  31. S. G. Narasimhan and S. K. Nayar, “Chromatic framework for vision in bad weather,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 598–605, Hilton Head Island, SC, USA, June 2000. View at: Google Scholar
  32. Y. Pang, M. Sun, X. Jiang, and X. Li, “Convolution in convolution for network in network,” IEEE Transactions on Neural Networks and Learning Systems, vol. 29, no. 5, pp. 1587–1597, 2018. View at: Publisher Site | Google Scholar
  33. X. Huang and S. Belongie, “Arbitrary style transfer in real-time with adaptive instance normalization,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 1510–1519, Venice, Italy, October 2017. View at: Google Scholar
  34. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017. View at: Publisher Site | Google Scholar
  35. Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multi-scale structural similarity for image quality assessment,” in Proceedings of the Asilomar Conference on Signals, Systems and Computers, pp. 1398–1402, Pacific Grove, CA, USA, November 2003. View at: Google Scholar
  36. K. Jiang, Z. Wang, P. Yi et al., “Multi-scale progressive fusion network for single image deraining,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8346–8355, Seattle, WA, USA, June 2020. View at: Google Scholar
  37. L. I. Rudin, S. Osher, and E. Fatemi, “Nonlinear total variation based noise removal algorithms,” Physica D Nonlinear Phenomena, vol. 60, no. 1–4, pp. 259–268, 1992. View at: Publisher Site | Google Scholar
  38. D. Berman, T. Treibitz, and S. Avidan, “Air-light estimation using haze-lines,” in Proceedings of the IEEE International Conference on Computer Photography, pp. 1–9, Stanford, CA, USA, May 2017. View at: Google Scholar
  39. Y. Zhu, G. Tang, X. Zhang, J. Jiang, and Q. Tian, “Haze removal method for natural restoration of images with sky,” Neurocomputing, vol. 275, no. 31, pp. 499–510, 2017. View at: Google Scholar
  40. Z. Wang and A. C. Bovik, “Mean squared error: love it or leave it?” IEEE Signal Processing Magazine, vol. 26, no. 1, pp. 98–117, 2009. View at: Google Scholar
  41. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004. View at: Publisher Site | Google Scholar
  42. L. Zhang, L. Zhang, X. Mou, and D. Zhang, “FSIM: a feature similarity index for image quality assessment,” IEEE Transactions on Image Processing: A Publication of the IEEE Signal Processing Society, vol. 20, no. 8, pp. 2378–2386, 2011. View at: Publisher Site | Google Scholar
  43. A. Mittal, R. Soundararajan, and A. C. Bovik, “Making a “completely blind” image quality analyzer,” IEEE Signal Processing Letters, vol. 20, no. 3, pp. 209–212, 2013. View at: Publisher Site | Google Scholar
  44. P. Adam, G. Sam, and C. Soumith, “Automatic differentiation in PyTorch,” in Proceedings of the Conference and Workshop on Neural Information Processing Systems, pp. 1–4, Long Beach, CA, USA, December 2017. View at: Google Scholar
  45. A. Bochkovskiy, C. Y. Wang, and H. Y. M. Liao, “YOLOv4: optimal speed and accuracy of object detection,” 2020, https://arxiv.org/abs/2004.10934. View at: Google Scholar
  46. S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: towards real-time object detection with region proposal networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 6, pp. 1137–1149, 2016. View at: Google Scholar

Copyright © 2021 Yuxu Lu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views436
Downloads429
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.