Table of Contents Author Guidelines Submit a Manuscript
Computational Intelligence and Neuroscience
Volume 2017 (2017), Article ID 8351232, 8 pages
https://doi.org/10.1155/2017/8351232
Research Article

Underwater Inherent Optical Properties Estimation Using a Depth Aided Deep Neural Network

1Department of Electronic Engineering, College of Information Science and Engineering, Ocean University of China, Qingdao, China
2School of Life Science and Technology, Xidian University, Xi’an, China

Correspondence should be addressed to Bing Zheng

Received 21 June 2017; Accepted 26 October 2017; Published 15 November 2017

Academic Editor: Leonardo Franco

Copyright © 2017 Zhibin Yu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Underwater inherent optical properties (IOPs) are the fundamental clues to many research fields such as marine optics, marine biology, and underwater vision. Currently, beam transmissometers and optical sensors are considered as the ideal IOPs measuring methods. But these methods are inflexible and expensive to be deployed. To overcome this problem, we aim to develop a novel measuring method using only a single underwater image with the help of deep artificial neural network. The power of artificial neural network has been proved in image processing and computer vision fields with deep learning technology. However, image-based IOPs estimation is a quite different and challenging task. Unlike the traditional applications such as image classification or localization, IOP estimation looks at the transparency of the water between the camera and the target objects to estimate multiple optical properties simultaneously. In this paper, we propose a novel Depth Aided (DA) deep neural network structure for IOPs estimation based on a single RGB image that is even noisy. The imaging depth information is considered as an aided input to help our model make better decision.

1. Introduction

Light always plays an important role in physics, chemistry, and biology of oceans research. The process of the light transmission in the seawater is the foundation of ocean optical research. And the optical properties are the key to describe the light transmission process. The wavelength of the visible light is widely spread from 400 nm to 700 nm. The optical properties of the medium are crucial for the further research of underwater vision, marine organism, pollution detection, and other ocean research areas. The optical properties of ocean can be roughly classified as inherent optical properties (IOPs) and apparent optical properties (AOPs). Important IOPs contain spectral absorption coefficient, , scattering coefficients, , and attenuation coefficient, . For a certain wavelength , these three properties can be simply described as

IOPs only correlate with medium itself and are irrelevant to the ambient light field or its geometric distribution. Measuring these coefficients is fundamental and important to ocean optical research. Beam transmissometers, such as ac-spectra (AC-S) produced by Wetlabs, are the most commonly used devices for IOPs measurement [13]. However, its inconvenient to place such an equipment because of the high price and the limited volume of underwater IOPs measured by the device. Moreover, most of the researchers consider the water as homogeneous medium when measuring IOPs [47]. Actually, a slight turbulence caused by robots, marine pollution, or organisms may lead to an inhomogeneous medium. But the beam transmissometers can only detect the IOPs in the surrounding area. In such a case, an underwater camera with a real-time system will be more flexible to capture efficient information. Besides, there are some other ways to deduct IOPs based on AOPs [8, 9].

AOPs are those properties that depend both on the medium (the IOPs) and on the geometric structure of the radiation distribution. AOPs can be measured by remote sensing. And many researchers prefer to measure IOPs with transmissometers as the ground truth to verify their AOPs based deduction. But such deduction is not accurate enough. On one hand, data based on remote sensing is obtained from satellites or airplane. The detailed information may be omitted. On the other hand, remote sensing cannot investigate the IOPs of undersea. AOPs based on remote sensing can only tell us the IOPs distribution of the surface water. Research also shows that the depth of water may bring difficulty when we want to calculate IOPs [10].

Underwater images are always with scattering and absorption. Because of these reasons, underwater images are always blurred because of different light field and IOPs, which may bring difficulty for us to build an accurate physical model [11]. However, there are many researchers showing that they can recover an underwater image with IOPs [1214] and physical models [15]. That means an underwater image contains plenty of IOPs information. That is why we can restore images with correct IOPs and suitable physical models. If we can estimate IOPs with a single image, it would be much convenient to measure IOPs undersea. Unlike the most of underwater image restoration tasks which aim to reduce the noise caused by scattering and absorption effect [16], the image noisy information will help us to deduct IOPs.

Since Hinton and his colleges proposed the deep learning concept from 2006 [17], deep neural network becomes more and more popular these years. Neural network models are proven not only on computer vision, but also on symmetric recognition, image quality assessment, image restoration, and even optical flow processing [1823]. Thus, we consider if it is possible to use deep neural network for analyzing underwater images to estimate IOPs in this study. Besides, if we want to build an end-to-end system with an image as inputs and IOPs as outputs, deep neural network is a suitable candidate to connect them together. In this paper, we used AC-S to provide 156 IOPs (78 attenuation coefficients and 78 absorption coefficients) as the ground truth for the neural network training.

2. Depth Aided Deep Neural Networks for IOPs Estimation

The framework of our system can be found in Figure 1. We used a color calibration board as the target underwater object and captured its RGB images with a video camera. We also measured the distance between the board and the camera as depth information. The depth information is used as an aided input to the deep convolutional neural network. Convolutional neural network (CNN), which is developed by Lecun and Bengio [24] in 1995, is a powerful model especially in computer vision research. CNN is improved by Krizhevsky and his colleges in 2012 as AlexNet [25]. We follow their idea to build our model for underwater image analysis. However, our inputs and outputs are different from their work.

Figure 1: The framework of our experiment.

Deep neural network can not only recognize what kind of the target object is, but also understand how much blur an image is. But the water quality is not the only factor which causes the image blur. In physical point of view, distance between the camera and the target is also an important factor. The underwater images would be blur if the water quality is low, and the images would also be unclear if the target object is far away from the camera. That is the reason why we considered depth information of the target object as another useful feature.

Unlike some common neural network application tasks such as image classification, IOPs including multiple coefficients are not binary values. Thus, softmax activation function which is usually used for image classification tasks [2628] in the last layer cannot be used for our goal. Instead, we applied min–max normalization and Euclidean loss function for IOPs regression:where is the desired IOP coefficient measured by AC-S, is the IOP coefficient after min–max normalization, is the estimated IOPs by deep neural network, and is the number of IOPs. The AC-S employed can provide 78 absorption and 78 attenuation coefficients. Hence, we have . Equation (3) shows the detail of min–max normalization which is used for depth normalization. We also use this method to normalize the depth information.

We design the Depth Aided (DA) neural network model for IOPs estimation as shown in Figure 2. The AC-S can provide 78 attenuation coefficients and 78 absorption coefficients from 400 nm to 730 nm as network output labels when we capture RGB images. These IOPs are used as the targets of the AlexNet. Because our target object is a flat board, depth information is considered as a single number. Thus there is no need to put the depth information into convolution layer. So we set the depth information as an aided input in the feedforward layer 7. The weights between depth information and feedforward layer 7 are fully connected. Error backpropagation algorithm, stochastic gradient descendent (SGD), and dropout algorithms [29] are used for evolving this network. The feedforward connections of each neurons in feedforward layer 7 can be described as where and are the neuron postsynaptic and presynaptic value of the feedforward layer 7, respectively; is the neural inputs from the feedforward layer 6; is the weights between layer 6 and layer 7, is the depth information; is the weight in depth information; is the bias; and is the activation function. The network weights updating can follow the error backpropagation rule aswhere represents the weight from the neuron to the neuron, is the error function defined in (2), is the weight updating value in the iteration, is the momentum, is the learning rate, and is the weight decay rate which can prevent overfitting.

Figure 2: The Depth Aided (DA) deep neural network structure.

Our datasets are collected in a large water tank as shown in Figures 3 and 4. We put a lifting platform inside the tank to hold the color calibration board. The digital camera is just above the water. The lifting platform can guarantee the board to always be inside the water and vertical to the camera. And it can also change the distance between the color board and the camera accurately. After we get enough data with different distances, we added the aluminium hydroxide into the water to change the water qualities and collect the data again. Meanwhile, we also use AC-S in the water to measure the real-time IOPs as ground truth. We did not use any additional light field in this experiment except indoor diffuse refection.

Figure 3: The framework of our experiment.
Figure 4: Experiment environment.

The data we collected are listed in Tables 1 and 2. Due to the size of the employed water tank (3.6 m (length) × 2.0 m (width) × 1.2 m (depth)), the precise concentration of aluminium hydroxide cannot be directly measured. However, we could estimate the concentration of aluminium hydroxide by using the volume of water filled in the tank and the weight of aluminium hydroxide added for each image collection. The results were given in Table 2. To ensure uniformity of aluminium hydroxide distribution, we used a circulating pump to stir the water before image collection started. Yet, it can be noticed that the average attenuation and absorption of image pack (4) look lower than the value obtained for image pack (3). The reason for that is the first 3 image packs and the remaining 6 image packs were taken on two consecutive days; some aluminium hydroxide settled at the bottom still standing after 10 hours.

Table 1: Datasets description.
Table 2: IOPs description.

We collected 3 groups of datasets. Please note that we do not add any aluminium hydroxide into the water when we take photos in image pack (1). Datasets A and B were taken under similar environments but at different time periods. Dataset C was taken with different distances and different IOPs. Each image pack was captured with a digital video camera during a short period. Lots of researchers used IOPs under 520 nm wavelength as reference properties [30, 31]. The average value of attenuation and absorption coefficients at 520 nm wavelength in each image pack can be found in Table 1. The frame rate of this camera is 25 frames per second. After we take enough images under a certain depth we can modify the distance between the camera and the board by adjusting the lifting platform. When we got enough photos in one pack, we modified the water IOPs by adding aluminium hydroxide and then started capturing the next image pack. The IOPs in one image pack are similar. But we still use the real-time results provided by AC-S as the training label of deep convolutional neural networks. Our camera type is Hikvision 2ZCN3007. It used a 1/2.8′′ Progressive Scan CMOS sensor. The camera can provide videos with resolution of 1080 p. We used the raw RGB camera pictures ( pixels) and chose the center part as the region of interest ( pixels) as shown in Figure 6. And then we resized them into for network training. The input of the neural network used 3 channels for RGB format. We collected images in the daytime. The camera used an automatic exposure system to record images. There is no other additional light source during our experiments except the diffuse reflection. No additional image preprocessing methods are used.

The sample images of 10 different image packs are displayed in Figure 5. Figures 5(a)5(c) are captured in pack (1) under 460 mm, 560 mm, and 660 mm, respectively, and Figures 5(d)5(l) are captured from pack (2) to pack (10) under the same distance (460 mm). The overview of IOPs estimation results can be found in Table 3. We use a single GTX1070 graphic card and Intel i7-6700 to train these networks. We set the learning rate as 0.0001. We use 3 kinds of deep neural network for IOPs estimation evaluation. We waited enough epochs until these networks converged. Cifar-Net, which is improved based on LeNet-5 [22], is used as benchmark for this experiment. Although we waited 100,000 epochs (about 1 hour), the results based on Cifar-Net are still poor even in training set. AlexNet, which costs us about 3 hours on training until we reach 30,000 epochs, performs better. And we get the minimum Euclidean loss if we consider the depth information. Lower loss means estimated IOPs are closer to the ground truth. That means depth information is helpful especially in clean medium. The DA Net costs about 3 and a half hours for 30,000 epochs. The training speed of our model is a little slower than AlexNet, but with better performance.

Table 3: IOPs estimation results.
Figure 5: Images captured under different situations. (a)–(c) are captured without any aluminium hydroxide (image pack (1)) under 460 mm, 560 mm, and 660 mm. (d)–(l) are captured under 460 mm corresponding to image packs (2)–(10).
Figure 6: The ROI position.

The detail of IOPs estimation results can be found in Figure 7. We choose 3 typical RGB images representing images captured in high, medium, and low turbidity, respectively. The performance of Cifar-Net is shown using blue lines; the regression curve of AlexNet is displayed with red dashed lines; the DA Net is shown with green asterisks and the ground truth provided by AC-S is represented using purple dot lines. In high turbidity case, both AlexNet and our method perform well on attenuation regression. A small amount error existed in blue-purple band (400–450 nm). Our method performs better than AlexNet in both absorption and attenuation coefficients regression task. In medium and low turbidity case, the curves of the DA Net get closer to the ground truth comparing with AlexNet. Although Cifar-Nets show a generally right regression result on three cases, its performance is much lower than the other two methods.

Figure 7: IOPs estimation results. (a)–(f) show the IOPs estimation results based on 3 typical images corresponding to high, medium, and low turbidity. (a) and (b) show the attenuation and absorption coefficients regression results based on Figure 5(i). (c) and (d) show the coefficients regression results based on Figure 5(f). (e) and (f) show the coefficients regression results based on Figure 5(d).

3. Discussion and Conclusion

In summary, we propose a DA deep neural network for IOPs estimation method based on a single RGBD image with a DA deep neural network. We argue that an underwater image contains enough IOPs information that is even noisy. So we are able to deduct IOPs on a single RGBD image with a suitable system. Comparing with traditional methods based on transmissometers, our method can archive enough accuracy but cost-effectively and more flexibly than traditional devices. Our method is able to predict both attenuation and absorption coefficients of the medium simultaneously. The experimental results in Table 3 show that even a single RGB image seems enough for IOPs estimation with deep learning technologies. We can get better estimation results if we consider depth information as an aided input.

In our experiment, we did not consider any complex light field conditions and target objects with complex shape case. These factors may bring difficulty to measure IOPs when we want to put this system in an opening environment. Fortunately, research on deep neural network shows that it is possible to estimate a depth map on a complex target object and even under different light fields with a single RGB image [32, 33]. That may be a possible solution for us to improve our model. On the other hand, back-scattering coefficients, which cannot be measured by AC-S, are very important to build an underwater image recovering model. How to estimate back-scattering coefficients is another challenge. We wish to leave these two parts in our future work.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Authors’ Contributions

Zhibin Yu and Yubo Wang contributed equally to this work.

Acknowledgments

This work was supported by the National Natural Science Foundation of China under Grant no. 61701463, Natural Science Foundation of Shandong Province of China under Grant no. ZR2017BF011, the Fundamental Research Funds for the Central Universities under Grant nos. 201713017 and 201713019, and the Qingdao Postdoctoral Science Foundation of China.

References

  1. J. H. Steele, S. A. Thorpe, and K. K. Turekian, Measurement Techniques, Sensors and Platforms - A Derivative of Encyclopedia of Ocean Sciences, vol. 272, 2009.
  2. H. Loisel, D. Stramski, B. G. Mitchell et al., “Comparison of the ocean inherent optical properties obtained from measurements and inverse modeling,” Applied Optics, vol. 40, no. 15, pp. 2384–2397, 2001. View at Publisher · View at Google Scholar · View at Scopus
  3. E. Rehm and N. J. McCormick, “Inherent optical property estimation in deep waters,” Optics Express, vol. 19, no. 25, pp. 24986–25005, 2011. View at Publisher · View at Google Scholar · View at Scopus
  4. H.-H. Lu, C.-Y. Li, H.-H. Lin et al., “An 8 m/9.6 Gbps underwater wireless optical communication system,” IEEE Photonics Journal, vol. 8, no. 5, 2016. View at Publisher · View at Google Scholar · View at Scopus
  5. R. Schettini and S. Corchs, “Underwater image processing: State of the art of restoration and image enhancement methods,” EURASIP Journal on Advances in Signal Processing, vol. 2010, Article ID 746052, 2010. View at Publisher · View at Google Scholar · View at Scopus
  6. W. Hou, D. J. Gray, A. D. Weidemann, G. R. Fournier, and J. L. Forand, “Automated underwater image restoration and retrieval of related optical properties,” in Proceedings of the 2007 IEEE International Geoscience and Remote Sensing Symposium, IGARSS 2007, pp. 1889–1892, Spain, June 2007. View at Publisher · View at Google Scholar · View at Scopus
  7. J. S. Jaffe, “Underwater Optical Imaging: The Past, the Present, and the Prospects,” IEEE Journal of Oceanic Engineering, vol. 40, no. 3, pp. 683–700, 2015. View at Publisher · View at Google Scholar · View at Scopus
  8. N. L. Swanson, B. D. Billard, V. M. Gehman, and T. L. Gennaro, “Application of the small-angle approximation to ocean water types,” Applied Optics, vol. 40, no. 21, pp. 3608–3613, 2001. View at Publisher · View at Google Scholar · View at Scopus
  9. L. Sipelgas and U. Raudsepp, “Comparison of hyperspectral measurements of the attenuation and scattering coefficients spectra with modeling results in the north-eastern Baltic Sea,” Estuarine, Coastal and Shelf Science, vol. 165, pp. 1–9, 2015. View at Publisher · View at Google Scholar · View at Scopus
  10. D. McKee, A. Cunningham, J. Slater, K. J. Jones, and C. R. Griffiths, “Inherent and apparent optical properties in coastal waters: A study of the Clyde Sea in early summer,” Estuarine, Coastal and Shelf Science, vol. 56, no. 2, pp. 369–376, 2003. View at Publisher · View at Google Scholar · View at Scopus
  11. A. Cunningham, L. Ramage, and D. McKee, “Relationships between inherent optical properties and the depth of penetration of solar radiation in optically complex coastal waters,” Journal of Geophysical Research: Oceans, vol. 118, no. 5, pp. 2310–2317, 2013. View at Publisher · View at Google Scholar · View at Scopus
  12. G. Wang, B. Zheng, and F. F. Sun, “Estimation-based approach for underwater image restoration,” Optics Expresss, vol. 36, no. 13, pp. 2384–2386, 2011. View at Publisher · View at Google Scholar · View at Scopus
  13. B. Huang, T. Liu, H. Hu, J. Han, and M. Yu, “Underwater image recovery considering polarization effects of objects,” Optics Express, vol. 24, no. 9, pp. 9826–9838, 2016. View at Publisher · View at Google Scholar · View at Scopus
  14. H. Mortazavi, J. P. Oakley, and B. Barkat, “Mitigating the effect of optical back-scatter in multispectral underwater imaging,” Measurement Science and Technology, vol. 24, no. 7, Article ID 074025, 2013. View at Publisher · View at Google Scholar · View at Scopus
  15. J. S. Jaffe, “Computer modeling and the design of optimal underwater imaging systems,” IEEE Journal of Oceanic Engineering, vol. 15, no. 2, pp. 101–111, 1990. View at Publisher · View at Google Scholar · View at Scopus
  16. R. Wang, Y. Wang, J. Zhang, and X. Fu, “Review on underwater image restoration and enhancement algorithms,” in Proceedings of the 7th International Conference on Internet Multimedia Computing and Service, ICIMCS 2015, pp. 277–282, China, August 2015. View at Publisher · View at Google Scholar · View at Scopus
  17. G. E. Hinton, S. Osindero, and Y.-W. Teh, “A fast learning algorithm for deep belief nets,” Neural Computation, vol. 18, no. 7, pp. 1527–1554, 2006. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  18. B. Cai, X. Xu, K. Jia, C. Qing, and D. Tao, “DehazeNet: an end-to-end system for single image haze removal,” IEEE Transactions on Image Processing, vol. 25, no. 11, pp. 5187–5198, 2016. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  19. J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '15), pp. 3431–3440, IEEE, Boston, Mass, USA, June 2015. View at Publisher · View at Google Scholar
  20. W. Hou, X. Gao, D. Tao, and X. Li, “Blind image quality assessment via deep learning,” IEEE Transactions on Neural Networks and Learning Systems, vol. 26, no. 6, pp. 1275–1286, 2015. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  21. L. Kang, P. Ye, Y. Li, and D. Doermann, “Convolutional neural networks for no-reference image quality assessment,” in Proceedings of the 27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014, pp. 1733–1740, USA, June 2014. View at Publisher · View at Google Scholar · View at Scopus
  22. S. Bianco, L. Celona, P. Napoletano, and R. Schettini, “On the use of deep learning for blind image quality assessment,” Signal, Image and Video Processing, pp. 1–8, 2017. View at Publisher · View at Google Scholar
  23. F. Weber, H. Eichner, H. Cuntz, and A. Borst, “Eigenanalysis of a neural network for optic flow processing,” New Journal of Physics , vol. 10, Article ID 015013, 2008. View at Publisher · View at Google Scholar · View at Scopus
  24. Y. Lecun and Y. Bengio, The Handbook of Brain Theory and Neural Networks, vol. 255, 1995.
  25. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proceedings of the 26th Annual Conference on Neural Information Processing Systems (NIPS '12), pp. 1097–1105, Lake Tahoe, Nev, USA, December 2012. View at Scopus
  26. T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean, “Distributed representations of words and phrases and their compositionality,” in Proceedings of the 27th Annual Conference on Neural Information Processing Systems (NIPS '13), pp. 3111–3119, December 2013. View at Scopus
  27. R. Socher, C. C.-Y. Lin, C. D. Manning, and A. Y. Ng, “Parsing natural scenes and natural language with recursive neural networks,” in Proceedings of the 28th International Conference on Machine Learning (ICML '11), pp. 129–136, Bellevue, Wash, USA, June 2011. View at Scopus
  28. N. Kalchbrenner, E. Grefenstette, and P. Blunsom, “A Convolutional Neural Network for Modelling Sentences,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 655–665, Baltimore, Md, USA, June 2014. View at Publisher · View at Google Scholar
  29. N. Kalchbrenner, E. Grefenstette, and P. Blunsom, “A convolutional neural network for modelling sentences,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014, pp. 655–665, June 2014. View at Scopus
  30. D. C. English and K. L. Carder, “Determining bottom reflectance and water optical properties using unmanned underwater vehicles under clear or cloudy skies,” Journal of Atmospheric and Oceanic Technology, vol. 23, no. 2, pp. 314–324, 2006. View at Publisher · View at Google Scholar · View at Scopus
  31. H. M. Oubei, C. Li, K. Park, T. K. Ng, M. Alouini, and B. S. Ooi, “2.3 Gbit/s underwater wireless optical communications using directly modulated 520 nm laser diode,” Optics Express, vol. 23, no. 16, pp. 20743–20748, 2015. View at Publisher · View at Google Scholar
  32. D. Eigen, C. Puhrsch, and R. Fergus, “Depth map prediction from a single image using a multi-scale deep network,” in Proceedings of the 28th Annual Conference on Neural Information Processing Systems 2014, NIPS 2014, pp. 2366–2374, December 2014. View at Scopus
  33. F. Liu, C. Shen, and G. Lin, “Deep convolutional neural fields for depth estimation from a single image,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, pp. 5162–5170, USA, June 2015. View at Publisher · View at Google Scholar · View at Scopus