About this Journal Submit a Manuscript Table of Contents
Computational and Mathematical Methods in Medicine
Volume 2013 (2013), Article ID 208402, 9 pages
Research Article

Nonlinear Radon Transform Using Zernike Moment for Shape Analysis

1School of Information and Technology, Northwest University, Xi'an 710120, China
2School of Information and Computing Sciences, North University for Nationalities, Yinchuan 750021, China
3College of Computing & Communication Engineering, Graduate University of Chinese Academy of Sciences, Beijing 100049, China
4College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China

Received 18 January 2013; Accepted 22 March 2013

Academic Editor: Carlo Cattani

Copyright © 2013 Ziping Ma et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


We extend the linear Radon transform to a nonlinear space and propose a method by applying the nonlinear Radon transform to Zernike moments to extract shape descriptors. These descriptors are obtained by computing Zernike moment on the radial and angular coordinates of the pattern image's nonlinear Radon matrix. Theoretical and experimental results validate the effectiveness and the robustness of the method. The experimental results show the performance of the proposed method in the case of nonlinear space equals or outperforms that in the case of linear Radon.

1. Introduction

Shape analysis methods have been broadly applied to biomedical signal processing, object recognition, image retrieval, target tracking, and so forth [1]. Moments methods [2, 3] can be referred to shape descriptors because of their good characterization in describing different shapes. The most important properties of shape descriptors achieved by different moments are invariance, including translation, rotation, scaling, and stretching, stability to noise, and completeness [4].

In the past twenty years, many attentions have been paid to the completeness property of the invariant descriptor set in pattern recognition and other similar application fields. These kinds of methods can be obtained by the following processes. Firstly, Fourier transform or Radon transform is employed to map the image into other space. Secondly, the different ideas can be conceived to construct invariant descriptors based on the information in new space. Sim et al. [5] gave a new method for texture image retrieval. They converted the images in Fourier domain and calculated modified Zernike moments to extract the texture descriptors. It is tested that the descriptor has higher accuracy comparing to Gabor, Radon, and wavelet based methods and requires low computational effort. However, it is not invariant to scale. Wang et al. [6] and Xiao et al. [7] introduced the Radon transform to Fourier-Mellin transform to achieve RST (rotation, scaling, and translation) invariance and RS invariance combined blur, respectively. In virtue of Xiao’s idea, Zhu et al. [8] constructed RST invariants using Radon transforms and complex moments in digital watermarking. Similarly, the Zernike moments can be connected with Radon transform. Rouze et al. [9] described a method to design an approach by calculating the Zernike moments of an image from its Radon transform using a polynomial transform in the position coordinate and a Fourier transform in the angular coordinate. However, the proposed descriptors are only invariant to rotation. Meanwhile, in order to improve the precision of image retrieval and noise robustness, Hoang and Tabbone [10] proposed a new method similar to Xiao’s descriptor to obtain RST invariance based on the Radon, Fourier, and Mellin transform.

Then, Radon transform is widely applied in many methods mainly because of its better properties in projection space [1115]. In the projective space, a rotation of the original image results in a translation in the angle variable, and a scaling of the original image leads to a scaling in the spatial variable together with an amplitude scaling [16, 17]. Based on these properties, a rotation and scaling invariant function is easy to construct and highly robust to noise.

Enlightened by the peers’ research works, we extend Radon transform to nonlinear Radon transform and propose a new set of complete invariant descriptors by applying Zernike moments to the radial coordinate of the pattern’s nonlinear Radon space of an image [1822].

The remainder of this paper is organized as follows. In Section 2, we briefly review the definition of nonlinear Radon transform and Zernike moments, and propose a new method based on Zernike moment and nonlinear Radon transform. In Section 3, the comparative experiments of the proposed approach with Hu moment invariance, Chong’s method is conducted in terms of image retrieval efficiency, different noise robustness. Section 4 concludes this paper.

2. Nonlinear Radon Transform and Zernike Moments

2.1. Nonlinear Radon Transform

The nonlinear Radon transform of an image function is defined as [10] where , is a real instance, denotes the angle vector formed by the function , and is a rotation function by with an angel of and defined by

The nonlinear Radon transform indicates curve integral of the image function along different curves. The parameter can control the shape of curve. Different curves can be obtained by the values of parameter and function .

Especially when and , . This reveals that the linear Radon transform is the special case of nonlinear Radon transform. The results of different curves’ Radon transform are shown in Table 1.

Table 1: The diagrams of results using different curves’ Radon transform.

The nonlinear Radon transform has some properties that are beneficial for pattern recognition as outlined below.(1) Periodicity: the nonlinear Radon transform of is periodic in the variable with period when is an arbitrarily parametric inference, (2) Resistance: if and are two images with little difference when is arbitrarily parametric inference, the corresponding nonlinear Radon transform of and are as followes: (3) Translation: a translation of by a vector results in a shift in the variable of by a distance and equals to the length of the projection of onto the line , (4) Rotation: a rotation of by an angle implies a shift in the variable of by a distance when is arbitrarily parametric inference, (5) Scaling: a scaling of by a factor of results in a scaling in the variable and of amplitude of , respectively, when represents ellipse or hyperbola curve,

2.2. Zernike Moment

The radial Zernike moments of order of an image function , is defined as where the radial Zernike moment of order is defined by the following equation: With

2.3. NRZM Descriptor Based on Nonlinear Radon Transform and Zernike Moment

The Zernike moment is carried out to be computed after the projective matrix of nonlinear Radon transform is mapped to the polar coordinate (NRZM). The computational process of our proposed method, NRZM, is illuminated in Figure 1.

Figure 1: The computation process of NRZM.

Supposed is the image rotated by rotational angle and scaled by scaling factor , and Radon transform of is given by The Zernike moments of is The radial Zernike polynomials can be expressed as a series of as follows: The derivation process of (13) is given in the Appendix. According to (12), we have Let , (14) can be rewritten as

Equation (15) shows that the radial Zernike moments of being rotated image can be expressed as a linear combination of the radial Zernike moments of original image. Based on this relationship, we can construct a set of rotation invariant which is described as follows: Then, is invariant to rotation and translation.

3. Experimental Results and Discussions

This section is intended to test the performance of a complete family of similarity invariants introduced in the previous section for images retrieval by comparison, Chong’s method presented in [12], Hu moment presented in [13]. In the experiments, the feature descriptors are calculated by

Three subsections are included in this section. In the first subsection, we test the retrieval efficiency of proposed descriptors in shape 216 dataset. This dataset is composed of 18 shape categories with 12 samples per category, and each of every category cannot be obtained by RST transforming from any other shape from the same category. In the second subsection, we test robustness of proposed descriptors in different noisy dataset. In the third subsection, we verify the rotation invariance of the proposed method.

3.1. Experiment 1

The kind of curves is changing with the controlled parameters varying. So, the retrieval efficiency is different with the controlled parameters. Many experiments are conducted to find the best parameters’ values of every curve in nonlinear Radon transform, and finally the most suitable values of parameters are listed in Table 2. In the subsequent experiments, we analyze the retrieval efficiency of linear Radon transform, ellipse Radon transform, hyperbola Radon transform, and parabola Radon transform with Zernike moment, respectively, which is referred to as NZ, EPZ, HPZ, and PRZ, respectively.

Table 2: The most suitable values of parameters.

In order to obtain the best retrieval efficiency of every curve Radon, the comparative precisions-recall curves in Shapes 216 are shown in Figure 2. It can be seen that the precision-recall curve of PRZ moves downward more slowly than those of others, which indicates that the retrieval efficient of PRZ is slightly higher than that of RZ while HRZ is weaker than PRZ and RZ.

Figure 2: The precision-recall curve of shape 216.

The comparative number of relevant image upon every category is a better insight into the performance of proposed method as shown in Figure 3. It is easy to see that almost the number of relevant image in every category is higher than 6, especially in bird, children, elephant, face, glass, hammer, heart, and misk.

Figure 3: The retrieved number of every category in shape 216.
3.2. Experiment 2

The robustness of the proposed descriptors is demonstrated using eight datasets added additive “salt & pepper” and “Gaussian” noise, respectively. The first seven datasets are generated from original shape 216 database, and each image is corrupted by “salt & pepper” noise with SNR varying from 16 to 4 dB with 2 dB decrements. The last one is generated from shape 216 added “Gaussian” noise with noise density = 0.01, …, 0.2.

The retrieval experiments are conducted again in the datasets mentioned above and the precision-recall curves of comparative descriptors are depicted in Figure 4. From Figures 4(a)4(g), it can be observed that efficiency of the PRZ and RZ are similar. It also can be seen that the PRZ and RZ descriptors have better performances than other comparative methods in “salt and pepper” noisy datasets from SNR = 16 to 8, while Hu moment and Chong’s descriptors have similarly the worse performance. However, when SNR = 6 and SNR = 4, the situation has changed. The deterioration appears in the PRZ and RZ because their precision-recall curves moves downward more rapidly than those of HPZ and EPZ, while they move downward more slowly than those of Chong’s method and CMI. This demonstrates that PRZ and RZ descriptor are sensitive than other nonlinear methods’ descriptors when the value of SNR is low of 8 though it has the stronger robustness than Chong’s method and Hu moment. In short, the impact of noise on RZ, ERZ, HRZ, and PRZ curves sometimes were little similar or sometimes differ from one to another. It is also observed that(1) as the values of SNR decrease, the curves of all the descriptors generally move downwards; (2) Hu moment and Chong’s descriptors are very sensitive to noise, and their performance has not changed much under different levels of noise; (3) Hu moment method has more resistance to “salt & pepper” noise than Chong’s descriptors; (4) among the RZ, ERZ, PRZ, and HRZ, the resistance of PRZ is the strongest to “salt & pepper” noise and that of RZ is close to PRZ when the values of SNR are higher than 6; (5) PRZ is always slightly more robust to “salt & pepper” noise than RZ except for SNR = 6 and SNR = 4;(6) EPZ and HPZ descriptors are more robust to “salt & pepper” noise than PRZ and RZ when values of SNR are higher than 6.

Figure 4: The precision upon recall curves of different descriptors on seven noisy datasets added “salt & pepper” and one “Gaussian’’ noisy dataset.

However, the retrieval results shown in Figure 4(h) are essentially different from those in Figures 4(a)4(g). It is clear that ERZ and HRZ are more robust to “Gaussian” noise than other methods because their precision-recall curves are absolutely on the top of others in the “Gaussian” noisy dataset. This indicates that “Gaussian” noise can result in poor performance in the case of linear transform. In these cases, the nonlinear Radon transform should be a top priority to be employed in the proposed method.

3.3. Experiment 3

The last test dataset is color objective dataset generated by choosing 7 sample images from Col and View subset. Each of the datasets is transformed by being rotated by 72 arbitrary angles (10–360) with 5 degree increment. As a result, the last dataset consists of 504 images, and the retrieval results are shown in Figure 5. From the figure, it can be concluded that the proposed descriptors are invariant to rotation, and the retrieval performance of PRZ is more efficient.

Figure 5: The precision-recall curves of different descriptors on rotated dataset.

4. Conclusion

In this paper, we proposed a method to derive a set of rotation invariants using Radon transform and Zernike moments and extend linear Radon transform to nonlinear Radon transform.

Comparing to linear Radon transform, the proposed method can perform better or similar. However, the numerical experiments show that different curve Radon transforms and Zernike moment perform different. In the noiseless dataset, the retrieval efficiency of PRZ is higher than comparative methods. In the “salt & pepper” noise and the PRZ consistently performs better except SNR = 6 and SNR = 4. While when SNR = 6, SNR = 4, the EPZ and HPZ are most robust than RZ. And in “Gaussian” noise dataset, the proposed method in the cases of nonlinear Radon transform is more robust to “Gaussian” noise than that in the case of linear Radon transform. Moreover, the nonlinear Radon transform can be exploited to other application fields for engineer application and recognition for the sake of the good characteristic, especially their robustness.


Proof of (13)

From (12), the radial Zernike polynomials can be expressed as a series of decreasing power of as follows: Since all the diagonal element are not zero, the matrix is nonsingular, thus


This work is supported by the National Natural Science Foundation of China under Grant no. 61261043 and 61102008, College Scientific research project of Ningxia province (no. NGY2012147). The authors would like to thank the anonymous referees for their valuable comments and suggestions.


  1. Z. Teng, J. He, A. J. Degnan, et al., “Critical mechanical conditions around neovessels in carotid atherosclerotic plaque may promote intraplaque hemorrhage,” Atherosclerosis, vol. 223, no. 2, pp. 321–326, 2012. View at Publisher · View at Google Scholar
  2. S. Y. Chen, J. Zhang, Q. Guan, and S. Liu, “Detection and amendment of shape distortions based on moment invariants for active shape models,” IET Image Processing, vol. 5, no. 3, pp. 273–285, 2011. View at Publisher · View at Google Scholar · View at Scopus
  3. J. Wood, “Invariant pattern recognition: a review,” Pattern Recognition, vol. 29, no. 1, pp. 1–17, 1996. View at Publisher · View at Google Scholar
  4. F. Ghorbel, S. Derrode, R. Mezhoud, T. Bannour, and S. Dhahbi, “Image reconstruction from a complete set of similarity invariants extracted from complex moments,” Pattern Recognition Letters, vol. 27, no. 12, pp. 1361–1369, 2006. View at Publisher · View at Google Scholar · View at Scopus
  5. D. G. Sim, H. K. Kim, and R. H. Park, “Invariant texture retrieval using modified Zernike moments,” Image and Vision Computing, vol. 22, no. 4, pp. 331–342, 2004. View at Publisher · View at Google Scholar · View at Scopus
  6. X. Wang, F. X. Guo, B. Xiao, and J. F. Ma, “Rotation invariant analysis and orientation estimation method for texture classification based on Radon transform and correlation analysis,” Journal of Visual Communication and Image Representation, vol. 21, no. 1, pp. 29–32, 2010. View at Publisher · View at Google Scholar · View at Scopus
  7. B. Xiao, J. Ma, and J. T. Cui, “Combined blur, translation, scale and rotation invariant image recognition by Radon and pseudo-Fourier-Mellin transforms,” Pattern Recognition, vol. 45, no. 1, pp. 314–321, 2012. View at Publisher · View at Google Scholar
  8. H. Q. Zhu, M. Liu, and Y. Li, “The RST invariant digital image watermarking using Radon transforms and complex moments,” Digital Signal Processing, vol. 20, no. 6, pp. 1612–1628, 2010. View at Publisher · View at Google Scholar · View at Scopus
  9. N. C. Rouze, V. C. Soon, and G. D. Hutchins, “On the connection between the Zernike moments and Radon transform of an image,” Pattern Recognition Letters, vol. 27, no. 6, pp. 636–642, 2006. View at Publisher · View at Google Scholar · View at Scopus
  10. T. V. Hoang and S. Tabbone, “Invariant pattern recognition using the RFM descriptor,” Pattern Recognition, vol. 45, no. 1, pp. 271–284, 2012. View at Publisher · View at Google Scholar
  11. S. R. Deans, The Radon Transform and Some of Its Applications, Wiley, New York, NY, USA, 1983.
  12. H. P. Hiriyannaiah and K. R. Ramakrishnan, “Moments estimation in Radon space,” Pattern Recognition Letters, vol. 15, no. 3, pp. 227–234, 1994. View at Scopus
  13. R. R. Galigekere, D. W. Holdsworth, M. N. S. Swamy, and A. Fenster, “Moment patterns in the Radon space,” Optical Engineering, vol. 39, no. 4, pp. 1088–1097, 2000. View at Publisher · View at Google Scholar · View at Scopus
  14. F. Peyrin and R. Goutte, “Image invariant via the Radon transform,” in Proceedings of the IEEE International Conference on Image Processing and its Applications, pp. 458–461, 1992.
  15. J. Flusser and T. Suk, “Degraded image analysis: an invariant approach,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20, no. 6, pp. 590–603, 1998. View at Publisher · View at Google Scholar
  16. C. W. Chong, P. Raveendran, and R. Mukundan, “Translation and scale invariants of Legendre moments,” Pattern Recognition, vol. 37, no. 1, pp. 119–129, 2004. View at Publisher · View at Google Scholar · View at Scopus
  17. X. Zhang, Y. Zhang, J. Zhang, X. Li, S. Chen, and D. Chen, “Unsupervised clustering for logo images using singular values region covariance matrices on Lie groups,” Optical Engineering, vol. 51, no. 4, 8 pages, 2012. View at Publisher · View at Google Scholar
  18. M. K. Hu, “Visual pattern recognition by moments invariants,” IRE Transactions on Information Theory, vol. 8, no. 2, pp. 179–187, 1962.
  19. T. B. Sebastian, P. N. Klein, and B. B. Kimia, “Recognition of shapes by editing their shock graphs,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 5, pp. 550–571, 2004. View at Publisher · View at Google Scholar · View at Scopus
  20. http://staff.science.uva.nl/~aloi/.
  21. H. Zhu, M. Liu, H. Ji, and Y. Li, “Combined invariants to blur and rotation using Zernike moment descriptors,” Pattern Analysis and Applications, vol. 13, no. 3, pp. 309–319, 2010. View at Publisher · View at Google Scholar · View at Scopus
  22. http://museumvictoria.com.au/bioinformatics/butter/images/bthumbliv.htm.