Abstract

This paper proposes a calibration technique of a stereo gamma detection camera. Calibration of the internal and external parameters of a stereo vision camera is a well-known research problem in the computer vision society. However, few or no stereo calibration has been investigated in the radiation measurement research. Since no visual information can be obtained from a stereo radiation camera, it is impossible to use a general stereo calibration algorithm directly. In this paper, we develop a hybrid-type stereo system which is equipped with both radiation and vision cameras. To calibrate the stereo radiation cameras, stereo images of a calibration pattern captured from the vision cameras are transformed in the view of the radiation cameras. The homography transformation is calibrated based on the geometric relationship between visual and radiation camera coordinates. The accuracy of the stereo parameters of the radiation camera is analyzed by distance measurements to both visual light and gamma sources. The experimental results show that the measurement error is about 3%.

1. Introduction

Detection of radioactive sources is one of the important topics in many research areas, particularly in nuclear instrumentation, nuclear physics, nuclear medicine, and many nondestructive tests [1]. Radiation detectors, also known as particle detectors, are used to detect, track, and identify the presence of radioactive sources within a given area or environment. As a generic definition, radiation is an energy that comes from a source and travels through some material or space. They can be found in many different formats such as alpha, beta, gamma, and X-ray. The detection of sources that emit radiation plays a very significant role in many areas of science and industry. Mainly, gamma radiation has managed to attract the interest of most of the scientists since the advance of the Anger camera [2].

Not only the detection of radioactive sources is crucial, but also the accurate estimation of their three-dimensional (3D) information is important. If the 3D position information of radioactive (such as gamma) sources is available, their activity (strength) can be estimated more precisely. However, detection of gamma sources which are widely distributed in the 3D environment is not an easy task to accomplish as it may seem. Determining or estimating the 3D coordinates of such widely spread gamma sources using a single camera is almost impossible. Therefore, in this paper, we use a stereo gamma detection device instead of using mono devices to estimate their coordinates precisely.

When detecting and extracting metric information from two-dimensional (2D) radiation images using a stereo gamma camera (SGC), we must consider the calibration of the camera. Proper calibration methods have to be implemented in order to obtain more accurate 3D coordinate measurement. Since accurate calibration of vision cameras is of interest in computer vision, many related works have been done throughout the last few decades [37].

Although much work related to the calibration of stereo and mono vision camera has been introduced, a proper calibration method of an SGC has not been introduced yet. Considering it as a fact, we propose a new simplified method to calibrate such SGC devices using planar homography relationships. The experiment system consists of a 1D gamma sensor and a 2D vision sensor, which are mounted on a pan-tilt motion device. Because the gamma sensor is a 1D point sensor, the pan-tilt device scans the sensor in horizontal and vertical directions to acquired 2D gamma images. In addition, the pan-tilt device also rotates both the gamma sensor and vision sensor in a symmetric position so that each sensor can obtain stereo images with the same baseline. Therefore the gamma sensor combined with the pan-tilt device works as a 2D radiation camera and the vision camera is used to acquire the corresponding 2D visual information.

To measure 3D information from the stereo gamma images, the stereo gamma camera must be calibrated. However, no visual information can be obtained from the camera, and we transform the visual images obtained from the vision camera to the view of the gamma camera. In this way, we can obtain virtual calibration images of the gamma camera to calibrate the stereo geometry of the gamma camera. Once the devices are fully calibrated, we can measure the parallax in the stereo gamma images and distance to the radioactive sources.

The paper is constructed as follows. Section 2 introduces the panning and tilting method implemented to generate radiation images. Section 3 describes the method used to calculate the planar homography relationships between gamma camera images and vision camera images. The proposed calibration method is described in Section 4. Section 5 presents the experiments we have performed to measure 3D distance to both visual light and gamma sources. In this section, we evaluate the accuracy of the proposed calibration method by comparing the acquired results with real data. Section 6 consists with the conclusion.

2. Obtaining Stereo Gamma Images Using a Gamma Detector

The approach of obtaining gamma images is based on scanning a 1D gamma detector using a pan-tilt motion module. Figure 1 shows a simplified diagram of this concept. A 1D gamma detector is mounted on a pan-tilt module which is connected to a general purpose PC. The rotation of the pan-tilt module in panning and tilting directions is controlled by a terminal program of the PC. According to the resolution of a gamma image, the sensing direction of the pan-tilt module is determined. At every sensing direction, the detector obtains the corresponding pixel’s grey value of the image. By obtaining all pixel values, a 2D gamma image is then generated. Therefore, we can consider that there is a virtual 2D gamma camera which can capture a gamma image.

As a gamma detector, a SPMT (small photo multiplier tube) is used with a Nal (Tl) type scintillator. The detection sensitivity of the SPMT is the maximum at 420 nm. The output type of the detector is the pulse count of photons, which are generated by the SPMT. For more information, see reference [8]. The pulse counts from the SPMT are recorded and mapped to the grey value of a pixel in the gamma image.

A 2D vision camera is also mounted on the opposite side of the gamma detector. The vision camera only captures two ways of visual information, the left and the right images, of the scanning area. The viewing angle of the vision camera is determined by the lens of the camera and it is usually different than the scanning area of the detector. However, it can be transformed to be matched with the scanning area of the detector by using the calibration parameters of the camera.

Figure 2 shows a CAD model and a real view of the pan-tile module equipped with two enclosures for mounting gamma and vision camera. The pan part of the module rotates the tilt part and two sensing devices more than 180 degrees. In addition, the tilt part rotates the gamma detector and the vision camera more than 180 degrees also. Figure 3 shows diagrams to explain the way of capturing stereo images of two sensing devices. In the beginning, suppose the gamma camera is on the right side of the pan-tile module and the vision camera is on the left side. Therefore, we can obtain the right gamma image () from the virtual gamma camera and the left vision image () from the vision camera. Next, let us rotate the pan and tilt parts 180 degrees, respectively. Then the origin of the sensor coordinate systems is exactly exchanged, while the image planes of the cameras are rotated 180 degrees with respect to the image center. The image rotation can be easily rotated to the right directions; finally, we can obtain the left gamma image () and the right vision image () in the final position.

In order to rotate the pan-tilt module accurately, we control the module using very accurate DC motors and a motion controller. The motors are equipped with encoders with the resolution of 2048 pulses per revolution. In addition the rotation of the pan and tilt module is reduced by 100 : 1 gear ratio. Therefore, the rotation accuracy is very high and we assume the error is less than 0.01 degrees. Table 1 shows the specification of some categories of the pan-tile module.

3. Homography Relationship between Gamma and Vision Cameras

The first step of the proposed method is calculating the homography relationship between two different image planes of the gamma and vision cameras. Currently two types of gamma detectors are used: pin-hole and coded aperture types [9]. As described in the previous section, our gamma sensor is a pin-hole detector which consists of a cylindrical collimator and an SPMT. Assuming that the gamma and vision cameras follow the pinhole camera model, the image formation between the gamma and vision cameras is related by a homography. In the previous section, we show that the coordinate origins of the gamma and vision cameras are the same because of the symmetric motion of the pan-tilt module. There are two homography relations between gamma and vision images: one is between left gamma and vision images ( and ) and the other is between right gamma and vision images ( and ).

Let and be the left and the right homography of the gamma and vision cameras, respectively. As shown in Figure 4, is the homography from to , and is the homography from to . To transform the vision images to gamma images, and are multiplied by and , respectively, as follows, and then the results are homogeneously equal to the gamma image: Because two sensor’s images captured from the same capture position, for example, from and , have the same coordinate origin, 2D homography relation is enough to transform one image plane to the other.

To calculate the planar homography matrix , we need at least four corresponding image points between the gamma and vision images. This means that we need to know at least four sets of 2D image coordinates of any object which can be captured in both gamma and vision cameras. Fortunately, the gamma detection sensor is designed to respond to visible light. Thus we can use four-point light sources to find the homography.

Figure 5 shows an example of capturing four corresponding image points between the gamma and vision cameras. Four LED light sources are fixed in front of the cameras as calibration objects and gamma and vision images of the light sources are obtained. Both in gamma and vision images, the image intensity of the LED sources is higher than those of their neighboring pixels. We apply the Gaussian fitting method [10] to get the exact coordinates of the corresponding points and calculate the homography and the homography using a linear algebra equation.

and are 3 × 3 matrices, but their scale is unknown. Therefore, using four pairs of matching points between gamma and vision images, we can compute eight unknown parameters of the homography. Equation (2) shows homography and homography computed by the images in Figure 5:

Figure 6 shows an example of applying the and to the vision images. In Figure 6(a), left and right vision images and are captured in 640 × 480 resolution. Using two homography matrices, virtual gamma images and are obtained in 400 × 400 resolution. Each pair of left and right images shares the same coordinate origin; thus, there are only scale and rotation transformation between the vision and gamma images (ideally speaking, there is no rotation transformation if the coordinate axes of the two sensors are parallel).

In Figure 7, more light sources are used to calibrate the homography between the two stereo cameras. Images of eight LED light points are captured in the images from both the vision and gamma cameras. Images in Figure 7(a) show the original captured images of light points from the vision and gamma cameras. Homography relation between the left and right sensors is shown in (3). In Figure 7(c) are virtual gamma images obtained from the homography relation:

4. Stereo Calibration of Vision and Gamma Cameras

In the field of computer vision, calibration of a stereo vision camera is a general but important task. Many calibration algorithms have been introduced; however, currently Zhang’s method is widely used. To use Zhang’s calibration method, it is recommended to capture at least six images of a calibration pattern which usually consists of checkerboard patterns. Zhang’s calibration algorithm is available in several computer vision libraries such as OpenCV and MATLAB. In this paper, we also use the OpenCV computer vision library for calibrating the vision and gamma cameras.

Obtaining the images of the calibration pattern from the gamma camera is a main problem of this paper. It is because that the gamma camera cannot capture the images of the printed calibration pattern. To solve this problem we use the homography relationship between the gamma and vision cameras. Once several images and of the calibration board from the vision camera are obtained, these images are transformed by and to generate virtual calibration images and from the view of the left and right gamma cameras. Then the transformed images are used as input of Zhang’s algorithm to obtain the stereo calibration parameters of the stereo gamma camera. The whole stereo calibration process is shown in Figure 8. The intrinsic and extrinsic parameters of both left and right radiation detectors are depicted in Tables 2 and 3. Here, , , and represent the rotation matrix, and , , and represent the translation components.

5. Distance Measurement to Radioactive Sources

In order to evaluate the accuracy of the proposed calibration method, we performed distance measurement experiments using the calibrated stereo gamma camera. The main approach of this is to calculate distances to the previously used bright LED spots which are installed at already known distance. Two kinds of experimental setup “P4” and “P8” are used in distance measurement which are shown already in Figures 5 and 7. In the first experimental setup, four LEDs are placed on different positions as shown in Figure 5(a). In the second experimental setup, eight LEDs are placed on different positions as shown in Figure 7(a).

The positions of the light spot are measured using a Bosch GLM 250 VF Professional laser rangefinder. After gamma camera calibration, the intrinsic and extrinsic camera parameters can be obtained and they are used to calculate left and right perspective projection matrices (PPM). Figure 9 shows left and right stereo gamma images where image processing methods such as adaptive thresholding, contour detection, and ellipse fitting are applied to find the center points (() and ()) of each LED light spot of the “P4” setup. Essential and fundamental matrices can also be calculated using the camera parameters according to (4) and the epipolar geometry is applied to identify the corresponding matching points of left and right gamma images. These correspondences are represented with identical colors in Figure 9:

and represent the essential and fundamental matrices of the stereo gamma camera, respectively. is the cross product of the extrinsic camera parameters whereas and represent the intrinsic parameters of the left and right cameras, respectively. The 3D coordinates for each matching point set can be calculated using a linear equation (5):Here represents a 4 × 3 matrix and represents a 4 × 1 matrix as below. The inverse of matrix is calculated using the singular value decomposition (SVD) method:

Figure 10 shows “P4” experimental setup. In this setup, we have done two experiments by changing the position of the four light points as shown in the figure. Figure 10(a) is the same one used in the homography calibration. Figure 10(b) shows four LED lights are placed at the same distance to the gamma camera. Using the two kinds of setup, two distance measurement experiments “P4E1” and “P4E2” are done. As shown in Table 4, the distance measurement is very accurate, where the average percentage error is about 1.13% in “P4E1” and 2.41% in “P4E2” experiments. To each point light in Figure 10(a), measurement error is about 1~4 cm from camera distance of 210~270 cm. If the distance to the LED light increases, the measurement error also increases.

We use more LED light points to analyze the accuracy of distance measurement. Figure 11 shows the second experimental setup “P8” using eight LED lights which are fixed in the different distances. As shown in the figure, “point1,” “point2,” “point6,” and “point7” are on the same plane. Four other light points are fixed at the different distances. With this setup, we have done two experiments “P8E1” and “P8E2” by placing the gamma camera at the different distance from this setup. The ground truth distance to each LED light is shown in Table 5.

Figure 12 shows the left and the right gamma images of the “P8E1” experiments. Stereo correspondences between the left and right images are shown in the same colored ellipse models. The center coordinates of the corresponding ellipse models are used to compute the distance of each LED light point. The calibration parameters of these experiments are also derived by using the same setup as described in Section 3.

Table 5 shows the results of experiments “P8E1” and “P8E2.” The average percentile distance error is about 2% in “P8E1” and 8% in “P8E2” experiment, respectively. The measurement error is a little higher than that of using four LEDs. This may be caused by using more light points in the computation of the homography relationship. We assume that the centers of the vision and gamma camera are the same, so that light points which are not in the plane can be used for homography computation. However, in reality, the alignment error of the centers of the two cameras can yield higher error not only in the homography computation but also in the distance measurement.

As the final experiment, we use a real radioactive source to obtain stereo gamma images and measure the distance to the source. Figure 13(a) shows a facility for general purpose experiments using a single radioactive source. In this facility, a radioactive source of Cs-137 exposes gamma radiation of 100 mSv/h at 2 m distance. As shown in the figure, the gamma source is exposed as a point source. We change the position of the stereo gamma camera module from 150 cm to 400 cm and measure the distance to the source using stereo gamma images captured in each distance.

Figure 13(b) shows gamma images captured from the stereo gamma camera. The position difference in stereo images becomes smaller when the distance to the radioactive source becomes farther due to the parallel stereo geometry. In each distance, the center coordinates of the gamma source are extracted and the real distances to the source are calculated as in Table 6. Average percentage error is about 3.29% and it is reasonably accurate results.

6. Conclusion

In this paper, we proposed a new method to calibrate stereo radiation detection devices using planar homography relationship. Because it is impossible to capture the calibration pattern images from a gamma camera, we use the planar homography relationship between gamma vision cameras to generate virtual calibration images. Bright LED point lights are used as calibration objects and a gamma camera is used as a radiation detection device. A set of calibration pattern images captured from a vision camera are converted into gamma images by applying homography relationships. Zhang’s calibration method is applied to calibrate the gamma camera afterwards. We performed distance measurement experiments using the calibrated gamma camera to check the accuracy of the proposed method. The 3D coordinates of LED point lights and a real gamma radiation source are obtained with high accuracy using stereo gamma images.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported by the Civil Military Technology Cooperation Center and the Korea Atomic Energy Research Institute.