About this Journal Submit a Manuscript Table of Contents
The Scientific World Journal
Volume 2013 (2013), Article ID 139738, 10 pages
http://dx.doi.org/10.1155/2013/139738
Research Article

IMU-Based Online Kinematic Calibration of Robot Manipulator

School of Computer Science & Engineering, South China University of Technology, Guangzhou 510006, China

Received 5 August 2013; Accepted 12 September 2013

Academic Editors: Y. Lu and M. Strojnik

Copyright © 2013 Guanglong Du and Ping Zhang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Robot calibration is a useful diagnostic method for improving the positioning accuracy in robot production and maintenance. An online robot self-calibration method based on inertial measurement unit (IMU) is presented in this paper. The method requires that the IMU is rigidly attached to the robot manipulator, which makes it possible to obtain the orientation of the manipulator with the orientation of the IMU in real time. This paper proposed an efficient approach which incorporates Factored Quaternion Algorithm (FQA) and Kalman Filter (KF) to estimate the orientation of the IMU. Then, an Extended Kalman Filter (EKF) is used to estimate kinematic parameter errors. Using this proposed orientation estimation method will result in improved reliability and accuracy in determining the orientation of the manipulator. Compared with the existing vision-based self-calibration methods, the great advantage of this method is that it does not need the complex steps, such as camera calibration, images capture, and corner detection, which make the robot calibration procedure more autonomous in a dynamic manufacturing environment. Experimental studies on a GOOGOL GRB3016 robot show that this method has better accuracy, convenience, and effectiveness than vision-based methods.

1. Introduction

Because of the manufacturing and assembly tolerance, the actual kinematic parameters of a robot deviate from their nominal values, which are referred to as kinematic errors. The kinematic errors would result in the errors of the robot tool if the nominal kinematics were used to estimate the pose of the robot. With the restriction of cost, the kinematic calibration is an effective way to improve the absolute accuracy of robots. Nowadays, calibration tasks use a lot of measurement techniques like coordinate measuring machines, laser tracking interferometer systems, and inexpensive customized fixtures [1, 2]. These systems are not only very expensive but also not friendly to use or with low working volume. A system which is used in a dynamic environment is expected to perform calibration without any external expensive calibration apparatus and elaborate setups, which means self-calibration.

Self-calibration techniques can be classified into two kinds: redundant sensor approach and motion constraint approach.

To increase the degrees-of-sensing over DOF, the redundant sensors approach includes one or more redundant rotary sensors to the proper passive joints of the manipulator. There is a self-calibration method for parallel mechanisms with a case study on Stewart platform which is proposed by Zhuang in [3]. He used forward and inverse kinematics with six rotary encoders for three objective functions of parameter identification. Khalil and Besnard [4] installed two orthogonally allocated inclinometers to the tool to calibrate the Stewart platform except the redundant sensors which are mentioned above. However, there are some limitations of these methods. One of them is that some kinematic parameters orthogonally are not independent of the error models and the position and/or orientation of the tool on the platform cannot be calibrated.

For the other approach, that is the motion constraint approach, the mobility of the resultant system will be lower than its inherent degrees-of-sensing by fixing one or more passive joints or constraining partial DOF of the manipulator so that the calibration algorithm can be performed [5]. Bennett and Hollerbach [6] lowered the mobility of the tool of a serial manipulator and performed self-calibration using only the inherent joint sensors in the manipulator. And this idea was used and extended to calibrate a robot system with a hand-mounted instrumented stereo camera [7]. However, the position and/or orientation of the tool on the platform cannot be calibrated, and some parameter errors related to the locked passive joints may become unobservable in the calibration algorithm because of the mobility constraints.

To solve these limitations, advances in robot calibration allow the researchers to use a hand-mounted camera to calibrate a robot instead of using measurements from passive joints or imposing mechanical constraints. Compared with those mechanical measuring devices, this camera system costs less and it is easier to use and more accurate. The traditional vision-based methods [815] to calibrate a robot require the precise 3D fixtures measured in a reference coordinate system and the procedure is inconvenient and time consuming and it may not be feasible for some applications. The self-calibration methods [16, 17] assume that the camera is rigidly attached to the robot tool. Closed-loop methods “virtual closed kinematic chain” proposed in [1820], use the joint angle measurements already in the robot and can be considered self-calibrating. A method uses laser to capture robot position data to model the stiffness of the manipulator [21] and predict kinematic parameters [2226]. O’Brian et al. [27] used a magnetic motion to capture robot data to estimate the kinematic parameters. Renaud et al. [28] and Rauf et al. [29] used a vision-based measuring device and a pose measurement device for kinematic calibration, respectively. Santolaria et al. [30] employed a continuous data capture method by using a ball bar gauge and a coupling probe to estimate the kinematic parameters. However, these approaches have a limitation; that is, the calibration is completed off-line. The optimization technique was based on the measured positions of the EE. The parameter error was minimized in the measured positions, but the error increased in very different positions. Moreover, the parameter error increased while the robot withstood different loads. When the robot is used in high-temperature or high-pressure environments, such as deep sea or outer space, the shapes of the robot links are easy to change. Therefore, online calibration is an indispensable method to rectify the kinematic parameters in real time.

In this paper, we propose an original approach of online robot calibration using IMU to measure the robot poses. In our method, an IMU is required to rigidly attach to the robot tool (Figure 1) to measure the robot pose in real time. In order to reduce the effect of the noise and improve the accuracy, we proposed a method combined FQA and KF to estimate the orientation of the IMU. Finally, an EKF is used to estimate differential errors of individual kinematic parameters. Unlike existing vision-based self-calibration methods, the described method does not require special complex steps such as camera calibration and corner detection. Moreover, this method does not require robot to make the motion for capturing the images, which makes our method more efficient.

139738.fig.001
Figure 1: Structure of the system.

The remainder of the paper is organized as follows. Section 2 provides kinematic modeling for the serial robot. In Section 3, a method of pose measurement using IMU is presented. Parameters identification algorithm is proposed in Section 4. In Section 5, an EKF is detailed to estimate the kinematic parameter errors. Finally, the experimental results are shown in Section 6 and we conclude the paper in Section 7.

2. Kinematic Modeling

A robot kinematic model relates the robot joint coordinate to the pose of the robot tool. A robot kinematic model should meet the following rules for the kinematic parameter identification [2123]. (1)Completeness: the robot kinematic model should have enough parameters to define any possible deviation from the nominal values [24].(2)Continuity: any small changes in the structure of the robot must correspond to small changes in kinematic parameters [21].(3)Minimality: the kinematic model must include only a minimal number of parameters [3].

Many researchers have found suitable kinematic models for robot since 1980s, such as Hayati et al. models [2527], Veitschegger and Wu’s model [28], Stone and Sanderson’s S-model [29], and Zhuang et al. model [30]. The standard Denavit-Hartenberg (DH) [31] convention is the most often used to describe the robot kinematics (Figure 2). The error models of DH are not continuous for robots that possess parallel joint axes. To avoid the singularity of DH convention, the DH modeling or Hayati modeling convention were used, respectively. The singularity-free calibration model prevents the use of a single minimal modeling convention which can be used to identify all possible robot parameters.

139738.fig.002
Figure 2: Forward kinematics of an -DOF robot.

The robot tool position and orientation are defined according to the controller conventions. Through the consecutive homogeneous transformations from the base coordinate to the robot tool coordinate (Figure 2), the kinematic equation can be defined as where is the translation matrix from coordinate to coordinate, is the number of joints (or joints coordinates), is the parameter vector for the robot, and is the link parameter vector, which includes the joint errors where is the nominal vector for the joint and is the link parameter error vector for the joint . The exact kinematic equation is Taking the joint variables into consideration, thus, where is a vector of joint variables. Thus,   is a function of and .

3. Pose Measurement Using IMU

3.1. Factored Quaternion Algorithm (FQA)

The FQA presented in [32], which is based on Earth gravity and magnetic field measurements, is for estimating the orientation of a rigid body. This algorithm is only applied for the static or slow-moving rigid body. In order to be applicable to situations in which relatively large linear accelerations occur, we use KF fusion algorithm together with angular rate information to estimate the orientation of (slow-moving or fast-moving) dynamic body in the next section.

In our application, a sensor module, which is a strap down inertial measurement unit (IMU) is attached to the robot tool whose orientation (roll, pitch, and yaw) is to be determined. Figure 3 shows an IMU of Xsens and its prototype board. The IMU sensor consists of one three-axis accelerometer, two two-axis gyroscopes, and one three-axis magnetometer.

fig3
Figure 3: IMU sensor (a) and the prototype board (b).

We define three frames: body frame , sensor frame , and Earth-fixed frame . The sensor frame corresponds to the axes of three orthogonally mounted accelerometers/magnetometers. Because the sensor is attached to the robot tool rigidly, the body frame is assumed to coincide with the sensor frame . The Earth-fixed frame is defined to follow the North-East-down (NED) convention, where points North, points East, and points down. The IMU can measure the orientation (roll, pitch, and yaw) of itself. Define that the rotation about the -axis represents roll, the rotation about -axis represents pitch, and the rotation about -axis represents yaw. According to Euler’s theorem [33] on finite rotations, the conversion from Euler angles to quaternions isand the four Euler parameters are constrained as [34] where is the scalar part and () is the vector part. So the direction cosine matrix from the sensor frame to the Earth-fixed frame can be represented as follows:

3.2. Quaternion KF

Because both gyroscopes and magnetometer have white noise and random walk, we use Kalman Filter to estimate the state of IMU from a set of noisy and incomplete measurements [35]. The Kalman Filter is a recursive stochastic technique and it estimates the state at time from the state at time . The state-transition equation at time is where is the state vector at time , is an state-transition matrix, is an system input matrix, is a vector with deterministic input at time , is an process noise vector at time , is an measurements vector at time , is an observation matrix, and is an measurement noise vector. In this paper, and . The differential equation of the quaternion with respect to time is where , , and are the angular velocity components of IMU in -, -, and -axis. Since the state includes the quaternion states and the angular velocities, has the following form: where , , , , , , and are the quaternion states and the angular velocities at time . From (10), the state-transition matrix is where is the sampling time. Let because there is no control inputs. We use angular velocities to estimate the quaternion states, so the process noise vector is where , , and are the process noises of the angular velocity. Because we use calibrated gyroscopes to measure the angular velocities, the observation matrix is

In order to satisfy (7), the determined quaternion at time should be normalized by

4. Parameter Identification

Kinematic identification is the process that identifies the kinematic model parameters of a robot manipulator by a given set of robot tool pose measurements and the corresponding joint position readings. The objective of a kinematic identification algorithm is to minimize the difference between the computed and the measured poses [15].

Assuming that the number of measured pose is , it can be stated that where () is the vector of joint variables for the measure pose.

All matrices or vectors in bold are functions of . The objective of the kinematic identification is the computation for the parameter vector , which is to minimize the discrepancy between the computed and the measured poses: is the function of pose of and is the measured function of joint variables .

For each measurement pose , it concludes orientation measurement and position measurement , and

If the measurement system can provide orientation measurement and position measurement, each pose can formulate six measurement equations. If only orientation measurement can be provided by the measurement system, each pose measurement can just formulate three measurement equations. In this paper, only orientation obtained from IMU is used to calibration the kinematic parameters. From(17), where is the discrepancy function of the orientation components of . Introducing the Jacobian matrix, and then when using Equation (20) can be rewritten:

5. Estimating Errors Using Extended Kalman Filter

Initially, the orientations of the tool are measured from the IMU. Since uncertainty exists in the measurement, Extended Kalman Filter (EKF) is used as an optimization algorithm and the Jacobian matrices are used to estimate the kinematic errors of DH parameters by the measured orientation values [5].

Since there are four parameters for revolute joints and four parameters for the transformation from the IMU to the tool, the number of total parameters to be considered is . So the predicted state is of the DH parameters in the prediction step of the EKF. The covariance matrix of the predicted state is where is the covariance matrix of the system noise at the iteration.

In the observation step of the EKF, Jacobian matrix (20), measurement residual , and residual covariance are calculated as follows: where and are the measured orientation value and the covariance matrix of measurement noise at the th iteration. means a prior estimate, and means a posteriori estimate.

In the update step, the state covariance matrix is updated by an optimal Kalman gain : where is the identity matrix. The norm values of the state vector are calculated for every iteration once the updating procedure is completed. Note that if and are set to zero, then EKF simply reduces to the Newton-Raphson method.

6. Experimental Results

6.1. Experiments Environment

To verify the above method, a GOOGOL GRB3016 robot model was used in this experiment. In the experiment, the robot could self-calibrate online in the working status. There were four steps in the self-calibration procedure.(1)Data collection: the orientation of the IMU and the corresponding joint angles were captured with the robot tool moving in different pose.(2)Manipulator orientation estimation: the orientations of the manipulator were estimated via the KFs from the obtained data.(3)Kinematic parameters identification: the manipulator kinematic parameters were identified from the estimated orientations and the joint data.(4)Calibration accuracy assessment: 3D pose errors were used to verify the calibration accuracy by inserting a peg into the holes (Figure 5).

Table 1 lists the nominal robot link parameters of the robot, which were chosen as the initial conditions for the above kinematic identification algorithm, and Figure 4 shows the skeleton of the GOOGOL GRB3016 robot with all coordinate frames and geometric features. As expected, the increase in the noise intensity will lead to the increase in the calibration errors.

tab1
Table 1: The nominal link parameters in DH model for the GOOGOL GRB3016 robot.
139738.fig.004
Figure 4: Skeleton of the GOOGOL GRB3016 robot with coordinate frames in the zero position.
139738.fig.005
Figure 5: Steel plate and errors measurement system.

From Table 1 note that a GOOGOL GRB3016 robot with 6 DOF needs 24 geometric parameters to be modeled. From (18), each 3D robot orientation provides 3 model equations. So a unique computation of the 24 parameters needs 8 pose measurements at least. And more pose measurements will decrease the calibration errors. But limited by the measurement accuracy (affected by the noise), the calibration errors intend to be stable as the pose measurements increase.

In order to validate the proposed IMU-based robot calibration, we have performed a number of peg-into-hole experiments. In our experiments, two calibration methods were used to carry out the experiments of peg-into-hole. Our method was initially compared with a standard vision-based robot calibration of the Meng and Zhuang [16]. There were 16 holes in the steel plate (Figure 5) and 16 tests of peg-into-hole were carried out in each experiment. The peg was the cylinder with 7.5 mm in radius and 150 mm in length. The radius of the hole was 8 mm. The size of the steel plate was 300 mm 500 mm. In the step of our method, an IMU (Xsens MTi-100 IMU) was rigidly attached to the robot tool flange to measure the pose of the robot tool. The static accuracy of roll and pitch was 0.02 deg and that of yaw was 0.05 deg. The dynamic accuracy of roll and pitch was 0.05 deg and that of yaw was 0.1 deg. The noise density of gyroscopes was   and that of the magnetometer was   . The IMU measurements were received at 100 Hz. Since the robot will be stopped after it executes a command, in order to improve the accuracy, the system collects the static IMU data after the robot stops. The coordinates of the center of the hole with respect to the base coordinate system were known. In the step of method [16], a camera, with 1280(H) 960(V) picture elements and 22 frame/s frequency, was mounted on the robot tool to capture the RGB images of the chessboard. A chessboard pattern, the position of which was unknown in the reference frame, was placed on the platform. The distances, measured by a caliper, between two adjacent corners of the square were known. Note that the robot has different kinematic parameter errors in the different position. In order to improve the accuracy, the system made a robot calibration before inserting the peg into each hole.

To evaluate the calibration accuracy, 3D position errors between the peg and the hole were proposed. A calibrated camera with 1280(H) 960(V) picture elements was used to measure the 3D position errors (Figure 5). In the initiation of the system, the camera measured the coordinates of the center of the hole with respect to the camera coordinate system. After the peg was inserted into a hole, the camera can measure the depth and the direction of insertion by detecting the edge of the peg. And then the system could calculate the center of the peg with respect to the camera coordinate system. Let and be the centers of the peg and the hole, respectively (Figure 6). Then the 3D position error can be written: The mean absolute error in position for peg-into-hole tests was where is the 3D position error for the th hole.

fig6
Figure 6: Definition of 3D errors.
6.2. Result Analysis

Following the initialization step, the system performs a series of tests for peg-into-hole. The KF processes the IMU measurements and concurrently estimates the state vector. For EKF, the number of maximum iteration was set as 2000. and were set as . When the state vector was less than , the iteration terminated. The proposed algorithm was evaluated by estimating the orientation from the IMU and the theoretical orientation obtained by using (3). Figure 7 shows the estimated DH parameter error values of the robot tool. As shown in Figure 7, the EKF algorithm quickly converged to the stable error value from about the fifteenth iteration.

fig7
Figure 7: Estimated - parameter errors with EKF.

Table 2 shows the estimated parameter errors when the calibration sets were used in EKF fro 200 iterations. Since the EKF algorithm quickly converged to the stable error value from about the fifteenth iteration, The iteration should be set as 15 so that the calibration could be carried out in real time.

tab2
Table 2: Estimated parameter errors of 6 DOF robot.

In the experiments of peg-into-hole, the iteration was set as 15 to estimate the parameter errors. The state-estimate errors with 15 pose measurements are shown in Table 3. As evident from Table 3, even with the noise error for the IMU measurements, the algorithm is still able to attain very accurate estimates of the calibration parameters. Table 3 shows the accuracy result of method [16] too. By comparing the results of Table 3, the calibration parameters were more accurate in our method.

tab3
Table 3: The 3D errors with 15 pose measurements.

In our method, a unique computation of the 24 kinematic parameters needs 8 pose measurements. In method [16], it needs 4 pose measurements at least. We compare two methods from 10 pose measurements to 15 pose measurements. The results presented in Figure 8 show that with more pose measurements, the parameter error decreases gradually. The mean absolute estimation errors from 10 pose measurements to 15 pose measurements in method [16] were 4.5 mm, 2.3 mm, 1.41 mm, 1.3 mm, 0.61 mm, and 0.52 mm with standard deviations (SDs) of 0.4 mm, 0.38 mm, 0.35 mm, 0.28 mm, 0.09 mm, and 0.09 mm. Compared with method [16], the mean absolute errors of our method drop about 0.51 mm, 0.32 mm, 0.41 mm, 0.23 mm, 0.13 mm, and 0.22 mm. The estimation errors trend slows down after two times of the minimum number of poses measurements.

139738.fig.008
Figure 8: Mean absolute errors in different number of measurements.

Compared with method [16], the great advantage of our method is that the system does not need to make a more motion to take the photo. After the robot executed a command, the robot would stop and the system concurrently obtained the static measurement data from the IMU. Figure 9(a) shows the execution time of two methods with 15 pose measurements. In method [16], the average time of taking a photo was 3 s. In our method, the station time of robot was about 1 s. The execution time of peg-into-hole is about 8 s. The time of parameter identification was 0.8 s. Figure 9(b) shows the comparison of execution time with different pose measurements. The time in method [16] is more than two times that in our method.

fig9
Figure 9: The comparison in execution time.

7. Conclusions

An IMU-based online autonomous calibration for serial robot has been proposed in this paper. In this approach, the IMU is rigidly attached to the robot tool to estimate the robot pose automatically during the working time. An efficient approach which incorporates Factored Quaternion Algorithm (FQA), Kalman Filter (KF), and Extended Kalman Filter (EKF) to estimate the orientation of the IMU is presented in this paper. After the robot poses are estimated, the kinematics identification can be carried out. Finally, the robot kinematic parameters can be corrected from the identification results in real time. The whole procedure of the robot calibration is automatic and without any manual intervention. The results of the experiments show the good accuracy, convenience, and effectiveness of the presented approach.

Compared with the existing expensive and complex approach, the method this paper proposed is easier to use and setup. Compared to the existing vision-based self-calibration method, the proposed method can conduct the calibration more accurately and with less execution time. In the future work, we will research the approach which can accurately estimate the robot pose without stopping the robot. With the dynamic pose measurements, the robot calibration will be more efficient.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  1. M. Vincze, J. P. Prenninger, and H. Gander, “Laser tracking system to measure position and orientation of robot end effectors under motion,” International Journal of Robotics Research, vol. 13, no. 4, pp. 305–314, 1994. View at Scopus
  2. H. Zhuang, J. Yan, and O. Masory, “Calibration of Stewart platforms and other parallel manipulators by minimizing inverse kinematic residuals,” Journal of Robotic Systems, vol. 15, no. 7, pp. 395–405, 1998. View at Scopus
  3. H. Zhuang, “Self-calibration of parallel mechanisms with a case study on Stewart platforms,” IEEE Transactions on Robotics and Automation, vol. 13, no. 3, pp. 387–397, 1997. View at Publisher · View at Google Scholar · View at Scopus
  4. W. Khalil and S. Besnard, “Self calibration of Stewart-Gough parallel robots without extra sensors,” IEEE Transactions on Robotics and Automation, vol. 15, no. 6, pp. 1116–1121, 1999. View at Publisher · View at Google Scholar · View at Scopus
  5. I.-W. Park, B.-J. Lee, S.-H. Cho, Y.-D. Hong, and J.-H. Kim, “Laser-based kinematic calibration of robot manipulator using differential kinematics,” IEEE/ASME Transactions on Mechatronics, vol. 17, no. 6, pp. 1059–1067, 2012. View at Publisher · View at Google Scholar · View at Scopus
  6. D. J. Bennett and J. M. Hollerbach, “Autonomous calibration of single-loop closed kinematic chains formed by manipulators with passive endpoint constraints,” IEEE Transactions on Robotics and Automation, vol. 7, no. 5, pp. 597–606, 1991. View at Publisher · View at Google Scholar · View at Scopus
  7. D. J. Bennett, D. Geiger, and J. M. Hollerbach, “Autonomous robot calibration for hand-eye coordination,” International Journal of Robotics Research, vol. 10, no. 5, pp. 550–559, 1991. View at Scopus
  8. S. K. Mustafa, G. Yang, S. H. Yeo, W. Lin, and I.-M. Chen, “Self-calibration of a biologically inspired 7 DOF cable-driven robotic arm,” IEEE/ASME Transactions on Mechatronics, vol. 13, no. 1, pp. 66–75, 2008. View at Publisher · View at Google Scholar · View at Scopus
  9. H. Zhuang, Z. S. Roth, and K. Wang, “Robot calibration by mobile camera systems,” Journal of Robotic Systems, vol. 11, no. 3, pp. 155–167, 1994. View at Scopus
  10. P. Renaud, N. Andreff, F. Marquet, and P. Martinet, “Vision-based kinematic calibration of a H4 parallel mechanism,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA '03), pp. 1191–1196, September 2003. View at Scopus
  11. D. Daney, N. Andreff, and Y. Papegay, Interval Method for Calibration of Parallel Robots: A Vision-Based Experimentation, Computational Kinematics, Cassino, Italy, 2005.
  12. P. Renaud, N. Andreff, J.-M. Lavest, and M. Dhome, “Simplifying the kinematic calibration of parallel mechanisms using vision-based metrology,” IEEE Transactions on Robotics, vol. 22, no. 1, pp. 12–22, 2006. View at Publisher · View at Google Scholar · View at Scopus
  13. P. Renaud, N. Andreff, M. Philippe, and G. Gogu, “Kinematic calibration of parallel mechanisms: a novel approach using legs observation,” IEEE Transactions on Robotics, vol. 21, no. 4, pp. 529–538, 2005. View at Publisher · View at Google Scholar · View at Scopus
  14. N. Andreff and P. Martinet, “Unifying kinematic modeling, identification, and control of a Gough-Stewart parallel robot into a vision-based framework,” IEEE Transactions on Robotics, vol. 22, no. 6, pp. 1077–1086, 2006. View at Publisher · View at Google Scholar · View at Scopus
  15. N. Andreff, P. Renaud, P. Martinet, and F. Pierrot, “Vision-based kinematic calibration of an H4 parallel mechanism: practical accuracies,” Industrial Robot, vol. 31, no. 3, pp. 273–283, 2004. View at Publisher · View at Google Scholar · View at Scopus
  16. Y. Meng and H. Zhuang, “Autonomous robot calibration using vision technology,” Robotics and Computer-Integrated Manufacturing, vol. 23, no. 4, pp. 436–446, 2007. View at Publisher · View at Google Scholar · View at Scopus
  17. J. M. S. T. Motta, G. C. De Carvalho, and R. S. McMaster, “Robot calibration using a 3D vision-based measurement system with a single camera,” Robotics and Computer-Integrated Manufacturing, vol. 17, no. 6, pp. 487–497, 2001. View at Publisher · View at Google Scholar · View at Scopus
  18. Y. Zhu, C. Hu, J. Hu, and K. Yang, “Accuracy and simplicity oriented self-calibration approach for two-dimensional precision stages,” IEEE Transactions on Industrial Electronics, vol. 60, no. 6, pp. 2264–2272, 2012.
  19. C. S. Gatla, R. Lumia, J. Wood, and G. Starr, “An automated method to calibrate industrial robots using a virtual closed kinematic chain,” IEEE Transactions on Robotics, vol. 23, no. 6, pp. 1105–1116, 2007. View at Publisher · View at Google Scholar · View at Scopus
  20. F. Aghili, “Adaptive control of manipulators forming closed kinematic chain with inaccurate kinematic model,” IEEE/ASME Transactions on Mechatronics, vol. 18, no. 5, pp. 1544–1554, 2012.
  21. G. Alici and B. Shirinzadeh, “Enhanced stiffness modeling, identification and characterization for robot manipulators,” IEEE Transactions on Robotics, vol. 21, no. 4, pp. 554–564, 2005. View at Publisher · View at Google Scholar · View at Scopus
  22. W. S. Newman, C. E. Birkhimer, R. J. Horning, and A. T. Wilkey, “Calibration of a Motoman P8 robot based on laser tracking,” in Proceedings of the IEEE International Conference on Robotics and Automation (ICRA '00), pp. 3597–3602, San Francisco, Calif, USA, April 2000. View at Scopus
  23. A. Omodei, G. Legnani, and R. Adamini, “Three methodologies for thecalibration of industrial manipulators: experimental results on a SCARArobot,” Journal of Robotic Systems, vol. 17, no. 6, pp. 291–307, 2000.
  24. A. Omodei, G. Legnani, and R. Adamini, “Calibration of a measuring robot: experimental results on a 5 DOF structure,” Journal of Robotic Systems, vol. 18, no. 5, pp. 237–250, 2001. View at Publisher · View at Google Scholar · View at Scopus
  25. G. Alici and B. Shirinzadeh, “A systematic technique to estimate positioning errors for robot accuracy improvement using laser interferometry based sensing,” Mechanism and Machine Theory, vol. 40, no. 8, pp. 879–906, 2005. View at Publisher · View at Google Scholar · View at Scopus
  26. G. Alici, R. Jagielski, Y. Şekercioǧlu, and B. Shirinzadeh, “Prediction of geometric errors of robot manipulators with Particle Swarm Optimisation method,” Robotics and Autonomous Systems, vol. 54, no. 12, pp. 956–966, 2006. View at Publisher · View at Google Scholar · View at Scopus
  27. J. O’Brian, R. Bodenheimer, G. Brostow, and J. Hodgins, “Automaticjoint parameter estimation from magnetic motion capture data,” in Proceedings of the Graph Interface, pp. 53–60, May 2000.
  28. P. Renaud, N. Andreff, J.-M. Lavest, and M. Dhome, “Simplifying the kinematic calibration of parallel mechanisms using vision-based metrology,” IEEE Transactions on Robotics, vol. 22, no. 1, pp. 12–22, 2006. View at Publisher · View at Google Scholar · View at Scopus
  29. A. Rauf, A. Pervez, and J. Ryu, “Experimental results on kinematic calibration of parallel manipulators using a partial pose measurement device,” IEEE Transactions on Robotics, vol. 22, no. 2, pp. 379–384, 2006. View at Publisher · View at Google Scholar · View at Scopus
  30. J. Santolaria, J.-J. Aguilar, J.-A. Yagüe, and J. Pastor, “Kinematic parameter estimation technique for calibration and repeatability improvement of articulated arm coordinate measuring machines,” Precision Engineering, vol. 32, no. 4, pp. 251–268, 2008. View at Publisher · View at Google Scholar · View at Scopus
  31. R. He, Y. Zhao, S. Yang, and S. Yang, “Kinematic-parameter identification for serial-robot calibration based on POE formula,” IEEE Transactions on Robotics, vol. 26, no. 3, pp. 411–423, 2010. View at Publisher · View at Google Scholar · View at Scopus
  32. G. Gatti and G. Danieli, “A practical approach to compensate for geometric errors in measuring arms: application to a six-degree-of-freedom kinematic structure,” Measurement Science and Technology, vol. 19, no. 1, Article ID 015107, 2008. View at Publisher · View at Google Scholar · View at Scopus
  33. K. Schröer, S. L. Albright, and M. Grethlein, “Complete, minimal and model-continuous kinematic models for robot calibration,” Robotics and Computer-Integrated Manufacturing, vol. 13, no. 1, pp. 73–85, 1997. View at Scopus
  34. H. Zhuang and Z. Roth, Camera -Aided Robot Calibration, CRC, Boca Raton, Fla, USA, 1996.
  35. S. A. Hayati, “Robot arm geometric link parameter estimation,” in Proceedings of the 22th IEEE Conference on Decision and Control, vol. 22, pp. 1477–1483, December 1983.