Literature DB >> 31623249

Robust and Accurate Hand-Eye Calibration Method Based on Schur Matric Decomposition.

Jinbo Liu1, Jinshui Wu2, And Xin Li3.   

Abstract

To improve the accuracy and robustness of hand-eye calibration, a hand-eye calibration method based on Schur matric decomposition is proposed in this paper. The accuracy of these methods strongly depends on the quality of observation data. Therefore, preprocessing observation data is essential. As with traditional two-step hand-eye calibration methods, we first solve the rotation parameters and then the translation vector can be immediately determined. A general solution was obtained from one observation through Schur matric decomposition and then the degrees of freedom were decreased from three to two. Observation data preprocessing is one of the basic unresolved problems with hand-eye calibration methods. A discriminant equation to delete outliers was deduced based on Schur matric decomposition. Finally, the basic problem of observation data preprocessing was solved using outlier detection, which significantly improved robustness. The proposed method was validated by both simulations and experiments. The results show that the prediction error of rotation and translation was 0.06 arcmin and 1.01 mm respectively, and the proposed method performed much better in outlier detection. A minimal configuration for the unique solution was proven from a new perspective.

Entities:  

Keywords:  Schur matric decomposition; hand–eye calibration; observation data preprocessing; outlier detection; robotics

Mesh:

Year:  2019        PMID: 31623249      PMCID: PMC6832585          DOI: 10.3390/s19204490

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

The combination of vision sensors and robots is a milestone in robotic intelligence, increasing the extent and efficacy of robot applications [1,2,3,4,5]. Hand–eye calibration is an important technique for bridging the transformation between a robot gripper and a robot vision sensor [6]. Its application is mainly reflected in the robot’s hand–eye coordination, guiding the robot gripper to accurately target and reach into a specified location using the machine vision system. From height work to surgery, the more sophisticated the operation, the better robot hand–eye coordination required. Many researchers have studied hand–eye calibration, and all current methods can be divided into two categories: linear methods and iterative methods. Linear methods are efficient and suitable for online hand–eye calibration. Shiu and Ahmad first introduced the dynamic equation AX = XB into hand–eye calibration and provided minimal configuration for a unique solution [6]. Tsai and Lens proposed a high-efficiency linear method for the equation AX = XB [7]. Chou and Kamel expressed rotation matrices using quaternions and obtained an analytical solution using Singular Value Decomposition (SVD) [8]. Lu and Chou used an eight-dimension vector to express rotation and translation and obtained a least squares solution [9]. Chen analyzed the relationship between screw movement and hand–eye calibration, and then proved that the movement of the robot gripper and vision sensor must satisfy certain geometric constraints [10]. Daniilidis solved rotation and translation simultaneously by means of a dual quaternion [11]. Park introduced canonical coordinates into the hand–eye calibration equation, which simplified the parameters [12]. Shah constructed a closed-form solution and derived the minimal configuration of the unique solution based on Kronecker product [13]. Compared with Daniilidis [11], Shah’s method was more reliable and accurate. Iterative methods are mainly used to improve the accuracy and robustness. Other authors [14,15] took the F norm of the rotation error and translation error as the cost function, and then optimized it using nonlinear methods. Horaud expressed rotation matrices using quaternions and simultaneously optimized the transformation between the robot-world and hand and eye [16]. Strobl and Hirzinger proposed a new adaptive error model that helped improve the solution to AX = XB and AX = ZB [17]. Ruland proposed a self-calibration method that took projection error as its cost function and optimized it using branch-and-bound [18]. The accuracies of the above methods strongly depend on the quality of the observation data. Therefore, preprocessing observation data is essential. Observation data preprocessing is rarely reported. Schmidt et al. [19] proposed a preprocessing method based on vector quantization, which improved the quality of observation data to a certain extent but could not identity outliers. The complexity increased from O(N) to O(N4), which considerably decreased the method’s efficiency.

2. Description of Hand–Eye Calibration Problem

Figure 1 describes the hand–eye calibration problem. The symbols are notated as follows: G is the robot gripper coordinate system, it is fixed on the robot gripper and moves together with it, C is the camera coordinate system fixed on the camera that moves together with it and the origin point is coincident with the camera’s optical center. The Z-axis is parallel to the optical axis, and the X and Y axes are parallel to the X and Y axes of the image coordinate system. CW is the world coordinate system and RW is the robot coordinate system that is fixed on the robot and moves together with it. When the robot gripper moves, its controlling device can identify the gripper’s pose in RW.
Figure 1

Description of the hand–eye calibration problem.

A is the homogenous transformation matric from G to RW, obtained from the robot controlling device: B is the homogenous transformation matric from CW to C, obtained using camera pose estimation methods: A is the homogenous transformation matric from G to G: B is the homogenous transformation matric from C to C: and X is the homogenous transformation matric from C to G: i and j represent the ith and jth state of the robot gripper and camera respectively, ranging from 0 to N. N is the number of movements. Since the robot gripper and camera are fixed, X is constant. The hand–eye calibration equation can be represented by notations: Two equations can be obtained based on the partition matric: Equation (7) shows that R is independent, but the accuracy of t is related to R.

3. Hand–Eye Calibration Method

3.1. Schur Matric Decomposition

A given matric can be simplified to a normalized form via similarity transformation. Considering numerical stability, the similarity transformation of a unitary matric is the most attractive. Schur matric decomposition can be simply described as: If A ∈ C×, then a unitary matric that satisfies UHAU = T = D + N exists, where D is a diagonal matric and N is a strictly upper triangular matric, implicating ∀I ≥ j n = 0. For a real matric A, U is restricted to an orthogonal matric: UTAU = T. T has the following form: T is a 1 × 1 or 2 × 2 matric consisting of complex conjugate eigenvalues. If is similar to and eigenvalues of and are the same, the matric T related to and are the same.

3.2. Hand–Eye Calibration Principle

A0, B0 is notated as the initial state of the robot gripper and camera. is a series of homogenous transformation matrices related to their initial states. Without the loss of generality, e.g., i = 1, only consider the equation related to the rotation in Equation (7): From Theorem 1, proved in the Appendix A, the general solution can be written as: And R only depends on c and d. For arbitrary i = 1, 2, …, N – 1, N: Substitute Equation (10) into Equation (12): where: Collate Equation (13) into equations only related to s = (c d)T. C is a matric generated by the coefficients of c and d. D is a matric generated by the constant term. Then, the final linear equation system can be constructed: where, This is a least squares problem with constraints: where, Notate the cost function as: From and : Notate and substitute it into previous equations: K is a symmetrical matric, so s = 1 is the same as y = 1: Because y = F: This is a symmetrical second eigenvalue problem [20]. Solve the least squares solution of the Langrage multiplier through methods previously published [21,22]. The least square solution of s is: Under the condition , the least squares solution of R is: An exists for each i = 1, 2, …, N – 1, N. To weaken the effect of noise, fuse the matrices based on the string distance of matrices. First, calculate the singular decomposition of the sum of , i = 1, 2, …, N – 1, N: Then: To solve for t, for the ith movement, the translation satisfies the following equation: Substitute Equation (29) into Equation (30): Then, a large linear equation system can be obtained: where, This problem can be solved using the least squares method [20].

3.3. Outlier Detection

In practice, matrices A and B contain an observation error, notated as and , respectively. B is more sensitive to image noises. A poor environment may lead to a large observation error and, in this case, the global optimization solution has no significance. This is a basic problem that considerably decreases the robustness of hand–eye calibration and has not been well solved. The form of Y is: and must satisfy Equation (13). For arbitrary c and d, Equation (36) is satisfied: which can be used to discriminate the quality of the observation data: if greater than a specific threshold ε, then the observation data are outliers and should be deleted. The threshold ε is an empirical value. Through setting its value, the observation data can be filtered. The lower the threshold ε, the higher the quality of the observation data. In simulations and experiments, ε was set to 0.01. In summary, the flowchart of the proposed method is described in Figure 2.
Figure 2

Flowchart of the proposed method.

3.4. Unique Solution Conditions

Assume the rotation matrices of two movements are A1, A2, B1, and B2, and X is known. From Theorem 1 (Appendix A), the general solution of A1X = XB1 is: where, Y is a matric only related to c and d. Substitute Equation (37) into the equation built by two movements: Substitute them into Equation (13) to obtain: And For Equation (39): Equation (41) is an identical equation. If rotation axes of two movements are not parallel, P2 and Q2 are independent: From Theorem 2 proved in the Appendix A, if the rotation axes of N movements of the robot gripper are parallel, there will be multiple solutions to the hand–eye calibration. Therefore, the minimal configuration of the unique solution is that the robot gripper and camera move at least twice, and the rotation axes cannot be parallel.

4. Results

4.1. Simulations

We designed simulations to test the performance of different hand–eye calibration methods. The hand–eye calibration equation can be written as: where, A and B are the movement of the robot gripper and camera from time i to time j, respectively. A and B were simulated as the observation data. X is simulated as the transformation from the camera to the robot gripper. A, B and X consist of rotation matrices and translation vectors. The rotation matric can be generated using three Euler angles. The simulations included three parts: analysis of noise sensitivity, relationship between the number of movements and accuracy, and outlier detection ability. All the simulations were performed using MATLAB. In addition to the proposed method, we selected another five popular methods for comparisons [7,11,12,13,23]. For the ith simulation, and are the ideal transformation from the camera to the robot gripper and and are the measured transformations. The error matric can be calculated as: where, and are the rotation axis and rotation angle of , respectively. The errors of rotation and translation are defined as: where, n is the number of simulations.

4.1.1. Analysis of Noise Sensitivity

Gaussian rotation noise (μ = 0, σ = 0°–5°) and translation noise (μ = 0, σ = 0–5 mm) were added into A and B (i = 1, 2, …, 9, 10). We ran 100 simulations at each noise level. The results were shown in Figure 3, in which ‘Rot.’ represents ‘Rotation’ and ‘Trans.’ represents ‘Translation’. Except for the dual quaternion method, translation perturbation had no effect on the rotation solution, because only the dual quaternion method solves rotation and translation simultaneously, whereas other methods solve rotation and translation by steps.
Figure 3

The relationship between calibration accuracy and observation errors: (a) Rotation errors in observations and calibration errors of R. (b) Rotation errors in observations and calibration errors of t. (c) Translation errors in observations and calibration errors of R. (d) Translation errors in observation and calibration errors of t. Each point on the figure is the Root Mean Square (RMS) of 100 simulations.

4.1.2. Relationship between Number of Movements and Accuracy

The simulation conditions included σ = 0.2°, σ = 2 mm, and the number of movements varied from 3 to 15. We ran 100 simulations at each number of movements. Figure 4a,b indicates that the accuracy of hand–eye calibration improves with the increase in the number of movements. When the number of movements increases from three to eight, the accuracy of hand–eye calibration improves considerably. Figure 4c,d demonstrates that the other five methods are more robust, except for the dual quaternion method being unstable.
Figure 4

The relationship between calibration accuracy and the number of movements: (a) The number of movements and the calibration errors of R. (b) The number of movements and the calibration errors of t. (c) The standard deviations of the calibration errors of R. (d) The standard deviations of the calibration errors of t. Each point on the figure is the RMS of 100 simulations.

4.1.3. Outlier Detection

The simulation conditions were σ = 0.2°, σ = 2 mm, and ε = 0.01. The robot gripper moved 10 times, in which large noise was added into n (n = 1, 2, 3, 4, 5, 6) movements randomly and these observations were regarded as outliers. We ran 100 simulations at each number of outliers. Figure 5a,b shows the relationship between calibration errors of R and t and the number of outliers, respectively. Figure 5c,d depicts the performance of the proposed method. The results indicate that the proposed method can detect outliers effectively and performs much better than the other five methods.
Figure 5

The relationship between calibration accuracy and the number of outliers: (a) Calibration errors of R and the number of outliers. (b) Calibration errors of t and the number of outliers. (c) Partial enlargers of (a). (d) Partial enlargers of (b). Each point on the figure is the RMS of 100 simulations.

4.2. Experiments

Determining poses of the robot gripper with high precision is costly, but movements of the robot gripper can be measured precisely. Thus, most researchers adopt the following program to validate hand–eye calibration methods: the camera moves N + n times, where the preview N times are called the calibration link and the last n times are called the verification link. The calibration link is used to solve the transformation between the robot gripper and the camera. The verification link is used to verify method accuracy by comparing its predicted movements with its true movements [3]. The predicted movements of the robot gripper can be solved from the camera’s movements using Equation (47). The true movements of the robot gripper can be obtained from its controlling device. A robot arm was fixed with a camera, as shown in Figure 6a.
Figure 6

(a) Robot arm, gripper and its controlling device. (b) Camera and feature points.

For the calibration link: (1) Fix 9 feature points on the platform as shown in Figure 6b. The three-dimensional (3D) coordinates of feature points can be measured by Leica Total Station. All the feature points’ coordinates remain unchanged during the experiment. (2) At time 0, capture an image of the feature points on the platform. Calculate the camera’s pose B0 through Perspective-n-Points (PnP) methods. The robot gripper’s pose A0 can be determined from its controlling device. (3) At time i, move the robot gripper and camera. (4) Capture an image of the feature points on the platform. Calculate the camera’s pose B through PnP methods. The robot gripper’s pose A can be determined from its controlling device. (5) Repeat step (3)–(4) N times and (A0, B0) (i = 1, …, N–1, N) can be obtained. (6) The transformation X from the camera to the robot gripper can be calibrated using all six hand–eye calibration methods. For the verification link: (7) Repeat step (3)–(4) n times and (A0, B0) (i = N+1, …, N+n–1, N+n) can be obtained. (8) The predicted movement of the robot gripper can be calculated through Equation (47). The true movement of the robot gripper A0 can be obtained from its controlling device. (9) Comparing with A0, the error matric can be calculated using Equation (48): and are the corresponding rotation axis and rotation angle of , respectively: The rotation and translation errors are defined as: The rotation error is in arcmin and the translation error is in mm. In the experiment, N = 2–9 and n = 200. The results are shown in Table 1. The experiment results indicate that the prediction error decreased with the increase in the number of movements and when the robot gripper moved 9 times, the proposed method’s prediction accuracy of rotation exceeded 6 arcsec, which is much higher than the calibration accuracy in the simulations. The reason is explained in the following.
Table 1

Prediction error: t in mm and θ in arcmin.

N ProposedTsaiInriaNavyDual QuaternionShah
θerror terror θerror terror θerror terror θerror terror θerror terror θerror terror
210.145.4910.147.0610.146.2310.175.2510.218.7010.145.63
310.104.6310.106.2110.146.2010.145.1010.147.0810.104.71
410.074.0610.105.7710.104.7410.104.9710.146.1810.104.16
59.833.9410.074.1510.073.7910.074.6210.104.179.864.04
60.962.460.963.671.303.612.163.543.813.981.342.60
70.441.570.513.510.513.600.511.871.783.640.721.75
80.371.150.372.760.412.510.441.770.442.270.371.20
90.061.010.272.470.342.270.411.190.411.820.201.05
Expand Equation (47) using a partition matric: The prediction error consists of hand–eye calibration error and camera pose estimation error. Hand–eye calibration error is notated as ∆R. Then, the prediction error of Equation (51) can be written as: Equation (52) can weaken the effect of the hand–eye calibration error. This conclusion also applies to the prediction error of translation. Thus, the prediction error in the experiment was much lower than the hand–eye calibration error in the simulations.

5. Conclusions

A hand–eye calibration method with high accuracy and robustness was proposed in this paper. Using this method, the basic problem of observation data preprocessing is solved by outlier detection, which significantly improves robustness. However, two aspects remain to be studied. To improve the method’s efficiency, we used the least squares optimization method with constraints. If no strict need exists for efficiency, an iterative method could be considered. We decreased the rotation matric’s dimension from three to two via Schur matric decomposition and unknown parameters satisfied the constraint c2 + d2 = 1. If the following triangle transformation is adopted, the degrees of freedom (DOFs) can be decreased from two to one. The Gröbner basis method can be used to solve polynomial equations [24]:
  7 in total

1.  Grid-free compressive mode extraction.

Authors:  Yongsung Park; Peter Gerstoft; Woojae Seong
Journal:  J Acoust Soc Am       Date:  2019-03       Impact factor: 1.840

2.  Comparison of prospective head motion correction with NMR field probes and an optical tracking system.

Authors:  Martin Eschelbach; Ali Aghaeifar; Jonas Bause; Jonas Handwerker; Jens Anders; Eva-Maria Engel; Axel Thielscher; Klaus Scheffler
Journal:  Magn Reson Med       Date:  2018-07-29       Impact factor: 4.668

3.  Pose Estimation from Line Correspondences: A Complete Analysis and a Series of Solutions.

Authors:  Chi Xu; Lilian Zhang; Li Cheng; Reinhard Koch
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-06-20       Impact factor: 6.226

4.  Adjoint Transformation Algorithm for Hand-Eye Calibration with Applications in Robotic Assisted Surgery.

Authors:  Krittin Pachtrachai; Francisco Vasconcelos; François Chadebecq; Max Allan; Stephen Hailes; Vijay Pawar; Danail Stoyanov
Journal:  Ann Biomed Eng       Date:  2018-07-26       Impact factor: 3.934

5.  Simultaneous Robot⁻World and Hand⁻Eye Calibration without a Calibration Object.

Authors:  Wei Li; Mingli Dong; Naiguang Lu; Xiaoping Lou; Peng Sun
Journal:  Sensors (Basel)       Date:  2018-11-15       Impact factor: 3.576

6.  Methods for Simultaneous Robot-World-Hand-Eye Calibration: A Comparative Study.

Authors:  Ihtisham Ali; Olli Suominen; Atanas Gotchev; Emilio Ruiz Morales
Journal:  Sensors (Basel)       Date:  2019-06-25       Impact factor: 3.576

7.  A computationally efficient method for hand-eye calibration.

Authors:  Zhiqiang Zhang; Lin Zhang; Guang-Zhong Yang
Journal:  Int J Comput Assist Radiol Surg       Date:  2017-07-19       Impact factor: 2.924

  7 in total
  2 in total

1.  Accuracy evaluation of hand-eye calibration techniques for vision-guided robots.

Authors:  Ikenna Enebuse; Babul K S M Kader Ibrahim; Mathias Foo; Ranveer S Matharu; Hafiz Ahmed
Journal:  PLoS One       Date:  2022-10-19       Impact factor: 3.752

2.  A Novel Concentric Circular Coded Target, and Its Positioning and Identifying Method for Vision Measurement under Challenging Conditions.

Authors:  Yan Liu; Xin Su; Xiang Guo; Tao Suo; Qifeng Yu
Journal:  Sensors (Basel)       Date:  2021-01-28       Impact factor: 3.576

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.