Elsevier

Pattern Recognition

Volume 35, Issue 7, July 2002, Pages 1617-1635
Pattern Recognition

A comparative review of camera calibrating methods with accuracy evaluation

https://doi.org/10.1016/S0031-3203(01)00126-1Get rights and content

Abstract

Camera calibrating is a crucial problem for further metric scene measurement. Many techniques and some studies concerning calibration have been presented in the last few years. However, it is still difficult to go into details of a determined calibrating technique and compare its accuracy with respect to other methods. Principally, this problem emerges from the lack of a standardized notation and the existence of various methods of accuracy evaluation to choose from. This article presents a detailed review of some of the most used calibrating techniques in which the principal idea has been to present them all with the same notation. Furthermore, the techniques surveyed have been tested and their accuracy evaluated. Comparative results are shown and discussed in the article. Moreover, code and results are available in internet.

Introduction

Camera calibration is the first step towards computational computer vision. Although some information concerning the measuring of scenes can be obtained by using uncalibrated cameras [1], calibration is essential when metric information is required. The use of precisely calibrated cameras makes the measurement of distances in a real world from their projections on the image plane possible [2], [3]. Some applications of this capability include:

  • 1.

    Dense reconstruction: Each image point determines an optical ray passing through the focal point of the camera towards the scene. The use of more than a single view of a motionless scene (taken from a stereoscopic system, a single moving camera, or even a structured light emitter) permits crossing both optical rays to get the metric position of the 3D points [4], [5], [6]. Obviously, the correspondence problem has to be previously solved [7].

  • 2.

    Visual inspection: Once a dense reconstruction of a measuring object is obtained, the reconstructed object can be compared with a stored model in order to detect any manufacturing imperfections such as bumps, dents or cracks. One potential application is visual inspection for quality control. Computerized visual inspection allows automatic and exhaustive examination of products, as opposed to the slow human inspection which usually implies a statistical approach [8].

  • 3.

    Object localization: When considering various image points from different objects, the relative position among these objects can be easily determined. This has many possible applications such as in industrial part assembly [9] and obstacle avoidance in robot navigation [10], [11], among others.

  • 4.

    Camera localization: When a camera is placed in the hand of a robot arm or on a mobile robot, the position and orientation of the camera can be computed by locating some known landmarks in the scene. If these measurements are stored, a temporal analysis allows the handler to determine the trajectory of the robot. This information can be used in robot control and path planning [12], [13], [14].

Camera calibration is divided into two phases. First, camera modelling deals with the mathematical approximation of the physical and optical behavior of the sensor by using a set of parameters. The second phase of camera calibration deals with the use of direct or iterative methods to estimate the values of these parameters. There are two kinds of parameters in the model which have to be considered. On the one hand, the intrinsic parameter set, which models the internal geometry and optical characteristics of the image sensor. Basically, intrinsic parameters determine how light is projected through the lens onto the image plane of the sensor. The other set of parameters are the extrinsic ones. The extrinsic parameters measure the position and orientation of the camera with respect to a world coordinate system, which, in turn, provides metric information with respect to a user-fixed coordinate system instead of the camera coordinate system.

Camera calibration can be classified according to several different criteria. For instance, (1) Linear versus non-linear camera calibration (usually differentiated depending on the modelling of lens distortion) [15]. (2) Intrinsic versus extrinsic camera calibration. Intrinsic calibration is concerned only with obtaining the physical and optical parameters of the camera [16], [17]. Besides, extrinsic calibration concerns the measurement of the position and orientation of the camera in the scene [18], [19]. (3) Implicit [20] versus explicit [21] calibration. Implicit calibration is the process of calibrating a camera without explicitly computing its physical parameters. Although, the results can be used for 3D measurement and the generation of image coordinates, they are useless for camera modelling as the obtained parameters do not correspond to the physical ones [22]. Finally, (4) the methods which use known 3D points as a calibrating pattern [23], [24] or even a reduced set of 3D points [25], [26], with respect to others which use geometrical properties in the scene such as vanishing lines [27] or other line features [28], [29].

These different approaches can also be classified regarding the calibration method used to estimate the parameters of the camera model:

  • 1.

    Non-linear optimization techniques. A calibrating technique becomes non-linear when any kind of lens imperfection is included in the camera model. In that case, the camera parameters are usually obtained through iteration with the constraint of minimizing a determined function. The minimizing function is usually the distance between the imaged points and the modelled projections obtained by iterating. The advantage of these iterating techniques is that almost any model can be calibrated and accuracy usually increases by increasing the number of iterations up to convergence. However, these techniques require a good initial guess in order to guarantee convergence. Some examples are described in classic photogrammetry [30] and Salvi [31].

  • 2.

    Linear techniques which compute the transformation matrix. These techniques use the least squares method to obtain a transformation matrix which relates 3D points with their 2D projections. The advantage here is the simplicity of the model which consists in a simple and rapid calibration. One drawback is that linear techniques are useless for lens distortion modelling, entailing a rough accuracy of the system. Moreover, it is sometimes difficult to extract the parameters from the matrix due to the implicit calibration used. Some references related to linear calibration can be found in Hall [20], Toscani-Faugeras [23], [32] and Ito [15].

  • 3.

    Two-step techniques. These techniques use a linear optimization to compute some of the parameters and, as a second step, the rest of the parameters are computed iteratively. These techniques permit a rapid calibration considerably reducing the number of iterations. Moreover, the convergence is nearly guaranteed due to the linear guess obtained in the first step. Two-step techniques make use of the advantages of the previously described methods. Some references are Tsai [24], Weng [33] and Wei [22].

This article is a detailed survey of some of the most frequently used calibrating techniques. The first technique was proposed by Hall in 1982 and is based on an implicit linear camera calibration by computing the 3×4 transformation matrix which relates 3D object points with their 2D image projections [20]. The latter work of Faugeras, proposed in 1986, was based on extracting the physical parameters of the camera from such a transformation technique, thus it is explained as the second technique [23], [32]. The following methods are based on non-linear explicit camera calibration, including the modelling of lens distortion. Hence, the first one is a simple adaptation of the Faugeras linear method with the aim of including radial lens distortion [31], [34]. The widely used method proposed by Tsai, which is based on a two-step technique modelling only radial lens distortion, is also detailed [24]. Finally, the complete model of Weng, which was proposed in 1992, including three different types of lens distortion, is explained as the last technique [33]. Note that one of the principal problems to understand a calibrating technique in detail is the lack of notation standardization in mathematical equations and the use of different sets of coordinate systems. Both limitations complicate the comparing of techniques, thus a great deal of effort has been made to present the survey using the same notation. All five techniques are explained herein and their 2D and 3D accuracy shown and discussed. A brief overview of camera accuracy evaluation [35] is included with the aim of using the same tools to compare different calibrating techniques implemented.

This article is structured as follows. Section 2 deals with camera modelling and how the camera model is gradually obtained by a sequence of geometrical transformations is explained. Section 3 describes the five different techniques of camera calibration, which estimate the parameters of the camera model. Then, a few methods for accuracy evaluation of camera calibrating techniques are explained in Section 4. Finally, both 2D and 3D accuracy of each calibration technique have been measured and their results are shown and compared. The paper ends with conclusions.

Section snippets

Camera model

A model is a mathematical formulation which approximates the behavior of any physical device by using a set of mathematical equations. Camera modelling is based on approximating the internal geometry along with the position and orientation of the camera in the scene. There are several camera models to choose from depending on the desired accuracy [15]. The simplest are based on linear transformations without modelling the lens distortion. However, there are also some non-linear models which

Calibrating methods

The calibrating method depends on the model used to approximate the behavior of the camera. The linear models, i.e. Hall and Faugeras–Toscani, use a least-squares technique to obtain the parameters of the model. However, non-linear calibrating methods, as with Faugeras–Toscani with distortion, Tsai and Weng, use a two-stage technique. As a first stage, they carry out a linear approximation with the aim of obtaining an initial guess and then a further iterative algorithm is used to optimize the

Accuracy evaluation

The systems used to evaluate the accuracy of camera calibration can be classified in two groups. The first group is based on analyzing the discrepancy between the real position of the 3D object point with respect to the 3D position estimated from its 2D projection. The second group compares the real position in pixels of a 2D image point with the calculated projection of the 3D object point on the image plane. In the following text, some of the most frequently used methods of accuracy

Experimental results

Instead of using our own experimental setup, we decided to download a list of corresponding points from the well-known Tsai's Camera Calibration Software Webpage (http://www.cs.cmu.edu/ ̃rgw/TsaiCode.html). Actually, results are always conditioned to the structure of the 3D points and the image processing tools used in segmentation and further points extraction. Hence, this decision was just taken to allow the scientific community to reproduce the same conditions. Then, the surveyed calibrating

Conclusions

This article surveys some of the most frequently used calibrating techniques. Effort has been made to unify the notation among these different methods, and they have been presented in a way the reader can easily understand. We can see that the differences among these techniques are mainly in the step concerning lens modelling. Also, the transformation from camera to image coordinates is slightly different in the method proposed by Tsai.

Furthermore, a survey on accuracy evaluation has been done.

Summary

In this article, we present a comparative study of the most commonly used camera calibrating methods of the last few decades. These techniques cover a wide range of the classical hard calibration of image sensors which begin from a previous knowledge of a set of 3D points and their corresponding 2D projections on an image plane in order to estimate the camera parameters. Hence, this study is presented describing a total of 5 different camera calibrating techniques which include implicit vs.

About the Author—JOAQUIM SALVI graduated in Computer Science in the Polytechnical University of Catalunya in 1993. He joined the Computer Vision and Robotics Group in the University of Girona, where he received the M.S. degree in Computer Science in July 1996 and the Ph.D. in Industrial Engineering in January 1998. He received the best thesis award in Industrial Engineering of the University of Girona. At present, he is an associate professor in the Electronics, Computer Engineering and

References (36)

  • Z. Zhang, The matching problem: the state of the art, Technical Report No. 2146, Institut National de Recherche en...
  • A. Casals

    Sensor Devices and Systems for Robotics, Vol. 52

    (1989)
  • A. Broggi

    Vision-based driving assistance in vehicles of the future

    IEEE Intell. Systems

    (1998)
  • L. Charbonnier, A. Fournier, Heading guidance and obstacles localization for an indoor mobile robot, IEEE International...
  • D. Khadraoui et al.

    Visual servoing in robotics scheme using a Camera/Laser-stripe sensor

    IEEE Int. J. Robotics Automat.

    (1996)
  • R.K. Lenz et al.

    Calibrating a cartesian robot with eye-on-hand configuration independent of eye-to-hand relationship

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1989)
  • M. Li, Camera calibration of a head-eye system for active vision, European Conference on Computer Vision, 1994, pp....
  • M. Ito

    Robot vision modelling—camera modelling and camera calibration

    Adv. Robotics

    (1991)
  • Cited by (452)

    • Adaptive adjustment of brightness and blur of the camera for high precision internal parameter calibration

      2024, Measurement: Journal of the International Measurement Confederation
    • Accurate weight coefficient estimation of multi-camera light field PIV through backpropagation neural network

      2024, Measurement: Journal of the International Measurement Confederation
    View all citing articles on Scopus

    About the Author—JOAQUIM SALVI graduated in Computer Science in the Polytechnical University of Catalunya in 1993. He joined the Computer Vision and Robotics Group in the University of Girona, where he received the M.S. degree in Computer Science in July 1996 and the Ph.D. in Industrial Engineering in January 1998. He received the best thesis award in Industrial Engineering of the University of Girona. At present, he is an associate professor in the Electronics, Computer Engineering and Automation Department of the University of Girona. His current interest are in the field of computer vision and mobile robotics, focusing on structured light, stereovision and camera calibration.

    About the Author—XAVIER ARMANGUE received the B.S. degree in Computer Science in the University of Girona in 1999 before joining the Computer Vision and Robotics Group. At present he is engaged in the study of stereovision systems for mobile robotics and he is working for his Ph.D. in the Computer Vision and Robotics Group in the University of Girona and in the Institute of Systems and Robotics in the University of Coimbra.

    About the Author—JOAN BATLLE graduated in Physics in the Autonomous University of Barcelona, received the Ph.D. in Computer Science in the Polytechnical University of Catalunya. At present, he is a professor in the Electronics, Computer Engineering and Automation Department; the leader of the Computer Vision and Robotics Group; and the director of the Institute of Informatics and Applications. His research activity is mainly focused on real-time vision and autonomous robots. He is actually involved in some governmental projects about underwater robots and technology transfer to industrial enterprises.

    This work has been supported by Spanish project CICYT TAP99-0443-CO5-01.

    View full text