[go: up one dir, main page]

Next Article in Journal
IEEE 802.15.4 Frame Aggregation Enhancement to Provide High Performance in Life-Critical Patient Monitoring Systems
Next Article in Special Issue
A Novel Strain-Based Method to Estimate Tire Conditions Using Fuzzy Logic for Intelligent Tires
Previous Article in Journal
Application of the PAMONO-Sensor for Quantification of Microvesicles and Determination of Nano-Particle Size Distribution
Previous Article in Special Issue
Practical Considerations in the Implementation of Collaborative Beamforming on Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Quantitative Comparison of Calibration Methods for RGB-D Sensors Using Different Technologies

by
Víctor Villena-Martínez
*,
Andrés Fuster-Guilló
,
Jorge Azorín-López
,
Marcelo Saval-Calvo
,
Jeronimo Mora-Pascual
,
Jose Garcia-Rodriguez
and
Alberto Garcia-Garcia
Department of Computer Technology, University of Alicante, Carretera San Vicente s/n, San Vicente del Raspeig 03690, Alicante, Spain
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(2), 243; https://doi.org/10.3390/s17020243
Submission received: 21 October 2016 / Accepted: 11 January 2017 / Published: 27 January 2017
(This article belongs to the Special Issue State-of-the-Art Sensors Technology in Spain 2016)
Figure 1
<p>Number of publications calculated from Scopus containing 3D imaging, depth cameras, RGB-D cameras, Microsoft Kinect sensors and other devices (including Asus Xtion, Primesense and Intel RealSense).</p> ">
Figure 2
<p>Projected pattern by Microsoft Kinect.</p> ">
Figure 3
<p>Time-of-Flight distance measurement [<a href="#B16-sensors-17-00243" class="html-bibr">16</a>].</p> ">
Figure 4
<p>Visual representation of the principal point.</p> ">
Figure 5
<p>Different models of optical distortions. (<b>a</b>) Barrel [<a href="#B51-sensors-17-00243" class="html-bibr">51</a>]; (<b>b</b>) Pincushion [<a href="#B52-sensors-17-00243" class="html-bibr">52</a>]; (<b>c</b>) Moustache [<a href="#B53-sensors-17-00243" class="html-bibr">53</a>].</p> ">
Figure 6
<p>Difference between spherical lens (<b>left</b>) and parabolical lens (<b>right</b>).</p> ">
Figure 7
<p>Infrared (IR) images of the chessboard. (<b>a</b>) Infrared image of the pattern; (<b>b</b>) Infrared image of the pattern without IR emitter; (<b>c</b>) Infrared image of the pattern without IR emitter and using a light bulb.</p> ">
Figure 8
<p>Some images of the chessboard used in the calibration process of Microsoft Kinect.</p> ">
Figure 9
<p>Plane fitting test, visual procedure. (Blue) 3D points of a wall. (Green) plane computed with RANSAC that best fit with the acquired points. The augmented part shows the point to plane orthogonal distances used to carry out this test.</p> ">
Figure 10
<p>Plane fitting test error for each calibration method for all cameras.</p> ">
Figure 11
<p>Plane fitting test error of each calibration method for Microsoft Kinect.</p> ">
Figure 12
<p>Plane fitting test error of each calibration method for Primesense Carmine 1.09.</p> ">
Figure 13
<p>Plane fitting test error of each calibration method for Microsoft Kinect V2.</p> ">
Figure 14
<p>Color (<b>left</b>) and depth (<b>right</b>) images of the markers distributed in the image.</p> ">
Figure 15
<p>Accuracy of the measurements. (<b>a</b>) Error of each method; (<b>b</b>) Error of each sensor group by method.</p> ">
Figure 16
<p>Registered objects. (<b>a</b>) Object 1 (Cube); (<b>b</b>) Object 2 (Taz); (<b>c</b>) Object 3 (Bob-omb).</p> ">
Figure 17
<p>Controlled environment.</p> ">
Figure 18
<p>Section of the cube acquired with Kinect v1 in the first row and the Primesense in the second row. The section shows the cube seen from the top. (<b>a</b>,<b>e</b>) Default; (<b>b</b>,<b>f</b>) Burrus; (<b>c</b>,<b>g</b>) Bouguet; (<b>d</b>,<b>h</b>) Herrera.</p> ">
Figure 19
<p>Frontal view of the reconstruction obtained with Primesense Carmine 1.09. (<b>a</b>) Default; (<b>b</b>) Burrus; (<b>c</b>) Bouguet; (<b>d</b>) Herrera.</p> ">
Figure 20
<p>Perspective view of the registration obtained with Microsoft Kinect v1. (<b>a</b>) Default; (<b>b</b>) Burrus; (<b>c</b>) Bouguet; (<b>d</b>) Herrera.</p> ">
Figure 21
<p>Side view of the reconstruction of the Object 1 obtained using different calibration methods with Microsoft Kinect v1. (<b>a</b>) Original; (<b>b</b>) Burrus; (<b>c</b>) Bouguet; (<b>d</b>) Herrera; (<b>e</b>) Real.</p> ">
Figure 22
<p>Frontal view of the reconstruction of the Object 2 obtained using different calibration methods with Microsoft Kinect v1. (<b>a</b>) Original; (<b>b</b>) Burrus; (<b>c</b>) Bouguet; (<b>d</b>) Herrera; (<b>e</b>) Real.</p> ">
Figure 23
<p>Noise distribution obtained with Kinect V2 in the acquisition of the cubes. <b>(a)</b> Perspective view; <b>(b)</b> Side view.</p> ">
Figure 24
<p>Registration error for different calibration methods in order to reconstruct a cube.</p> ">
Versions Notes

Abstract

:
RGB-D (Red Green Blue and Depth) sensors are devices that can provide color and depth information from a scene at the same time. Recently, they have been widely used in many solutions due to their commercial growth from the entertainment market to many diverse areas (e.g., robotics, CAD, etc.). In the research community, these devices have had good uptake due to their acceptable level of accuracy for many applications and their low cost, but in some cases, they work at the limit of their sensitivity, near to the minimum feature size that can be perceived. For this reason, calibration processes are critical in order to increase their accuracy and enable them to meet the requirements of such kinds of applications. To the best of our knowledge, there is not a comparative study of calibration algorithms evaluating its results in multiple RGB-D sensors. Specifically, in this paper, a comparison of the three most used calibration methods have been applied to three different RGB-D sensors based on structured light and time-of-flight. The comparison of methods has been carried out by a set of experiments to evaluate the accuracy of depth measurements. Additionally, an object reconstruction application has been used as example of an application for which the sensor works at the limit of its sensitivity. The obtained results of reconstruction have been evaluated through visual inspection and quantitative measurements.

1. Introduction

Broadly, in a 3D vision system, three main stages can be identified: acquisition, data processing and analysis. All of these stages are constrained by the application requirements. In the analysis stage, useful measures of the data are obtained depending on the requirements of the final application. The data processing stage modifies the data in order to align the acquired views. In the first stage, the data is acquired by the sensor, therefore it is crucial because its quality affects later stages.
In order to meet the requirements, the acquisition stage is constrained by three parameters [1]: the scene, the subject of interest and the camera. The light, shadows or the point of view are some factors of the scene that affect the captured data. For example, most sensors that project a pattern to determine the depth of the scene are limited to working indoors, because they would not be able to identify the pattern under intense sunlight. The subject of interest also affects the acquisition, for example, it could include specular surfaces causing the reflection of the pattern. Finally, the acquisition is affected by the camera, which is conditioned by its sensitivity, calibration and technology.
Focused on the sensor, the acquisition of 3D data could be performed with different types of devices, broadly classified into two groups:
  • Contact devices. They need a direct contact with the subject of interest to provide 3D information.
  • Contactless devices. They are able to provide 3D information from the distance.
This paper is focused on the second group, specifically on optical sensors because they are faster, more flexible and can provide complementary information about the scene’s colour. Comprehensive reviews of these sensors for 3D measurement have been presented in several papers [2,3,4,5]. Moreover, different taxonomies have been proposed to classify this kind of sensors as the interesting unifying framework proposed by Davis et al. [6]. However, this group has been widely classified into passive and active methods [4,7,8]:
  • Passive methods measure the scene radiance as a function of the object surface and environment characteristics using (usually) non-controlled ambient light external to the imaging system. Hence, only visible features of the scene are measured, providing high accuracy for well-defined features, such as targets and edges. However, unmarked surfaces are hard to measure [9]. In this category, techniques such as shape-from-X (e.g., shading, defocus, silhouettes, etc.), structure-from-motion and stereo are included. Stereo vision has received significant attention over the past decade in order to provide more accurate results and obtain them faster [10]. Usually, the methods use two or more calibrated RGB cameras to get the depth image by computing the disparity information from the images that conform to the system [11]. Stereoscopic cameras have been used for many purposes, including 3D reconstruction [12]. This technology can provide both colour and depth information, but it is required to be calibrated every time its location is changed, making its portability more difficult. Besides, they need the presence of texture to obtain the 3D information. In some devices, the distance between both cameras could be changed to fit the working range of the system.
  • Active methods use their own light source in the imaging system for the active illumination of the scene [13]. The sensor is usually focused on known features from this light source. Then, the illumination and the features are designed to be easily measured in most environments. Since they have difficulties with varying surface finish or sharp discontinuities such as edges [9], compared with the passive approach, active visual sensing techniques are in general more accurate and reliable [14]. Active sensors could be classified into two broad categories [15]: triangulation and time delay. The former rely on the triangulation principle using the light system, the scene and the sensor. The main differences between the methods include the nature of the controlled illumination (laser or incoherent light) and its geometry (beam, sheet, or projected pattern). Laser triangulators, structured light and moiré methods are examples that fall into this level. Time delay systems measure the time between emission and detection of light reflected by the scene (Time-of-flight, ToF) or the phase difference between two waves (Interferometry). Focusing on the ToF, pulsed-light and continuous wave modulation are the technologies available nowadays. Pulsed-light sensors directly measure the round-trip time of a light pulse. In order to obtain a range map, they use either rotating mirrors (LIDAR - Light Detection and Ranging o Laser Imaging Detection and Ranging) or a light diffuser (Flash LIDAR). LIDAR cameras usually operate outdoors and their range can be up to a few kilometers. Continuous wave sensors measure the phase difference between the emitted and received signals and usually operate indoors. Thier ambiguity-free range is usually fixed from 30 cm to 7 m [16,17]. A extensive comparison of ToF technologies can be found in [18].
Depending on the application requirements, a specific imaging device is selected according its characteristics. A comprehensive review including advantages and disadvantages for different applications of the most important techniques and sensors for the optical 3D measurement of a scene was presented by Sansoni et al. [5]. One of the conclusive remarks of the study was that most of the equipment available was significantly expensive, being an obstacle to a much wider distribution of 3D systems. However, they observed a trend towards a decrease in costs due to the increased competition of manufacturers and the technology evolution decreasing costs. One year later, popularized by Microsoft releasing the first generation of Kinect in November 2010 and focused on the entertainment market, consumer RGB-D (Red Green Blue and Depth) sensors emerged, which underlines this fact. As we can see in Figure 1, the introduction of Kinect sensor boosted the number of publications related to depth cameras and 3D imaging systems in general. Although the RGB-D topic was named in a few papers before 2011, it is in that year when the technological term was adopted, after the first Kinect version appeared. Moreover, the number of papers citing the Kinect camera has represented on average about 77% of the research publications on RGB-D sensors. These sensors combine one of the previous techniques (Structured Light, Time-of-Flight...) with an RGB camera to provide the colour and depth images of the scene using a common CMOS sensor for the color image and a different infrared technology to acquire the depth information.
RGB-D devices are mainly characterized by their low cost [19,20,21] but also they provide portability, high framerate, multidimensional perception with good accuracy for a wide range of applications. Hence, the use of these sensors has grown and been generalized from home entertainment systems to areas like robotics, medical informatics, etc. [22,23]. However, in some applications a calibration process could be needed in order to increase their accuracy, because they work in the limit of its sensitivity and some characteristics of the subject of interest might not be perceptible.
Khoshelham and Elberink [20] carried out a study into the accuracy of the Microsoft Kinect sensor. Some works propose algorithms and applications using RGB-D sensors. Han et al. [24] carried out a review focused on the Microsoft Kinect, but a more general review could be found in [25]. In [26], the authors performed a comparative of registration methods for RGB-D sensors. Weiss et al. [27] used a Microsoft Kinect to obtain a 3D model of a subject from multiple views around the body, avoiding the use of expensive devices. In [28], the authors were able to obtain a 3D model of a foot from multiple images around it using a PrimeSnese Carmine 1.09 and augmented reality markers. Jedvert [29] also used a Microsoft Kinect to obtain a 3D model of the head with hight quality textures. The work of Paier [30] aims to obtain a 3D model of a face for subject identification in security systems. In [31], the authors demonstrated that default parameters used by a Microsoft Kinect are not good enough for many applications.
To increase the accuracy of the data provided by these sensors, some works perform a calibration process. In [32], the authors propose an algorithm to calibrate the intrinsic parameters of both cameras, providing the necessary information to convert the disparity to meters. Zhang and Zhang [33] extend this work looking for correspondences between colour and depth images of a calibration pattern. Burrus [34] performed the calibration of a Kinect sensor based on the OpenCV calibration to obtain the intrinsic and extrinsic parameters. In [31], the authors noticed residual errors in depth images after calibration, estimating a fixed error for each pixel and calculating a correction pattern. Later, Daniel Herrera et al. [35] proposed a distortion correction on disparity images, taking into account that the magnitude of this error decreases as the distance from the subject increases. Subsequently, Raposo et al. [36] improved that proposal using fewer images for the input of the calibration process. Recently, Staranowicz et al. [37] proposed a method to estimate the parameters of a depth camera from images of a spheric object, using the Hough transform and performing a non linear minimization to obtain the results.
From the study of the state-of-the-art practice, it can be demonstrated that RGB-D sensors are useful for many applications, but their sensitivity is not suitable for problems in which higher accuracy of the data is required. To the best of our knowledge, there exist works carrying out a comparative of calibration methods for RGB-D sensors. For example, Staranowicz and Mariottini [38] made a comparison of three calibration methods [32,33,37] and, recently, Xiang et al. [39] did the same for others [34,35,40]. Meanwhile, Lachat et al. [13] made a comparison of the provided raw data, a calibration method and photogrammetry. Also, it is important to highlight the work carried out by Staranowicz et al. [41] in which they propose a comparative of different calibration methods using different known objects: checkerboard and spheres. However, all of them performed the evaluation of the methods and sensors using only a single consumer RGB-D sensor, the Microsoft Kinect v1 or v2. Analysing the technologies used by RGB-D sensors and the available calibration algorithms to improve their accuracy are very important topics that could be addressed simultaneously (i.e., quantifying the result of each algorithm in different technologies). In order to do so, in this paper, a comparative study of calibration algorithms applied to RGB-D sensors is presented in order to analyse the accuracy limits of this sensors.
The rest of the paper is structured as follows: in Section 2 the common technologies used by these sensors are explained and so are the three most common calibration methods. Section 3 shows the calibration results for each sensor and the developed experiments to test the accuracy for each one. An example of object reconstruction is used in order to evaluate the accuracy of the results in an application wherein the sensor works at the limit of its sensitivity. Finally, in Section 4 the conclusions of this work are shown.

2. Materials and Methods

In this section, a description of the RGB-D cameras and the calibration methods used in the quantitative comparison are presented.

2.1. RGB-D Cameras

Popularized by Microsoft, releasing the first generation of Kinect in November 2010 focused on the entertainment market, RGB-D sensors have been improved by different companies. Nowadays, many sensors could be found to provide depth. Table 1 shows the technical specifications of different low cost RGB-D cameras (price is less than 200 euros). Generally, consumer RGB-D cameras are mainly based on Structured light and ToF:
  • Structured Light (SL) based sensors are composed of a near-infrared emitter and an infrared (IR) camera. The infrared emitter projects a known pattern over the scene, simultaneously the IR camera gets the pattern and computes the disparity between the known and the observed pattern [42,43,44]. Usually, the infrared is chosen as the bandwidth of the projected pattern to avoid interfering with visible light in the scene. Nevertheless, a drawback of this technology is the impossibility of working in places where the illumination hinders the perception of the pattern [45]. More information about this technology can be found in [20]. For example, consumer RGB-D as Microsoft Kinect, Asux Xtion Pro or PrimeSense Carmine use structured light by projecting a speckle pattern over the scene (see Figure 2).
  • Time-of-Flight (ToF). As has previously been stated, ToF sensors obtain the distance to a subject of interest by measuring the time between the emission of a signal and its reflection from the subject. Consumer cameras that use this technology are based on a continuous wave sensor combined with a calibrated and internally synchronized RGB camera. A near-infrared emitter emits incoherent light, which is a modulated signal with a frequency ω. This light incises in the scene, producing a reflected signal with a phase shift ϕ with respect to the emitted signal (see Figure 3). Hence, the distance is given by the Equation (1), where c is the speed of light [46]. Microsoft Kinect V2 is the best representative example of this kind of cameras, achieving one of the best image resolutions among ToF cameras commercially available and an excellent compromise between depth accuracy and phase-wrapping ambiguity [18].
    d = c ϕ 4 π ω
In this study, Microsoft Kinect, Primesense Carmine 1.09 and Microsoft Kinect v2 have been selected. Structured light and ToF technologies for RGB-D cameras are represented by these sensors. Specifically, Kinect sensors have been selected because they are the most used and popular RGB-D cameras in the research community nowadays. As it was suggested by Figure 1, they represent the most papers dealing with this kind of sensors. The Microsoft Kinect V2 sensor has significant differences compared to its previous version. It is based on ToF technology with better resolution of 1920 × 1080 for the colour camera, but keeps the operation range of the depth camera. In case of the Primesense Carmine 1.09, although it is based on the same principle as Microsoft Kinect v1 and ASUS Xtion as popular cameras (in fact, it is the same Primesense patent [47] and uses the same PS1080 processor developed by the Primesense company), the operation range is different. The Carmine 1.09 sensor is a short range sensor, so its depth camera can operate between 0.35 m and 1.4 m, approximately, while the Microsoft Kinect works between 0.5 m and 4.5 m. The goal to include this sensor is to analyse the use of a specific short range device compared to the standard measuring range of the popular kinect devices for applications that require a short range, such as 3D reconstruction, for example.

2.2. Camera Calibration Parameters

Camera calibration is a necessary step in 3D computer vision in order to extract metric information from images [48]. It enables the determination of the camera geometric and optical characteristics and/or the relative position and orientation of the camera frame with respect to a world coordinate system [49]. One of the criteria that has to accomplish the calibration procedure is to be accurate because it is necessary to infer accurate 3D information from images. In other words, the calibration goal is to provide the measured data as close as possible to the real value. The more accurate the calibration model is, the more accurate the data that is provided by the camera system.
The calibration parameters of the cameras could be divided into two groups. One of them refers to those parameters which are specific to the lens geometry (intrinsic parameters), and those that are related to the relative position and orientation of the camera frame (extrinsic parameters). We refer to Hartley and Zisserman [50] for an extensive work of multi-view geometry in computer vision where the parameters are studied. For the sake of completeness, we briefly introduce the most relevant ones for this work below.

2.2.1. Intrinsic Parameters

Intrinsic parameters refer to the internal camera geometric and optical characteristics: focal length, distance between the optical centre of the lens and the photosensitive sensor; the principal point, represents the displacement of the optical axis, producing a displacement of the projection centre in the image (see Figure 4); and the distortion coefficients, which is the optical distortion model of a camera (see Figure 5). The latter refers to the variation of a straight projection due to the aberration of the lens. It is zero in the principal point, and increases with the distance.
The 3D point cloud is then calculated using the focal length and the principal point parameters, along with the depth image. The distortion is corrected to accurately provide the 3D data reducing the lens shape defects. For further details about the 3D point set estimation and optical parameter modeling, refer to Appendix A.
The distortion coefficients represent the optical distortion model of a camera. The two most common are the radial and the tangential ones. This is produced by the imperfect parabolic shape of the lenses, which are more spherical, producing the misalignment of the rays and resulting in a distorted image (see Figure 6).

2.2.2. Extrinsic Parameters

Extrinsic parameters refer to the relative position and orientation of the camera frame with respect to a world coordinate system. Specifically, in a multiple camera system, such as the stereo cameras or the RGB-D sensors studied here, multiple images are obtained from different coordinate systems. In this case, the extrinsic parameters describe the geometric relationship between the cameras that might be needed. Schulze [54] presented several calibration methods and discussed the accuracy to calibrate extrinsic parameters for aligning range sensors and colour cameras. For stereo matching, or RGB and Depth matching, it is necessary to align those images to a common coordinate system. This matching is carried out using the extrinsic parameters, which define the rotations and translations, the baseline which is the distance between the sensors, and the orientation of each camera because they are not perfectly parallel each other. For further information about the extrinsic parameters modeling, refer to Appendix A.

2.3. Calibration Methods

There are several methods to calibrate 3D sensors, most of them can be applied to RGB-D cameras. A comprehensive overview of the current approaches adopted for camera calibration in close-range photogrammetry and computer vision could be found in [55]. According to Xiang et al. [39], these methods can be classified in supervised and unsupervised calibration. The first ones perform the process acquiring images of targets with a particular shape or size, while unsupervised methods use the environment. This paper is focused in the first group, supervised calibration, due to the performance of unsupervised methods usually being lower. Besides, the methods can be distinguished between classical and those that are focused on 3D cameras of different technologies, ToF, SL, stereoscopic, etc. Usually, 3D cameras have two independent lenses and employ a technology to compute the depth. Classical methods calibrate cameras with one lens. Tt could be applied to calibrate each lens independently obtaining the intrinsic parameters, but they do not provide the extrinsic ones. Moreover, some calibration methods for RGB-D sensors are able to calibrate the parameters tat are employed to compute the depth information, which is specific to this technology.
A classification of calibration methods can be found in [41] including supervised and non-supervised calibration methods. This paper is focused on the most common supervised methods to calibrate 3D sensors (see Table 2). Usually, they use a set of images of a pattern composed of squares, known as chessboard or checkerboard. The corners of the chessboard are easily detected by a corner detector algorithm, but other kinds of patterns could be used. The methods have been evaluated according to a set of characteristics: the year of publication; the number of citations obtained; if the method performs a joint calibration, which is the calibration of both cameras simultaneously; the input data required by the algorithm; the type of target employed in the images; if the target is known by the algorithm or not; the number of images required to calibrate the sensor and the availability of the code.
The results obtained by Xiang et al. [39] showed that the best results where obtained with the methods of Daniel Herrera et al. [35] and Burrus [34]. In this paper, these two methods have been selected for the comparative. Moreover, the method of Bouguet [56] has been included in the comparative, due to it having been widely used in the literature and, in contrast to the other methods, the calibration of the infrared camera has to be performed with the infrared images only.

2.3.1. Bouguet Method

Bouguet [56] published a generic method for camera calibration based on the work of Zhang [64], Zhang [48] and Tsai [49]. The algorithm proposed by Zhang [64] only requires images of a planar pattern at different orientations. Later, this algorithm was adapted to work with 3D cameras [48], while Tsai [49] proposed a tow stage technique also for 3D camera calibration. The Bouguet method also includes an add-on to calibrate stereoscopic systems that allows us to calibrate both colour and depth cameras of an RGB-D sensor. Smisek et al. [31] used this method to calibrate a Microsoft Kinect, while Van Den Bergh and Van Gool [65] did the calibration of a ToF sensor with a colour camera coupled to it. The input to the calibration algorithm are the colour and infrared images obtained simultaneously.
Due to the IR emitter, the obtained infrared images are very noisy (see Figure 7a) and the corners of the calibration pattern can not be detected properly. In the images obtained without the IR emitter (Figure 7b) the chessboard is not perceptible because the image is very dark. In order to get images that could be used in the calibration process, a light bulb focused to the chessboard is needed (see Figure 7c).

2.3.2. Burrus Method

RGB Demo [34] is a set of tools and libraries to work with the data provided by a Microsoft Kinect sensor, but also could be used with devices supporting the same driver. One of the included tools is for calibrating this device using the calibration algorithm implemented in OpenCV based on Bouguet [56] and Hartley [66]. The calibration process is performed as if it were a stereoscopic system. Firstly, RGB, infrared and disparity images are obtained to calibrate the intrinsic parameters of both cameras individually. Then, a stereoscopic calibration process is done to get the extrinsic parameters.

2.3.3. Herrera Method

The method proposed by Daniel Herrera et al. [35] to calibrate colour cameras and a depth camera simultaneously have been developed with the objective of being accurate, practical and applicable to multiple sensors. The algorithm implements the intrinsic error model of the Microsoft Kinect depth camera, but it could be replaced to work with similar devices.
The intrinsic error model allows us to correct the distortion of the depth camera in the disparity image, and is based on the constant error in depth measurements that appear in this kind of sensors. Besides, this error decreases when the distances to the sensor increase.
The result of the algorithm provides a spatial distortion pattern D σ , which is a matrix of the same size of the depth image; the values α 0 , α 1 that represent the decadence of the distortion effect with the distance, and the values c 1 , c 0 to convert the disparity to meters. Then, the disparity for a given pixel ( u , v ) of the depth image can be corrected with Equation (2).
d k = d + D σ ( u , v ) · e x p ( α 0 α 1 d ˙ )
Where d is the disparity in the pixel ( u , v ) and d k is the corrected value of the disparity in that pixel. Once the disparity has been corrected, the distance in meters Z d for that pixel could be computed by Equation (3).
Z d = 1 c 1 · d k + c 0

3. Experimentation

In order to comparatively analyse the performance of the Bouguet, Burrus and Herrera calibration methods, three different sensors have been used as was stated in Section 2: Microsoft Kinect, Primesense Carmine 1.09 and Microsoft Kinect V2. For each sensor, different images of a chessboard pattern varying its position and orientation from the camera have been acquired (see Figure 8). This pattern is composed of 7 × 11 squares of 0.034 m of size. A subset of 60 images for each camera have been selected, which have been used for the calibration methods.

3.1. Calibration Results

The results obtained with each calibration method for Microsoft Kinect and Primesense Carmine 1.09 are shown in Table 3 and Table 4, respectively. The results for Microsoft Kinect V2 are in Table 5. Note that the values for the principal point ( c x , c y ) for the RGB camera obtained with the Burrus method for this sensor are not correct, it should be located near the center of the image with resolution 1920 ×1080, but the obtained coordinates are ( 345 . 85 , 251 . 59 ) due to the assumption of a fixed standard resolution. Additionally, the Microsoft Kinect V2 could not be calibrated using the method of Herrera because this camera can not provide the disparity images used by the method.

3.2. Experimental Results

We have carried out three experiments to evaluate the results of each calibration method. The plane fitting test was used by Khoshelham and Elberink [20] to evaluate the error in the distance, while with the measurements of the height and the markers, the accuracy of each method in combination with each camera is evaluated.

3.2.1. Plane Fitting Test

We have obtained different images of a wall at various distances (0.7 m, 0.8 m, 0.9 m, 1 m, 1.1 m, 1.2 m and 1.3 m) with each sensor, applying the corrections and computing the point cloud with the parameters provided by each calibration method. Also, images without any correction have been used to compare the accuracy obtained with the default parameters, which are unknown. Due to the difficultly of placing the sensor perfectly parallel to a wall, a square of 100 × 100 pixels from the center of the image has been extracted (blue points in Figure 9) computing the best plane that fits those points (green plane in Figure 9) using RANSAC [67] (Random Sample Consensus). Then, the outliers have been removed and the point-to-plane orthogonal distances with the remaining points have been computed. The error has been computed as the distance d from a point P = ( x 0 , y 0 , z 0 ) to a plane π A x + B y + C z + D = 0 . This distance corresponds to the perpendicular line from the point to the plane and its given by Equation (4).
d ( P , π ) = A x 0 + B y 0 + C z 0 + D A 2 + B 2 + C 2
Figure 10 shows the arithmetic mean error of each calibration method. As it can been seen, all methods improve the results obtained with the default parameters which gives an error of 12.18 mm. In particular, the method of Herrera provided the smallest error (7.67 mm), while the error for the Bouguet and Burrus methods were very close each other, 9.36 mm and 10.28 mm, respectively.
Analysing the standard deviation error for each sensor, it is possible to observe that the method of Herrera provides the best result in general terms for Microsoft Kinect (see Figure 11), with an std error of 5.73 mm, and Primesense Carmine 1.09 (see Figure 12), with an std error of 9.61 mm due to the distortion correction for the depth camera that provides this method. However, in the case of Primesense Carmine 1.09, the lowest average error is obtained with the Burrus method (9.044 mm). In case of Microsoft Kinect V2 (see Figure 13) there is not much difference between the default results and the calibrated ones, but in most cases the smallest error has been obtained with the default parameters. This is because the error in the depth is evaluated in this test, but only the calibration of the internal parameters of the infrared camera is performed. The calibration of the depth computation in a ToF sensor is complex due to it being difficult to calculate the frequency of the modulated signal and the phase shift of the reflected one. For this reason, there is not much difference between the calibrated and the raw results. Nevertheless, the lowest error for the Bouguet method has been obtained with this sensor (5.20 mm) which is based on ToF.

3.2.2. Measurement Error

The accuracy of the measurements of planar targets of size 10 cm × 20 cm have been compared. The targets have been acquired at two different distances from the camera, 1.5 m and 2 m, distributed among the image space (see Figure 14). Then, the obtained images have been corrected with the parameters provided by the calibration methods, and the height and width of the markers have been measured, analysing the error with the real size. Then, the arithmetic mean for each sensor and method has been calculated.
Figure 15a shows the arithmetic mean error group by method obtained in the corrected data in comparison with that obtained with the default parameters. The smallest error is provided by the data corrected with the parameters of the method of Herrera (with a mean error of 0.26 cm), while the default parameters provide the higher error (0.69 cm). The difference between the Bouguet and Burrus method is bigger than in the previous experiment, with values of 0.41 cm and 0.67, respectively.
Looking at this arithmetic mean error according to various sensors in Figure 15b, the best result is obtained with the Primesense Carmine 1.09 calibrated with the method of Herrera. However, the results of Microsoft Kinect with the same method are very close. It is important to highlight the results obtained for each camera with the calibration method of Bouguet, in which the sensor based on ToF gets better results than those based on structured light.

3.2.3. Object Registration

Additionally, since one of the most common applications for RGB-D sensor is 3D reconstruction, it has been used as example to compare the accuracy of each calibration method. This application is an example in which the sensor works at the limit of its sensitivity.In order to do that, the μ-MAR method [68] has been used, which performs a 3D reconstruction of an object from different views based on 3D markers (Figure 16a) to properly compute the transformation to align the views. Concretely, μ-MAR models the markers to reduce the effects of noise and register the scene with these models. This application is an example in which the sensor works at the limit of its sensitivity. The data have been acquired in a controlled environment (showed in Figure 17) including a set of 79 images for each object shown in Figure 16. The objects studied here are an 8 cm side cube (Figure 16a), that is one of the markers; Figure 16b is a 20-cm high Taz toy; a 5-cm Bom-omb toy shown in Figure 16c. The reason for using a marker as an object to analyze is because its registration quality is direct applied to the objects’ registration result, since the μ-MAR registration is based on the markers. The data from the acquisition has been corrected using the parameters from different methods, and then registered.
In order to evaluate the registration, visual inspection and a quantitative analysis have been performed. Regarding the visual inspection, we are going to pay attention to some details of the shapes to compare calibration methods. Concretely, the easiest shape is the cube. A section of the registered cube is presented to easily appreciate the quality of registration. Moreover, for the other objects, the arms of the Taz and the eyes of the Bob-omb are the regions that will be used to perceive the registration accuracy.
The first experimentation presented is the cube object for the Microsoft Kinect v1. The registration results are presented in the first row of Figure 18. The different views are well aligned when they are perfectly overlapped. On the other hand, if there is an error in the registration, it is possible to see how a single view is displaced from the rest. The default calibration (Figure 18a is clearly the worst registered result since many views are wrongly registered. Bouguet, shown in Figure 18c, has some error, like the left side of the cube, where a view is not accurately registered. Burrus achieves good results but has some views on the top and left side, which are not adequately registered (a slim gap can be seen between views). Herrera achieves the best result providing the most compact and accurate registration presented in Figure 18d. The second row of Figure 18 shows the results for the data acquired using the Primesense. In this case, despite the result being better than in the Kinect v1, the default calibration (Figure 18e) is the worst in terms of registration result. For the rest of the calibration methods, the registration result has some errors (areas where cube sides are misaligned) with the Herrera being the best in general terms since the average shape is more square-like.
After the study of the cube, which shows in a simple view the main accuracy of the registration with data from different calibrations, the experimentation with the second and third objects (Figure 16c,d) are studied.
The registration results using the Primesense Carmine 1.09 are presented in Figure 19. The data from Herrera (Figure 19d) achieves the best registration, confirming the previous experimentation. The Taz arms in this method are more compact and the Bob-omb eyes are better defined. Bouguet method Figure 19c) provides adequate data, but slightly worse (Bob-omb eyes and the spherical shape is distorted). Observing the original data, registration is clearly the worst in the four results, as is clear by looking at the poorly registered arms of the Taz.
Similar results are obtained with the Microsoft Kinect v1, showed in Figure 20. Figure 21 and Figure 22 show a side view of Taz and frontal of Bob-omb, for the registration results. As before, the arm registration could be used as a visual feature to visually evaluate the performance of Taz and the eyes of Bob-omb. Again, the data provided by Herrera (Figure 21d and Figure 22d) calibration achieves the best registration.
The Microsoft Kinect V2 RGB-D sensor has a particular noise distribution, presented in Figure 23. Due to this noise, the markers used by the μ M A R registration method cannot be accurately modeled, and hence, the registration cannot be done. Since the cubes are formed by planar surfaces, the model is obtained as a set of planes with ninety degrees between each other, located by fitting the points of each face in the point cloud. However, as it is possible to see in Figure 23a,b that the faces of the cube are prolonged in a non planar direction (e.g., the top face leans downwards). This makes it impossible to fit a planar model in a face and hence the method algorithm does not provide satisfactory results.
Finally, regarding the quantitative analysis, the study has been carried out using the cubes because the mathematical model of the shape to be reconstructed is known and can be used as the ground truth. Specifically, the euclidean distance from a point to the corresponding plane of the face of the cube has been used as a measure of the registration error. Figure 24 shows the average mean of distances for all points to the corresponding planes of the cube. The conclusions reached using visual inspection, related to the effect of calibration methods in the registration quality, are coherent with the measure of the error calculated by means of the mean point to plane distance. The highest registration error corresponds to the default calibration. Bouguet and Burrus have similar error levels. Finally, Herrera method achieves the lowest error. This conclusions are valid for both Kinect v1 and Primesense devices.

4. Conclusions

In this paper, three calibration algorithms have been compared by applying their results to three different RGB-D sensors. The obtained parameters for each camera haven been tested in different situations and the results have been analyzed. In addition, these parameters have been applied in 3D reconstruction of objects, which is an application for this kind of sensors where they work at the limit of their sensitivity. In the experiments, the results obtained through the calibration method of Herrera were the most accurate. Additionally, the best reconstruction was obtained with the data processed with this algorithm for Microsoft Kinect. In the reconstruction, the Microsoft Kinect sensor showed better results than the Primense Carmine 1.09, which is a short range camera. Besides, based on the quantitative experiments, it is possible to observe that from 1 meter to the structured light cameras, the error begins to increase. Moreover, the quantitative experiments showed the best results for Microsoft Kinect V2 applying the Bouguet method. However, it has not been possible to use this sensor for 3D reconstruction due to the flying pixels problem, which complicates the perception of the markers. Future works will include the use of other calibration methods, like the one proposed by Staranowicz et al. [37] and the denoising of the data acquired with Microsoft Kinect V2 so that it can be used with the μ M A R method, and compare the results of both versions of Kinect.

Author Contributions

V. Villena-Martinez, A. Fuster-Guilló, J. Azorín-Lopez and M. Saval-Calvo conceived the idea of the paper and reviewed the state-of-the-art publications to extract the knowledge from them, designing the experimentation; J. Mora-Pascual, J. Garcia-Rodriguez and A. Garcia-Garcia, carried out the experiments and analyzed the data. All the authors participated writing the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Appendix A.1. Intrinsic Parameters

Intrinsic parameters refer to the internal camera geometric and optical characteristics. Focal length, principal point, and distortion are the most common parameters that represent the visual capability of a camera lens.
Usually, the focal length and the principal point are represented in a 3 × 3 matrix which is named the intrinsic matrix (Equation (A1)) where f x , f y , c x and c y are the focal length and the principal point, respectively.
f x 0 c x 0 f y c y 0 0 1
For a given pixel (u,v) of the depth image with depth Z d , the coordinates of a point P is obtained with the Equation (A2), being f x d , f y d the focal length and c x d , c y d the principal point of the infrared camera.
P x = ( u c x d ) Z d f x d P y = ( v c y d ) Z d f y d P z = Z d
The distortion is given by the first terms of the Taylor series, but it could be simplified to the Equation (A3) for the radial distortion and Equation (A4) for the tangential [69], where ( i , j ) is the distorted position of a pixel in the image, r is the radius to the principal point and i c , j c is the corrected position. For more information about distortions, see the work of Weng et al. [70].
i c = i · ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 ) j c = j · ( 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 )
i c = i + [ 2 p 1 y + p 2 ( r 2 + 2 x 2 ) ] j c = j + [ p 1 ( r 2 + 2 y 2 ) + 2 p 2 x ]

Appendix A.2. Extrinsic Parameters

Extrinsic parameters refer to the relative position and orientation of the camera frame with respect to a world coordinate system. Specifically, in a multiple camera system, multiple images are obtained using different coordinate systems. In this case, the extrinsic parameters describe the geometric relationship between the cameras that might be needed to align the multiple images into a common coordinate system. Usually, using the extrinsic parameters of each camera (rotation and translation) and the calculated baseline, and the orientation between cameras, it is possible to correlate the information from multiple sources.
The extrinsic parameters are a 3 × 3 rotation matrix and a translation vector of size 3 × 1 . To align images from RGB-D devices, the point cloud is transformed and is projected on the colour sensor. Given a three dimensional point P of the point cloud, the transformations are applied through the Equation (A5).
P = R * P + T
where R and T are the rotation and translation matrices, respectively, and P is the transformed 3D point of the point cloud. Next, the data is re-projected on the colour sensor using its intrinsic parameters (Equation (A6)).
i = ( P . x · f x r g b P . z ) + c x r g b j = ( P . y · f y r g b P . z ) + c y r g b
where f x r g b , f y r g b , c x r g b and c y r g b are the focal length and principal point of the RGB camera. Applying these equations the bidimensional coordinates ( i , j ) in the colour image corresponding to that 3D point (P) of the point cloud are obtained.

References

  1. Saval-Calvo, M. Methodology Based on Registration Techniques for Representing Subjects and Their Deformations Acquired from General Purpose 3D Sensors. Ph.D. Thesis, University of Alicante, San Vicente del Raspeig, Spain, May 2015. [Google Scholar]
  2. Blais, F. Review of 20 years of range sensor development. J. Electron. Imaging 2004, 13, 231–240. [Google Scholar] [CrossRef]
  3. Chen, F.; Brown, G.M.; Song, M. Overview of three-dimensional shape measurement using optical methods. Opt. Eng. 2000, 39, 10–22. [Google Scholar]
  4. Besl, P.J. Active, optical range imaging sensors. Mach. Vis. Appl. 1988, 1, 127–152. [Google Scholar] [CrossRef]
  5. Sansoni, G.; Trebeschi, M.; Docchio, F. State-of-the-art and applications of 3D imaging sensors in industry, cultural heritage, medicine, and criminal investigation. Sensors 2009, 9, 568–601. [Google Scholar] [CrossRef] [PubMed]
  6. Davis, J.; Ramamoorthi, R.; Rusinkiewicz, S. Spacetime stereo: A unifying framework for depth from triangulation. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 2, pp. II-359–II-366.
  7. Curless, B. Overview of active vision techniques. In Proceedings of the SIGGRAPH, 2000, Los Angesles, CA, USA, 9 August 1999; Volume 99.
  8. Poussart, D.; Laurendeau, D. 3-D sensing for industrial computer vision. In Advances in Machine Vision; Springer: New York, NY, USA, 1989; pp. 122–159. [Google Scholar]
  9. El-Hakim, S.F.; Beraldin, J.A.; Blais, F. Comparative evaluation of the performance of passive and active 3D vision systems. Proc. SPIE 1995, 2646, 14–25. [Google Scholar]
  10. Tippetts, B.; Lee, D.J.; Lillywhite, K.; Archibald, J. Review of stereo vision algorithms and their suitability for resource-limited systems. J. Real-Time Image Process. 2016, 11, 5–25. [Google Scholar] [CrossRef]
  11. Lazaros, N.; Sirakoulis, G.C.; Gasteratos, A. Review of Stereo Vision Algorithms: From Software to Hardware. Int. J. Optomech. 2008, 2, 435–462. [Google Scholar] [CrossRef]
  12. Kasper, A.; Xue, Z.; Dillmann, R. The KIT object models database: An object model database for object recognition, localization and manipulation in service robotics. Int. J. Robot. Res. 2012, 31, 927–934. [Google Scholar] [CrossRef]
  13. Lachat, E.; Macher, H.; Landes, T.; Grussenmeyer, P. Assessment and calibration of a RGB-D camera (Kinect v2 Sensor) towards a potential use for close-range 3D modeling. Remote Sens. 2015, 7, 13070–13097. [Google Scholar] [CrossRef]
  14. Chen, S.; Li, Y.; Wang, W.; Zhang, J. Active Sensor Planning for Multiview Vision Tasks; Springer: Heidelberg, Germany, 2008; Volume 1. [Google Scholar]
  15. Godin, G.; Beraldin, J.A.; Taylor, J.; Cournoyer, L.; Rioux, M.; El-Hakim, S.; Baribeau, R.; Blais, F.; Boulanger, P.; Domey, J.; et al. Active Optical 3D Imaging for Heritage Applications. IEEE Comput. Graph. Appl. 2002, 22, 24–36. [Google Scholar] [CrossRef]
  16. Foix, S.; Alenyà, G.; Torras, C. Lock-in time-of-flight (ToF) cameras: A survey. IEEE Sens. J. 2011, 11, 1917–1926. [Google Scholar] [CrossRef] [Green Version]
  17. Cui, Y.; Schuon, S.; Chan, D.; Thrun, S.; Theobalt, C. 3D shape scanning with a time-of-flight camera. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1173–1180.
  18. Horaud, R.; Hansard, M.; Evangelidis, G.; Clément, M. An Overview of Depth Cameras and Range Scanners Based on Time-of-Flight Technologies. Mach. Vis. Appl. J. 2016, 27, 1005–1020. [Google Scholar] [CrossRef] [Green Version]
  19. Lai, K.; Bo, L.; Ren, X.; Fox, D. Consumer Depth Cameras for Computer Vision. In Consumer Depth Cameras for Computer Vision; Springer: London, UK, 2013; p. 167. [Google Scholar]
  20. Khoshelham, K.; Elberink, S.O. Accuracy and resolution of Kinect depth data for indoor mapping applications. Sensors 2012, 12, 1437–1454. [Google Scholar] [CrossRef] [PubMed]
  21. Henry, P.; Krainin, M.; Herbst, E.; Ren, X.; Fox, D. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. Int. J. Robot. Res. 2012, 31, 647–663. [Google Scholar] [CrossRef]
  22. Meng, M.; Fallavollita, P.; Blum, T.; Eck, U.; Sandor, C.; Weidert, S.; Waschke, J.; Navab, N. Kinect for interactive AR anatomy learning. In Proceedings of the 2013 IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2013, Adelaide, Australia, 1–4 October 2013; pp. 277–278.
  23. Zondervan, D.K.; Secoli, R.; Darling, A.M.; Farris, J.; Furumasu, J.; Reinkensmeyer, D.J. Design and Evaluation of the Kinect-Wheelchair Interface Controlled (KWIC) Smart Wheelchair for Pediatric Powered Mobility Training. Assist. Technol. 2015, 27, 183–192. [Google Scholar] [CrossRef] [PubMed]
  24. Han, J.; Shao, L.; Xu, D.; Shotton, J. Enhanced computer vision with Microsoft Kinect sensor: A review. IEEE Trans. Cybern. 2013, 43, 1318–1334. [Google Scholar] [PubMed]
  25. Shao, L.; Han, J.; Kohli, P.; Zhang, Z. Computer Vision and Machine Learning with RGB-D Sensors; Springer: Cham, Switzerland, 2014; p. 313. [Google Scholar]
  26. Morell-Gimenez, V.; Saval-Calvo, M.; Azorin-Lopez, J.; Garcia-Rodriguez, J.; Cazorla, M.; Orts-Escolano, S.; Fuster-Guillo, A. A comparative study of registration methods for RGB-D video of static scenes. Sensors 2014, 14, 8547–8576. [Google Scholar] [CrossRef] [PubMed]
  27. Weiss, A.; Hirshberg, D.; Black, M.J. Home 3D Body Scans from Noisy Image and Range Data. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1951–1958.
  28. Lovato, C.; Bissolo, E.; Lanza, N.; Stella, A.; Giachetti, A. A Low Cost and Easy to Use Setup for Foot Scanning. In Proceedings of the 5th International Conference on 3D Body Scanning Technologies, Lugano, Switzerland, 21–22 October 2014; pp. 365–371.
  29. Jedvert, M. 3D Head Scanner. Master’s Thesis, Chalmers University of Technology, Göteborg, Sweden, 2013. [Google Scholar]
  30. Paier, W. Acquisition of 3D-Head-Models Using SLR-Cameras and RGBZ-Sensors. Master’s thesis, Freie Universität Berlin, Berlin, Germany, 2013. [Google Scholar]
  31. Smisek, J.; Jancosek, M.; Pajdla, T. 3D with Kinect. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1154–1160.
  32. Herrera, C.D.; Kannala, J.; Heikkilä, J. Accurate and practical calibration of a depth and color camera pair. In Proceedings of the 14th International Conference on Computer Analysis of Images and Patterns, Seville, Spain, 29–31 August 2011; pp. 437–445.
  33. Zhang, C.; Zhang, Z. Calibration between depth and color sensors for commodity depth cameras. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6.
  34. Burrus, N. Kinect RGB Demo. Manctl Labs. Available online: http://rgbdemo.org/ (accessed on 21 January 2017).
  35. Daniel Herrera, C.; Kannala, J.; Heikkilä, J. Joint depth and color camera calibration with distortion correction. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2058–2064. [Google Scholar] [CrossRef] [PubMed]
  36. Raposo, C.; Barreto, J.P.; Nunes, U. Fast and accurate calibration of a kinect sensor. In Proceedings of the 2013 International Conference on 3DTV-Conference, Seattle, WA, USA, 29 June–1 July 2013; pp. 342–349.
  37. Staranowicz, A.; Brown, G.R.; Morbidi, F.; Mariottini, G.L. Easy-to-use and accurate calibration of RGB-D cameras from spheres. In Proceedings of the 6th Pacific-Rim Symposium on Image and Video Technology, PSIVT 2013, Guanajuato, Mexico, 28 October–1 November 2013; pp. 265–278.
  38. Staranowicz, A.; Mariottini, G.L. A comparative study of calibration methods for Kinect-style cameras. In Proceedings of the 5th International Conference on PErvasive Technologies Related to Assistive Environments—PETRA ’12, Heraklion, Greece, 6–9 June 2012; ACM Press: New York, NY, USA, 2012; p. 1. [Google Scholar]
  39. Xiang, W.; Conly, C.; McMurrough, C.D.; Athitsos, V. A review and quantitative comparison of methods for kinect calibration. In Proceedings of the 2nd international Workshop on Sensor-Based Activity Recognition and Interaction—WOAR ’15, Rostock, Germany, 25–26 June 2015; ACM Press: New York, NY, USA, 2015; pp. 1–6. [Google Scholar]
  40. Teichman, A.; Miller, S.; Thrun, S. Unsupervised Intrinsic Calibration of Depth Sensors via SLAM. In Proceedings of the Robotics Science and Systems 2013, Berlin, Germany, 24–28 June 2013; Volume 248.
  41. Staranowicz, A.N.; Brown, G.R.; Morbidi, F.; Mariottini, G.L. Practical and accurate calibration of RGB-D cameras using spheres. Comput. Vis. Image Underst. 2015, 137, 102–114. [Google Scholar] [CrossRef]
  42. Salvi, J.; Pagès, J.; Batlle, J. Pattern codification strategies in structured light systems. Pattern Recognit. 2004, 37, 827–849. [Google Scholar] [CrossRef]
  43. Salvi, J.; Fernandez, S.; Pribanic, T.; Llado, X. A state of the art in structured light patterns for surface profilometry. Pattern Recognit. 2010, 43, 2666–2680. [Google Scholar] [CrossRef]
  44. Herakleous, K.; Poullis, C. 3DUNDERWORLD-SLS: An Open-Source Structured-Light Scanning System for Rapid Geometry Acquisition. arXiv, 2014; 1–30arXiv:1406.6595. [Google Scholar]
  45. Gupta, M.; Yin, Q.; Nayar, S.K. Structured light in sunlight. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 545–552.
  46. Fuchs, S.; Hirzinger, G. Extrinsic and depth calibration of ToF-cameras. In Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, Anchorage, AK, USA, 23–28 June 2008.
  47. Freedman, B.; Shpunt, A.; Machline, M.; Arieli, Y. Depth Mapping Using Projected Patterns. U.S. Patent 8,493,496 B2, 23 July 2013. [Google Scholar]
  48. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  49. Tsai, R.Y. A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and Lenses. IEEE J. Robot. Autom. 1987, 3, 323–344. [Google Scholar] [CrossRef]
  50. Hartley, R.I.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  51. WolfWings File:Barrel distortion.svg. Available online: https://en.wikipedia.org/wiki/File:Barrel_distortion.svg (accessed on 22 January 2017).
  52. WolfWings File:Pincushion distortion.svg. Available online: https://en.wikipedia.org/wiki/File:Pincushion_distortion.svg (accessed on 22 January 2017).
  53. WolfWings File:Mustache distortion.svg. Available online: https://en.wikipedia.org/wiki/File:Mustache_distortion.svg (accessed on 22 January 2017).
  54. Schulze, M. An Approach for Calibration of a Combined RGB-Sensor and 3D Camera Device; Institute of Photogrammetry and Remote Sensing, Technische Universität Dresde: Dresden, Germany, 2011. [Google Scholar]
  55. Remondino, F.; Fraser, C. Digital camera calibration methods: Considerations and comparisons. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2006, 36, 266–272. [Google Scholar]
  56. Bouguet, J.Y. Camera Calibration Toolbox for Matlab. 2004. Available online: https://www.vision.caltech.edu/bouguetj/calib_doc/ (accessed on 22 December 2016).
  57. Daniel, H.C. Kinect Calibration Toolbox; Center for Machine Vision Research, University of Oulu: Oulu, Finland, 2012; Available online: http://www.ee.oulu.fi/~dherrera/kinect/ (accessed on 22 December 2016).
  58. Manuel Fernandez, E.L.; Lucas, T.; Marcelo, G. ANSI C Implementation of Classical Camera Calibration Algorithms: Tsai and Zhang. Available online: http://webserver2.tecgraf.puc-rio.br/~mgattass/calibration/ (accessed on 22 December 2016).
  59. Raposo, C.; Barreto, J.P.; Nunes, U. EasyKinCal. Available online: http://arthronav.isr.uc.pt/~carolina/kinectcalib/ (accessed on 22 December 2016).
  60. Staranowicz, A.; Brown, G.; Morbidi, F.; Mariottini, G. Easy-to-Use and Accurate Calibration of RGB-D Cameras from Spheres. Available online: http://ranger.uta.edu/~gianluca/research/assistiverobotics_rgbdcalibration.html (accessed on 22 December 2016).
  61. Lichti, D.D. Self-calibration of a 3D range camera. Archives 2008, 37, 1–6. [Google Scholar]
  62. Zhu, J.; Wang, L.; Yang, R.; Davis, J. Fusion of time-of-flight depth and stereo for high accuracy depth maps. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8.
  63. Lindner, M.; Kolb, A. Lateral and Depth Calibration of PMD-Distance Sensors. In Advances in Visual Computing; Springer: Berlin/Heidelberg, Germany, 2006; pp. 524–533. [Google Scholar]
  64. Zhang, Z.Z.Z. Flexible camera calibration by viewing a plane from unknown orientations. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–27 September 1999; Volume 1, pp. 666–673.
  65. Van Den Bergh, M.; Van Gool, L. Combining RGB and ToF cameras for real-time 3D hand gesture interaction. In Proceedings of the 2011 IEEE Workshop on Applications of Computer Vision, WACV 2011, Kona, HI, USA, 5–7 January 2011; pp. 66–72.
  66. Hartley, R.I. Theory and Practice of Projective. Int. J. Comput. Vis. 1998, 35, 115–127. [Google Scholar] [CrossRef]
  67. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  68. Saval-Calvo, M.; Azorin-Lopez, J.; Fuster-Guillo, A.; Mora-Mora, H. μ-MAR: Multiplane 3D Marker based Registration for depth-sensing cameras. Expert Syst. Appl. 2015, 42, 9353–9365. [Google Scholar] [CrossRef]
  69. Bradski, G.; Kaehler, A. Learning OpenCV: Computer Vision with the OpenCV Library; O’Reilly Media, Inc.: Newton, MA, USA, 2008. [Google Scholar]
  70. Weng, J.; Coher, P.; Herniou, M. Camera Calibration with Distortion Models and Accuracy Evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 965–980. [Google Scholar] [CrossRef]
Figure 1. Number of publications calculated from Scopus containing 3D imaging, depth cameras, RGB-D cameras, Microsoft Kinect sensors and other devices (including Asus Xtion, Primesense and Intel RealSense).
Figure 1. Number of publications calculated from Scopus containing 3D imaging, depth cameras, RGB-D cameras, Microsoft Kinect sensors and other devices (including Asus Xtion, Primesense and Intel RealSense).
Sensors 17 00243 g001
Figure 2. Projected pattern by Microsoft Kinect.
Figure 2. Projected pattern by Microsoft Kinect.
Sensors 17 00243 g002
Figure 3. Time-of-Flight distance measurement [16].
Figure 3. Time-of-Flight distance measurement [16].
Sensors 17 00243 g003
Figure 4. Visual representation of the principal point.
Figure 4. Visual representation of the principal point.
Sensors 17 00243 g004
Figure 5. Different models of optical distortions. (a) Barrel [51]; (b) Pincushion [52]; (c) Moustache [53].
Figure 5. Different models of optical distortions. (a) Barrel [51]; (b) Pincushion [52]; (c) Moustache [53].
Sensors 17 00243 g005
Figure 6. Difference between spherical lens (left) and parabolical lens (right).
Figure 6. Difference between spherical lens (left) and parabolical lens (right).
Sensors 17 00243 g006
Figure 7. Infrared (IR) images of the chessboard. (a) Infrared image of the pattern; (b) Infrared image of the pattern without IR emitter; (c) Infrared image of the pattern without IR emitter and using a light bulb.
Figure 7. Infrared (IR) images of the chessboard. (a) Infrared image of the pattern; (b) Infrared image of the pattern without IR emitter; (c) Infrared image of the pattern without IR emitter and using a light bulb.
Sensors 17 00243 g007
Figure 8. Some images of the chessboard used in the calibration process of Microsoft Kinect.
Figure 8. Some images of the chessboard used in the calibration process of Microsoft Kinect.
Sensors 17 00243 g008
Figure 9. Plane fitting test, visual procedure. (Blue) 3D points of a wall. (Green) plane computed with RANSAC that best fit with the acquired points. The augmented part shows the point to plane orthogonal distances used to carry out this test.
Figure 9. Plane fitting test, visual procedure. (Blue) 3D points of a wall. (Green) plane computed with RANSAC that best fit with the acquired points. The augmented part shows the point to plane orthogonal distances used to carry out this test.
Sensors 17 00243 g009
Figure 10. Plane fitting test error for each calibration method for all cameras.
Figure 10. Plane fitting test error for each calibration method for all cameras.
Sensors 17 00243 g010
Figure 11. Plane fitting test error of each calibration method for Microsoft Kinect.
Figure 11. Plane fitting test error of each calibration method for Microsoft Kinect.
Sensors 17 00243 g011
Figure 12. Plane fitting test error of each calibration method for Primesense Carmine 1.09.
Figure 12. Plane fitting test error of each calibration method for Primesense Carmine 1.09.
Sensors 17 00243 g012
Figure 13. Plane fitting test error of each calibration method for Microsoft Kinect V2.
Figure 13. Plane fitting test error of each calibration method for Microsoft Kinect V2.
Sensors 17 00243 g013
Figure 14. Color (left) and depth (right) images of the markers distributed in the image.
Figure 14. Color (left) and depth (right) images of the markers distributed in the image.
Sensors 17 00243 g014
Figure 15. Accuracy of the measurements. (a) Error of each method; (b) Error of each sensor group by method.
Figure 15. Accuracy of the measurements. (a) Error of each method; (b) Error of each sensor group by method.
Sensors 17 00243 g015
Figure 16. Registered objects. (a) Object 1 (Cube); (b) Object 2 (Taz); (c) Object 3 (Bob-omb).
Figure 16. Registered objects. (a) Object 1 (Cube); (b) Object 2 (Taz); (c) Object 3 (Bob-omb).
Sensors 17 00243 g016
Figure 17. Controlled environment.
Figure 17. Controlled environment.
Sensors 17 00243 g017
Figure 18. Section of the cube acquired with Kinect v1 in the first row and the Primesense in the second row. The section shows the cube seen from the top. (a,e) Default; (b,f) Burrus; (c,g) Bouguet; (d,h) Herrera.
Figure 18. Section of the cube acquired with Kinect v1 in the first row and the Primesense in the second row. The section shows the cube seen from the top. (a,e) Default; (b,f) Burrus; (c,g) Bouguet; (d,h) Herrera.
Sensors 17 00243 g018
Figure 19. Frontal view of the reconstruction obtained with Primesense Carmine 1.09. (a) Default; (b) Burrus; (c) Bouguet; (d) Herrera.
Figure 19. Frontal view of the reconstruction obtained with Primesense Carmine 1.09. (a) Default; (b) Burrus; (c) Bouguet; (d) Herrera.
Sensors 17 00243 g019
Figure 20. Perspective view of the registration obtained with Microsoft Kinect v1. (a) Default; (b) Burrus; (c) Bouguet; (d) Herrera.
Figure 20. Perspective view of the registration obtained with Microsoft Kinect v1. (a) Default; (b) Burrus; (c) Bouguet; (d) Herrera.
Sensors 17 00243 g020
Figure 21. Side view of the reconstruction of the Object 1 obtained using different calibration methods with Microsoft Kinect v1. (a) Original; (b) Burrus; (c) Bouguet; (d) Herrera; (e) Real.
Figure 21. Side view of the reconstruction of the Object 1 obtained using different calibration methods with Microsoft Kinect v1. (a) Original; (b) Burrus; (c) Bouguet; (d) Herrera; (e) Real.
Sensors 17 00243 g021
Figure 22. Frontal view of the reconstruction of the Object 2 obtained using different calibration methods with Microsoft Kinect v1. (a) Original; (b) Burrus; (c) Bouguet; (d) Herrera; (e) Real.
Figure 22. Frontal view of the reconstruction of the Object 2 obtained using different calibration methods with Microsoft Kinect v1. (a) Original; (b) Burrus; (c) Bouguet; (d) Herrera; (e) Real.
Sensors 17 00243 g022
Figure 23. Noise distribution obtained with Kinect V2 in the acquisition of the cubes. (a) Perspective view; (b) Side view.
Figure 23. Noise distribution obtained with Kinect V2 in the acquisition of the cubes. (a) Perspective view; (b) Side view.
Sensors 17 00243 g023
Figure 24. Registration error for different calibration methods in order to reconstruct a cube.
Figure 24. Registration error for different calibration methods in order to reconstruct a cube.
Sensors 17 00243 g024
Table 1. Technical specifications of consumer RGB-D cameras. SL: Structured light, ToF: Time of Flight.
Table 1. Technical specifications of consumer RGB-D cameras. SL: Structured light, ToF: Time of Flight.
SensorMeasuring Range (m)ErrorField of View HxV (Degrees)Resolution Colour/DepthDepth Resolution (cm)TechnologyFPS
Kinect v10.8–3.5<4 cm57 × 43640 × 480
640 × 480
1 @ 2 mSL15/30
Carmine 1.080.8–3.5-57.5 × 45640 × 480
640 × 480
1.2 @ 2 mSL60
Carmine 1.090.35–1.4-57.5 × 45640 × 480
640 × 480
0.1 @ 0.5 mSL60
Xtion Pro0. 8–3. 5-58 × 451280 × 1024
640 × 480
1 @ 2 mSL30/60
Real Sense0. 2–1. 21%59 × 461920 × 1080
640 × 480
-SL30/60
Kinect v20. 5–4. 50.5%70 × 601920 × 1080
512 × 424
2 @ 2 mToF15/30
Senz3D0. 2–1. 0-74 × 41. 61080 × 720
320 × 240
-ToF30
Table 2. Common supervised calibration methods (I = IR, D = Disparity, Z = Depth, C = Color).
Table 2. Common supervised calibration methods (I = IR, D = Disparity, Z = Depth, C = Color).
MethodYearCitationsJoint CalibrationInput DataType of TargetKnown TargetNumber of Images (Approx.)Available Code
Daniel Herrera et al. [35]2012223YD,CChessboardY20Y [57]
Zhang and Zhang [33]2011107YZ,CChessboardY12Y [58]
[34]201137YI,Z,CChessboardY30Y [34]
Bouguet [56]20042721NI,CChessboardY20Y [56]
Raposo et al. [36]201330YD,CChessboardY10Y [59]
Staranowicz et al. [37]201413YZ,CSpheresN-Y [60]
Tsai [49]19877113NCFlat surface with squaresY1–8Y [58]
Fuchs and Hirzinger [46]2008150NZChessboard + robotic armY50N
Lichti [61]2008452NZRectangular targets of different sizesN-N
Jiejie Zhu et al. [62]2008251NZChessboardY-N
Lindner and Kolb [63]200776NZChessboardY68N
Table 3. Calibration results for Microsoft Kinect.
Table 3. Calibration results for Microsoft Kinect.
BurrusBouguetHerrera
RGB CameraIR CameraRGB CameraIR CameraRGB CameraIR Camera
f x 523 . 24 595 . 99 523 . 16 ± 1 . 40 588 . 18 ± 1 . 58 522 . 55 ± 0 . 25 586 . 80 ± 0 . 45
f y 521 . 68 592 . 44 521 . 32 ± 1 . 35 586 . 00 ± 1 . 52 520 . 24 ± 0 . 25 577 . 70 ± 0 . 59
c x 328 . 65 314 . 43 330 . 14 ± 1 . 04 315 . 83 ± 1 . 21 329 . 76 ± 0 . 33 318 . 92 ± 0 . 35
c y 257 . 03 227 . 05 257 . 01 ± 1 . 14 245 . 20 ± 1 . 25 257 . 59 ± 0 . 37 231 . 46 ± 0 . 37
k 1 0 . 0215 0 . 1567 0 . 1475 ± 0 . 00609 0 . 0724 ± 0 . 0052 0 . 1930 ± 0 . 0024 0
k 2 0 . 6927 0 . 6467 0 . 2735 ± 0 . 0116 0 . 1306 ± 0 . 01 0 . 5651 ± 0 . 012 0
k 3 0 . 7170 0 . 8859 00 0 . 4843 ± 0 . 0176 0
p 1 0 . 0007 0 . 0012 0 . 0014 ± 0 . 00082 0 . 0009 ± 0 . 0007 0 . 0006 ± 0 . 0004 0
p 2 0 . 0005 0 . 0004 0 0 . 0013 ± 0 . 00071 0 . 0003 ± 0 . 0002 0
c 0 3 . 0946 ± 0 . 0035
c 1 0 . 0028 ± 3 . 7100 × 10 6
α 0 1 . 2521 ± 0 . 0510
α 1 0 . 0022 ± 7 . 4073 × 10 5
R 0 . 9995 0 . 0082 0 . 0052 0 . 0081 0 . 9988 0 . 0125 0 . 0053 0 . 0125 0 . 9999 0 . 0076 0 . 0031 0 . 0078 ± 0 . 0012 0 . 0016 0 . 0004 1 0 . 0077 0 . 0047 0 . 0077 0 . 9999 0 . 0084 0 . 0048 0 . 0084 1 ± 6 . 9160 × 10 4 5 . 9122 × 10 4 3 . 4263 × 10 4
T 0 . 0255 0 . 0026 0 . 0068 0 . 0250 0 . 0004 0 . 0003 ± 0 . 0001 0 . 0001 0 . 0004 ] 0 . 0269 0 . 0026 0 . 0024 ± 3 . 9870 × 10 4 4 . 5291 × 10 4 6 . 1674 × 10 4
Table 4. Calibration results for Primesense Carmine 1.09.
Table 4. Calibration results for Primesense Carmine 1.09.
BurrusBouguetHerrera
RGB CameraIR CameraRGB CameraIR CameraRGB CameraIR Camera
f x 540 . 84 580 . 04 540 . 58 ± 0 . 64 575 . 46 ± 0 . 68 541 . 67 ± 0 . 16 574 . 98 ± 0 . 23
f y 539 . 48 576 . 45 538 . 95 ± 0 . 62 573 . 98 ± 0 . 65 539 . 48 ± 0 . 16 570 . 58 ± 0 . 31
c x 318 . 38 307 318 . 62 ± 0 . 99 318 . 79 ± 1 . 05 316 . 87 ± 0 . 27 323 . 97 ± 0 . 23
c y 237 . 82 232 . 75 238 . 32 ± 0 . 86 245 . 13 ± 0 . 90 235 . 48 ± 0 . 24 227 . 71 ± 0 . 2
k 1 0 . 0512 0 . 0687 0 . 0232 ± 0 . 0023 0 . 0401 ± 0 . 0029 0 . 0578 ± 0 . 0015 0
k 2 0 . 2236 0 . 2196 0 . 0939 ± 0 . 0059 0 . 0304 ± 0 . 0061 0 . 2610 ± 0 . 0069 0
k 3 0 . 1785 0 . 4167 00 0 . 2430 ± 0 . 0098 0
p 1 0 . 0010 0 . 0007 0 . 0012 ± 0 . 00045 0 . 00011 ± 0 . 00044 0 . 0003 ± 0 . 0001 0
p 2 0 . 0009 0 . 004 0 . 00064 ± 0 . 00055 0 . 00014 ± 0 . 00054 0 . 0017 ± 0 . 0001 0
c 0 4 . 0054 ± 0 . 0021
c 1 0 . 0029 ± 1 . 68 × 10 6
α 0 1 . 6229 ± 0 . 0304
α 1 0 . 0021 ± 4 . 06 × 10 5
R 0 . 9999 0 . 0049 0 . 0089 0 . 005 9 . 9992 0 . 005 0 . 0089 0 . 0112 0 . 9989 0 . 00214 0 . 00201 0 . 00429 ± 0 . 00089 0 . 00121 0 . 0001 1 0 . 0040 0 . 0086 0 . 0042 0 . 9998 0 . 0169 0 . 0086 0 . 0169 0 . 9998 ± 4 . 3043 × 10 4 4 . 6892 × 10 4 2 . 1396 × 10 4
T 0 . 0257 0 . 0005 0 . 0037 0 . 0262 0 . 0001 0 . 0002 ± 0 . 00005 0 . 00005 0 . 00021 ] 0 . 0265 0 . 0007 0 . 0030 ± 1 . 7632 × 10 4 1 . 3328 × 10 4 2 . 0493 × 10 4
Table 5. Calibration results for Microsoft Kinect V2.
Table 5. Calibration results for Microsoft Kinect V2.
BurrusBouguet
RGB CameraIR CameraRGB CameraIR Camera
f x 1669 . 54 364 . 92 1057 . 58 ± 1 . 83 369 . 15 ± 0 . 66
f y 1588 . 27 364 . 29 1055 . 33 ± 1 . 77 368 . 01 ± 0 . 64
c x 345 . 85 256 . 67 971 . 26 ± 1 . 71 260 . 68 ± 0 . 53
c y 251 . 59 205 . 33 538 . 00 ± 1 . 68 205 . 92 ± 0 . 60
k 1 0 . 0343 0 . 0934 0 . 0413 ± 0 . 0023 0 . 0631 ± 0 . 0034
k 2 0 . 0697 0 . 2748 0 . 0389 ± 0 . 0019 0 . 1758 ± 0 . 0044
k 3 0 . 0257 0 . 0963 00
p 1 0 . 0209 0 . 0004 0 . 00107 ± 0 . 00044 0 . 00096 ± 0 . 00037
p 2 0 . 0518 0 . 00004 0 . 00001 ± 0 . 00049 0 . 00062 ± 0 . 00032
R 0 . 9261 0 . 0471 0 . 3740 0 . 0258 0 . 9818 0 . 1879 0 . 3761 0 . 1837 0 . 9081 0 . 00156 0 . 00402 0 . 00691 ± 0 . 00087 0 . 00112 0 . 00013
T 0 . 0468 0 . 0080 0 . 3432 0 . 05211 0 . 00061 0 . 00319 ± 0 . 00011 0 . 00011 0 . 00040

Share and Cite

MDPI and ACS Style

Villena-Martínez, V.; Fuster-Guilló, A.; Azorín-López, J.; Saval-Calvo, M.; Mora-Pascual, J.; Garcia-Rodriguez, J.; Garcia-Garcia, A. A Quantitative Comparison of Calibration Methods for RGB-D Sensors Using Different Technologies. Sensors 2017, 17, 243. https://doi.org/10.3390/s17020243

AMA Style

Villena-Martínez V, Fuster-Guilló A, Azorín-López J, Saval-Calvo M, Mora-Pascual J, Garcia-Rodriguez J, Garcia-Garcia A. A Quantitative Comparison of Calibration Methods for RGB-D Sensors Using Different Technologies. Sensors. 2017; 17(2):243. https://doi.org/10.3390/s17020243

Chicago/Turabian Style

Villena-Martínez, Víctor, Andrés Fuster-Guilló, Jorge Azorín-López, Marcelo Saval-Calvo, Jeronimo Mora-Pascual, Jose Garcia-Rodriguez, and Alberto Garcia-Garcia. 2017. "A Quantitative Comparison of Calibration Methods for RGB-D Sensors Using Different Technologies" Sensors 17, no. 2: 243. https://doi.org/10.3390/s17020243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop