Extrinsic Calibration of Camera Networks Based on Pedestrians
Abstract
:1. Introduction
2. Related Work
3. Preliminaries
3.1. Camera Model
3.2. Notations and Architecture
4. 3D Head and Feet Positions in Local Camera Coordinates
4.1. Extract Image Positions of Head and Feet
- Ellipse fitting: We fit an ellipse to the blob of the person. The end points of the major axis of the ellipse are then taken as and .
- Bounding box and line fitting: We fit a line and a bounding box to a person’s blob. The intersections between the line and the bounding box of the blob are taken as and .
4.2. Estimate 3D Positions Based on Image Positions
5. Multi-Camera Calibration
5.1. Pairwise Calibration Based on Orthogonal Procrustes
- The points are not coplanar. In this case, the rotation matrix is uniquely found and calculated by .
- The points are coplanar, but not collinear. In this case, the rotation matrix is calculated using:
- The points are collinear. In this case, cannot be uniquely found, which is the failure case of our method.
5.2. Robust Calibration Using RANSAC
- Select three pairs of 3D points randomly and compute extrinsic parameters using the method from Section 5.1.
- Count the number of pairs agreeing with the extrinsic parameters (inliers). A pair is considered to agree with the extrinsic parameters if for some threshold ϵ:
- Repeat Steps 1 and 2 until the number of inliers reaches a certain threshold.
- Re-compute extrinsic parameters using all of the inliers based on the method from Section 5.1.
5.3. Refinement through Gradient Descent
5.4. Alignment with a World Coordinate System
6. Performance Measures
6.1. Measures without Ground Truth
- . The triangulation error is a measure of how the calibration matrices influence the accuracy of multi-camera triangulation. Let be the ground truth position of the i-th test sample and its position estimated using the triangulation method of [19], which takes as input the estimated extrinsic parameters and image positions of the i-th test markers. This is also the classical method for 3D reconstruction: it represents how well we can measure the 3D world with estimated extrinsic parameters. The discrepancy between real and estimated positions is compared. The error is expressed in physical units (e.g., centimeter) and is defined as:
- . The projection error is a measure of how the calibration matrices influence the accuracy of projections of 3D points on image planes. Let be the observed pixel coordinates of the i-th sample in the k-th camera’s image, while is the estimated position through projection. Accuracy is obtained by measuring the discrepancy between the real 2D points (obtained from image segmentation) and the estimated ones (obtained by using the camera model). The error is expressed in and is defined as:
- . The reprojection error is used to quantify how closely we can recreate the point’s true projection with an estimate of a 3D point . Different from the projection error, the 3D points are firstly obtained from triangulation based on estimated extrinsic parameters and image points. Then, image feature points are projected from these 3D points. The discrepancy between the real 2D points (obtained from image segmentation) and the estimated ones (obtained through reprojection) is computed. The error is expressed in pixels and is defined as:
6.2. Measures with Ground Truth
- . The relative translation error is used to quantify how closely we can estimate the distance between the camera center and the origin of the world coordinate system. We get for the k-th camera from classical methods [1]. We also estimate translation with the proposed calibration method. Cameras are usually mounted high above the ground plane for getting a good view of the scene and increasing the viewing area. The origin of the world coordinate system mostly lies on the ground plane. Therefore, the distance between a camera center and the world origin is usually large (at least 200 cm). Thus, we propose to calculate relative translation error by:
- . The rotation error is used to quantify how accurately we can estimate the orientation of all cameras. We get three ground truth angles () for the k-th camera From classical methods [1]. We also estimate the three angles () with the proposed calibration method. Then, we calculate rotation error by:
7. Experiments and Results
7.1. Calibration when the Person Does not Walk along a Straight Line
7.1.1. Comparison Using Different Feet and Head Detections
7.1.2. Comparison Using Different Numbers of Locations
7.2. Calibration When the Person Runs along a Straight Line
7.3. Calibration with Multiple Pedestrians
8. Conclusions
Acknowledgments
Author Contributions
Conflicts of Interest
Abbreviations
2D | Two-dimensional |
3D | Three-dimensional |
VSS | Visual surveillance system |
DLT | Direct linear transformation |
RANSAC | Random sample consensus |
SVD | Singular value decomposition |
References
- Tsai, R.Y. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE J. Robot. Autom. 1987, 3, 323–344. [Google Scholar] [CrossRef]
- Faugeras, O.D.; Toscani, G. The calibration problem for stereo. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 22–26 June 1986; pp. 15–20.
- Hall, E.L.; Tio, J.; McPherson, C.; Sadjadi, F. Measuring Curved Surfaces for Robot Vision. Computer 1982, 15, 42–54. [Google Scholar] [CrossRef]
- Weng, J.; Cohen, P.; Herniou, M. Camera calibration with distortion models and accuracy evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 965–980. [Google Scholar] [CrossRef]
- Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
- Svoboda, T.; Martinec, D.; Pajdla, T. A Convenient Multi-Camera Self-Calibration for Virtual Environments. PRESENCE Teleoper. Virtual Environ. 2005, 14, 407–422. [Google Scholar] [CrossRef]
- Chen, X.; Davis, J.; Slusallek, P. Wide area camera calibration using virtual calibration objects. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head Island, SC, USA, 13–15 June 2000; Volume 2, pp. 520–527.
- Kurillo, G.; Li, Z.; Bajcsy, R. Wide-area external multi-camera calibration using vision graphs and virtual calibration object. In Proceedings of the ACM/IEEE International Conference on Distributed Smart Cameras, Stanford, CA, USA, 7–11 September 2008; pp. 1–9.
- Sturm, P.; Triggs, B. A Factorization Based Algorithm for Multi-Image Projective Structure and Motion. In Proceedings of the European Conference on Computer Vision, Cambridge, UK, 14–18 April 1996; Volume II, pp. 709–720.
- Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: New York, NY, USA, 2003. [Google Scholar]
- Pollefeys, M.; Koch, R.; Van Gool, L. Self-calibration and metric reconstruction in spite of varying and unknown internal camera parameters. In Proceedings of the International Conference on Computer Vision, Bombay, India, 4–7 January 1998; pp. 90–95.
- Triggs, B.; McLauchlan, P.; Hartley, R.; Fitzgibbon, A. Bundle Adjustment-A Modern Synthesis. In Vision Algorithms: Theory and Practice; Triggs, B., Zisserman, A., Szeliski, R., Eds.; Springer: Berlin/Heidelberg, Germany, 2000; Volume 1883, pp. 298–372. [Google Scholar]
- Shen, J.; Xu, W.; Luo, Y.; Su, P.C.; Cheung, S. Extrinsic calibration for wide-baseline RGB-D camera network. In Proceedings of the IEEE International Workshop on Multimedia Signal Processing, Jakarta, Indonesia, 22–24 September 2014; pp. 1–6.
- Guan, J.; Deboeverie, F.; Slembrouck, M.; van Haerenborgh, D.; van Cauwelaert, D.; Veelaert, P.; Philips, W. Extrinsic Calibration of Camera Networks Using a Sphere. Sensors 2015, 15, 18985–19003. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Arun, K.; Huang, T.; Blostein, S. Least-Squares Fitting of Two 3-D Point Sets. IEEE Trans. Pattern Anal. Mach. Intell. 1987, 9, 698–700. [Google Scholar] [CrossRef] [PubMed]
- Bouguet, J.Y. Camera calibration toolbox for Matlab. Available online: http://www.vision.caltech.edu/bouguetj/calib_doc/ (accessed on 19 February 2015).
- Hödlmoser, M.; Kampel, M. Multiple Camera Self-calibration and 3D Reconstruction Using Pedestrians. In Proceedings of the International Conference on Advances in Visual Computing, Las Vegas, NE, USA, 29 November–1 December 2010; pp. 1–10.
- Longuet-Higgins, H.C. A computer algorithm for reconstructing a scene from two projections. Nature 1981, 293, 133–135. [Google Scholar] [CrossRef]
- Hartley, R.; Sturm, P. Triangulation. In Computer Analysis of Images and Patterns; Hlaváč, V., Šára, R., Eds.; Springer: Berlin/Heidelberg, Germany, 1995; Volume 970, pp. 190–197. [Google Scholar]
- Lv, F.; Zhao, T.; Nevatia, R. Camera calibration from video of a walking human. IEEE Trans. Pattern Anal. Machine Intell. 2006, 28, 1513–1518. [Google Scholar]
- Krahnstoever, N.; Mendonca, P.R.S. Bayesian autocalibration for surveillance. In Proceedings of the IEEE International Conference on Computer Vision, Beijing, China, 17–21 October 2005; Volume 2, pp. 1858–1865.
- Krahnstoever, N.; Mendonca, P.R.S. Autocalibration from Tracks of Walking People. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 4–7 September 2006; pp. 107–116.
- Liu, J.; Collins, R.T.; Liu, Y. Surveillance camera autocalibration based on pedestrian height distributions. In Proceedings of the British Machine Vision Conference, Dundee, UK, 29 August–2 September 2011.
- Liu, J.; Collins, R.T.; Liu, Y. Robust autocalibration for a surveillance camera network. In Proceedings of the IEEE Workshop on Applications of Computer Vision, Tampa, FL, USA, 17–18 January 2013; pp. 433–440.
- Caprile, B.; Torre, V. Using Vanishing Points for Camera Calibration. Int. J. Comput. Vis. 1990, 4, 127–140. [Google Scholar] [CrossRef]
- Gardel, A.; García, J.; Bravo, I.; Espinosa, F.; Chateau, T. Camera calibration parameters for oriented person re-identification. In Proceedings of the International Conference on Distributed Smart Camera, Seville, Spain, 8–11 September 2015; pp. 110–115.
- Teixeira, L.; Maffra, F.; Badii, A. Scene Understanding for Auto-Calibration of Surveillance Cameras. In Proceedings of the Advances in Visual Computing: International Symposium on Visual Computing, Las Vegas, NV, USA, 8–10 December 2014; pp. 671–682.
- Schels, J.; Liebelt, J.; Lienhart, R. Self-calibrating 3D context for retrieving people with luggage. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Barcelona, Spain, 6–13 November 2011; pp. 1920–1927.
- Micusik, B.; Pajdla, T. Simultaneous surveillance camera calibration and foot-head homology estimation from human detections. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1562–1569.
- Possegger, H.; Rüther, M.; Sternig, S.; Mauthner, T.; Klopschitz, M.; Roth, P.M.; Bischof, H. Unsupervised Calibration of Camera Networks and Virtual PTZ Cameras. In Proceedings of the Computer Vision Winter Workshop, Prlekija, Slovenia, 1–3 February 2012.
- Hartley, R.I. Chirality. Int. J. Comput. Vis. 1998, 26, 41–61. [Google Scholar] [CrossRef]
- Werner, T.; Pajdla, T. Cheirality in epipolar geometry. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 1, pp. 548–553.
- Weng, J.; Ahuja, N.; Huang, T.S. Motion and structure from point correspondences with error estimation: Planar surfaces. IEEE Trans. Sign. Process. 1991, 39, 2691–2717. [Google Scholar] [CrossRef]
- Faugeras, O.D.; Lustman, F. Motion and structure from motion in a piecewise planar environment. Int. J. Pattern Recogn. Artif. Intell. 1988, 2, 485–508. [Google Scholar] [CrossRef]
- Eggert, D.W.; Lorusso, A.; Fisher, R.B. Estimating 3-D Rigid Body Transformations: A Comparison of Four Major Algorithms. Mach. Vision Appl. 1997, 9, 272–290. [Google Scholar] [CrossRef]
- Tardif, J.P.; Sturm, P.; Roy, S. Self-calibration of a General Radially Symmetric Distortion Model. In Proceedings of the European Conference on Computer Vision, Leonardis, Graz, Austria, 7–13 May 2006; pp. 186–199.
- Houben, S. Towards the intrinsic self-calibration of a vehicle-mounted omni-directional radially symmetric camera. In Proceedings of the IEEE Intelligent Vehicles Symposium Proceedings, Dearborn, MI, USA, 8–11 June 2014; pp. 878–883.
- Zivkovic, Z. Improved adaptive Gaussian mixture model for background subtraction. In Proceedings of the IEEE International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; Volume 2, pp. 28–31.
- Walker, M.W.; Shao, L.; Volz, R.A. Estimating 3-D Location Parameters Using Dual Number Quaternions. CVGIP Image Underst. 1991, 54, 358–367. [Google Scholar] [CrossRef]
- Horn, B.K.P.; Hilden, H.M.; Negahdaripour, S. Closed-form solution of absolute orientation using orthonormal matrices. J. Opt. Soc. Am. A 1988, 5, 1127–1135. [Google Scholar] [CrossRef]
- Horn, B.K.P. Closed-form solution of absolute orientation using unit quaternions. J. Opt. Soc. Am. A 1987, 4, 629–642. [Google Scholar] [CrossRef]
- Schönemann, P. A generalized solution of the orthogonal Procrustes problem. Psychometrika 1966, 31, 1–10. [Google Scholar] [CrossRef]
- Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
- Multi-camera pedestrians video. Available online: http://cvlab.epfl.ch/data/pom (accessed on 19 April 2016).
Variables | Definition |
---|---|
i, k | Index for frames and cameras, respectively |
, | Rotation and translation parameters for camera k |
h | Height of the person |
, , , | Image position, normalized image position, Z coordinate and 3D position of the head of a person, respectively, for camera k |
, , , | Image position, normalized image position, Z coordinate and 3D position of the feet of a person, respectively, for camera k |
Cross product of and for camera k | |
Unit vector of the center line of a person w.r.t camera k | |
Centroid of all 3D positions of the head and feet w.r.t camera k |
Proposed-Refined | Proposed | Essential | Essential-Refined | |
---|---|---|---|---|
Triangulation error (cm) | 1.9/0.8 | 5.4/1.5 | 10.0/7.6 | 2.0/1.7 |
Projection error (pixel) | 4.6/0.6 | 8.2/2.2 | 33.2/20.9 | 4.8/1.8 |
Reprojection error (pixel) | 4.4/0.6 | 7.1/2.1 | 32.1/20.9 | 4.6/1.6 |
Rotation error () | 0.9/0.2 | 2.2/0.6 | 4.1/3.8 | 1.1/2.2 |
Relative translation error | 1.9%/0.8% | 6.6%/1.8% | 14.6%/9.9% | 2.6%/7.4% |
2 | 3 | 4 | 5 | 6 | 7 | |
---|---|---|---|---|---|---|
Essential | NA | NA | 35.3% | 76.1% | 93.6% | 97.9% |
Proposed | 63.3% | 91.8% | 97.8% | 99.7% | 99.9% | 100% |
Camera 1 | Camera 2 | Camera 3 | |
---|---|---|---|
Rotation error () | 2.0 | 5.1 | 3.1 |
Relative translation error | 5.7% | 1.5% | 0.8% |
© 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Guan, J.; Deboeverie, F.; Slembrouck, M.; Van Haerenborgh, D.; Van Cauwelaert, D.; Veelaert, P.; Philips, W. Extrinsic Calibration of Camera Networks Based on Pedestrians. Sensors 2016, 16, 654. https://doi.org/10.3390/s16050654
Guan J, Deboeverie F, Slembrouck M, Van Haerenborgh D, Van Cauwelaert D, Veelaert P, Philips W. Extrinsic Calibration of Camera Networks Based on Pedestrians. Sensors. 2016; 16(5):654. https://doi.org/10.3390/s16050654
Chicago/Turabian StyleGuan, Junzhi, Francis Deboeverie, Maarten Slembrouck, Dirk Van Haerenborgh, Dimitri Van Cauwelaert, Peter Veelaert, and Wilfried Philips. 2016. "Extrinsic Calibration of Camera Networks Based on Pedestrians" Sensors 16, no. 5: 654. https://doi.org/10.3390/s16050654
APA StyleGuan, J., Deboeverie, F., Slembrouck, M., Van Haerenborgh, D., Van Cauwelaert, D., Veelaert, P., & Philips, W. (2016). Extrinsic Calibration of Camera Networks Based on Pedestrians. Sensors, 16(5), 654. https://doi.org/10.3390/s16050654