skip to main content
10.1145/3204949.3204969acmconferencesArticle/Chapter ViewAbstractPublication PagesmmsysConference Proceedingsconference-collections
research-article
Public Access

Skeleton-based continuous extrinsic calibration of multiple RGB-D kinect cameras

Published: 12 June 2018 Publication History

Abstract

Applications involving 3D scanning and reconstruction & 3D Tele-immersion provide an immersive experience by capturing a scene using multiple RGB-D cameras, such as Kinect. Prior knowledge of intrinsic calibration of each of the cameras, and extrinsic calibration between cameras, is essential to reconstruct the captured data. The intrinsic calibration for a given camera rarely ever changes, so only needs to be estimated once. However, the extrinsic calibration between cameras can change, even with a small nudge to the camera. Calibration accuracy depends on sensor noise, features used, sampling method, etc., resulting in the need for iterative calibration to achieve good calibration.
In this paper, we introduce a skeleton based approach to calibrate multiple RGB-D Kinect cameras in a closed setup, automatically without any intervention, within a few seconds. The method uses only the person present in the scene to calibrate, removing the need for manually inserting, detecting and extracting other objects like plane, checker-board, sphere, etc. 3D joints of the extracted skeleton are used as correspondence points between cameras, after undergoing accuracy and orientation checks. Temporal, spatial, and motion constraints are applied during the point selection strategy. Our calibration error checking is inexpensive in terms of computational cost and time and hence is continuously run in the background. Automatic re-calibration of the cameras can be performed when the calibration error goes beyond a threshold due to any possible camera movement. Evaluations show that the method can provide fast, accurate and continuous calibration, as long as a human is moving around in the captured scene.

References

[1]
D. Alexiadis, D. Zarpalas, and P. Daras. Fast and smooth 3d reconstruction using multiple rgb-depth sensors. In Visual Communications and Image Processing Conference, 2014 IEEE, pages 173--176, Dec 2014.
[2]
K. S. Arun, T. S. Huang, and S. D. Blostein. Least-squares fitting of two 3-d point sets. IEEE Trans. Pattern Anal. Mach. Intell., 9(5):698--700, May 1987.
[3]
E. Auvinet, J. Meunier, and F. Multon. Multiple depth cameras calibration and body volume reconstruction for gait analysis. In Information Science, Signal Processing and their Applications (ISSPA), 2012 11th International Conference on, pages 478--483, July 2012.
[4]
S. Beck and B. Froehlich. Volumetric calibration and registration of multiple rgbd-sensors into a joint coordinate system. In 3D User Interfaces (3DUI), 2015 IEEE Symposium on, pages 89--96. IEEE, 2015.
[5]
M. Carraro, M. Munaro, A. Roitberg, and E. Menegatti. Improved Skeleton Estimation by Means of Depth Data Fusion from Multiple Depth Cameras, pages 1155--1167. Springer International Publishing, Cham, 2017.
[6]
C. H. Chen, J. Favre, G. Kurillo, T. P. Andriacchi, R. Bajcsy, and R. Chellappa. Camera networks for healthcare, teleimmersion, and surveillance. Computer, 47(5):26--36, May 2014.
[7]
S. Chen, Z. Gao, K. Nahrstedt, and I. Gupta. 3dti amphitheater: Towards 3dti broadcasting. ACM Trans. Multimedia Comput. Commun. Appl., 11(2s):47:1--47:22, Feb. 2015.
[8]
L. Cruz, D. Lucio, and L. Velho. Kinect and rgbd images: Challenges and applications. In Graphics, Patterns and Images Tutorials (SIBGRAPI-T), 2012 25th SIBGRAPI Conference on, pages 36--49, Aug 2012.
[9]
K. Desai, K. Bahirat, S. Ramalingam, B. Prabhakaran, T. Annaswamy, and U. E. Makris. Augmented reality-based exergames for rehabilitation. In Proceedings of the 7th International Conference on Multimedia Systems, MMSys '16, pages 22:1--22:10, 2016.
[10]
K. Desai, U. H. H. Belmonte, R. Jin, B. Prabhakaran, P. Diehl, V. A. Ramirez, V. Johnson, and M. Gans. Experiences with multi-modal collaborative virtual laboratory (mmcvl). In 2017 IEEE Third International Conference on Multimedia Big Data (BigMM), pages 376--383, April 2017.
[11]
K. Desai, S. Raghuraman, R. Jin, and B. Prabhakaran. Qoe studies on interactive 3d tele-immersion. In 2017 IEEE International Symposium on Multimedia (ISM), Dec 2017.
[12]
D. Dwarakanath, C. Griwodz, P. Halvorsen, and J. Lildballe. Study the effects of camera misalignment on 3d measurements for efficient design of vision-based inspection systems. In A. Likas, K. Blekas, and D. Kalles, editors, Artificial Intelligence: Methods and Applications, pages 150--163, Cham, 2014. Springer International Publishing.
[13]
T. Elgamal and K. Nahrstedt. Multicamera summarization of rehabilitation sessions in home environment. In Proceedings of the 2017 ACM on Multimedia Conference, MM '17, pages 1381--1389, 2017.
[14]
Y. Furukawa and J. Ponce. Accurate camera calibration from multi-view stereo and bundle adjustment. International Journal of Computer Vision, 84(3):257--268, Sep 2009.
[15]
D. Herrera, J. Kannala, and J. Heikkilä. Joint depth and color camera calibration with distortion correction. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(10):2058--2064, 2012.
[16]
B. K. P. Horn, H. Hilden, and S. Negahdaripour. Closed-form solution of absolute orientation using orthonormal matrices. JOURNAL OF THE OPTICAL SOCIETY AMERICA, 5(7):1127--1135, 1988.
[17]
M. Kowalski, J. Naruniec, and M. Daniluk. Livescan3d: A fast and inexpensive 3d data acquisition system for multiple kinect v2 sensors. In 3D Vision (3DV), 2015 International Conference on, pages 318--325, Oct 2015.
[18]
G. Kurillo and R. Bajcsy. 3d teleimmersion for collaboration and interaction of geographically distributed users. Virtual Reality, 17(1):29--43, 2013.
[19]
B. Li, L. Heng, K. Koser, and M. Pollefeys. A multiple-camera system calibration toolbox using a feature descriptor-based calibration pattern. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 1301--1307. IEEE, 2013.
[20]
M. A. Livingston, J. Sebastian, Z. Ai, and J. W. Decker. Performance measurements for the microsoft kinect skeleton. In 2012 IEEE Virtual Reality Workshops (VRW), pages 119--120, March 2012.
[21]
M. I. A. Lourakis and A. A. Argyros. Sba: A software package for generic sparse bundle adjustment. ACM Trans. Math. Softw., 36(1):2:1--2:30, Mar. 2009.
[22]
A. Maimone, J. Bidwell, K. Peng, and H. Fuchs. Enhanced personal autostereoscopic telepresence system using commodity depth cameras. Computers & Graphics, 36(7):791--807, 2012.
[23]
R. Mekuria, M. Sanna, E. Izquierdo, D. Bulterman, and P. Cesar. Enabling geometry-based 3-d tele-immersion with fast mesh compression and linear rate-less coding. Multimedia, IEEE Transactions on, 16(7):1809--1820, Nov 2014.
[24]
K. A. F. Mora and J. M. Odobez. Gaze estimation from multimodal kinect data. In 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, pages 25--30, June 2012.
[25]
M. Munaro, F. Basso, and E. Menegatti. Openptrack: Open source multi-camera calibration and people tracking for rgb-d camera networks. Robotics and Autonomous Systems, 75(Part B):525 -- 538, 2016.
[26]
J. C. Núñez, R. Cabido, A. S. Montemayor, and J. J. Pantrigo. Real-time human body tracking based on data fusion from multiple rgb-d sensors. Multimedia Tools Appl., 76(3):4249--4271, Feb. 2017.
[27]
S. Obdrzálek, G. Kurillo, F. Ofli, R. Bajcsy, E. Seto, H. Jimison, and M. Pavel. Accuracy and robustness of kinect pose estimation in the context of coaching of elderly population. In 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pages 1188--1193, Aug 2012.
[28]
M. Ruan and D. Huber. Calibration of 3d sensors using a spherical target. In 2014 2nd International Conference on 3D Vision, volume 1, pages 187--193, Dec 2014.
[29]
T. Schöps, J. L. Schönberger, S. Galliani, T. Sattler, K. Schindler, M. Pollefeys, and A. Geiger. A multi-view stereo benchmark with high-resolution images and multi-camera videos. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2538--2547, July 2017.
[30]
J. Shotton, A. Fitzgibbon, M. Cook, T. Sharp, M. Finocchio, R. Moore, A. Kipman, and A. Blake. Real-time human pose recognition in parts from single depth images. In R. Cipolla, S. Battiato, and G. M. Farinella, editors, Machine Learning for Computer Vision, volume 411 of Studies in Computational Intelligence, pages 119--135. Springer Berlin Heidelberg, 2013.
[31]
A. Staranowicz, G. R. Brown, F. Morbidi, and G. L. Mariottini. Easy-to-use and accurate calibration of rgb-d cameras from spheres. In Pacific-Rim Symposium on Image and Video Technology, pages 265--278. Springer, 2013.
[32]
C. Strecha, W. von Hansen, L. V. Gool, P. Fua, and U. Thoennessen. On bench-marking camera calibration and multi-view stereo for high resolution imagery. In 2008 IEEE Conference on Computer Vision and Pattern Recognition, pages 1--8, June 2008.
[33]
P. C. Su, J. Shen, and M. U. Rafique. Rgb-d camera network calibration and streaming for 3d telepresence in large environment. In 2017 IEEE Third International Conference on Multimedia Big Data (BigMM), pages 362--369, April 2017.
[34]
S. Umeyama. Least-squares estimation of transformation parameters between two point patterns. IEEE Trans. Pattern Anal. Mach. Intell., 13(4):376--380, Apr. 1991.
[35]
L. Yang, L. Zhang, H. Dong, A. Alelaiwi, and A. El Saddik. Evaluating and improving the depth accuracy of kinect for windows v2. Sensors Journal, IEEE, 15(8):4275--4285, Aug 2015.
[36]
Z. Zhang. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell., 22(11):1330--1334, Nov. 2000.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MMSys '18: Proceedings of the 9th ACM Multimedia Systems Conference
June 2018
604 pages
ISBN:9781450351928
DOI:10.1145/3204949
  • General Chair:
  • Pablo Cesar,
  • Program Chairs:
  • Michael Zink,
  • Niall Murray
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 June 2018

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. 3D calibration
  2. 3D skeleton & point-cloud
  3. interactive 3D tele-immersion
  4. real-time

Qualifiers

  • Research-article

Funding Sources

Conference

MMSys '18
Sponsor:
MMSys '18: 9th ACM Multimedia Systems Conference
June 12 - 15, 2018
Amsterdam, Netherlands

Acceptance Rates

Overall Acceptance Rate 176 of 530 submissions, 33%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)171
  • Downloads (Last 6 weeks)34
Reflects downloads up to 25 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media