skip to main content
10.1145/2304496.2304502acmotherconferencesArticle/Chapter ViewAbstractPublication PagesvigtaConference Proceedingsconference-collections
research-article

A semi-automatic tool for detection and tracking ground truth generation in videos

Published: 21 May 2012 Publication History

Abstract

In this paper we present a tool for the generation of ground-truth data for object detection, tracking and recognition applications. Compared to state of the art methods, such as ViPER-GT, our tool improves the user experience by providing edit shortcuts such as hotkeys and drag-and-drop, and by integrating computer vision algorithms to automate, under the supervision of the user, the extraction of contours and the identification of objects across frames. A comparison between our application and ViPER-GT tool was performed, which showed how our tool allows users to label a video in a shorter time, while at the same time providing a higher ground truth quality.

References

[1]
M.-Y. Liao, D.-Y. Chen, C.-W. Sua, and H.-R. Tyan, "Real-time event detection and its application to surveillance systems," in Circuits and Systems, 2006. ISCAS 2006. Proceedings. 2006 IEEE International Symposium on, 2006.
[2]
A. Faro, D. Giordano, and C. Spampinato, "Soft-computing agents processing webcam images to optimize metropolitan traffic systems," in Computer Vision and Graphics, ser. Computational Imaging and Vision, K. Wojciechowski, B. Smolka, H. Palus, R. Kozera, W. Skarbek, and L. Noakes, Eds. Springer Netherlands, 2006, vol. 32, pp. 968--974.
[3]
C. Spampinato, J. Chen-Burger, G. Nadarajan, and R. Fisher, "Detecting, tracking and counting fish in low quality unconstrained underwater videos," in Proc. 3rd Int. Conf. on Computer Vision Theory and Applications (VISAPP), 2008, pp. 514--520.
[4]
A. Faro, D. Giordano, and C. Spampinato, "Adaptive background modeling integrated with luminosity sensors and occlusion processing for reliable vehicle detection," Intelligent Transportation Systems, IEEE Transactions on, vol. 12, no. 4, pp. 1398--1412, dec. 2011.
[5]
C. Spampinato, S. Palazzo, D. Giordano, I. Kavasidis, F. Lin, and Y. Lin, "Covariance based fish tracking in real-life underwater environment," in International Conference on Computer Vision Theory and Applications, VISAPP 2012, 2012, pp. 409--414.
[6]
C. Spampinato, D. Giordano, R. D. Salvo, Y. J. Chen-Burger, R. B. Fisher, and G. Nadarajan, "Automatic fish classification for underwater species behavior understanding," in First ACM International Workshop on Analysis and Retrieval of Tracked Events and Motion in Imagery Streams, 2010, pp. 45--50.
[7]
D. Doerman and D. Mihalcik, "Tools and techniques for video performance evaluation," in Pattern Recognition, 2000. Proceedings. 15th International Conference on, vol. 4, 2000, pp. 167--170.
[8]
T. D'Orazio, M. Leo, N. Mosca, P. Spagnolo, and P. L. Mazzeo, "A semi-automatic system for ground truth generation of soccer video sequences," in Advanced Video and Signal Based Surveillance, 2009. AVSS '09. Sixth IEEE International Conference on, Genova, 2009, pp. 559--564.
[9]
A. Ambardekar and M. Nicolescu, "Ground Truth Verification Tool (GTVT) for Video Surveillance Systems," in Advances in Computer-Human Interactions, 2009. ACHI '09. Second International Conferences on, Cancun, 2009, pp. 354--359.
[10]
C. Lin and B. Tseng, "Video collaborative annotation forum: Establishing ground-truth labels on large multimedia datasets," Proceedings of the TRECVID 2003, 2003.
[11]
M. Kass, A. Witkin, and D. Terzopoulos, "Snakes: Active contour models," International Journal of Computer Vision, vol. 1, no. 4, pp. 321--331, Jan. 1988.
[12]
C. Rother, V. Kolmogorov, and A. Blake, ""GrabCut": interactive foreground extraction using iterated graph cuts," ACM Trans. Graph., vol. 23, no. 3, pp. 309--314, 2004.
[13]
C. Stauffer and W. E. L. Grimson, "Adaptive background mixture models for real-time tracking," Computer Vision and Pattern Recognition, IEEE Computer Society Conference on, vol. 2, pp. 246--252, 1999.
[14]
G. R. Bradski, "Computer Vision Face Tracking For Use in a Perceptual User Interface," Intel Technology Journal, pp. 1--15, 1998.
[15]
J. P. Chin, V. A. Diehl, and K. L. Norman, "Development of an instrument measuring user satisfaction of the human-computer interface," in Proceedings of the SIGCHI conference on Human factors in computing systems, ser. CHI '88. New York, NY, USA: ACM, 1988, pp. 213--218.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
VIGTA '12: Proceedings of the 1st International Workshop on Visual Interfaces for Ground Truth Collection in Computer Vision Applications
May 2012
74 pages
ISBN:9781450314053
DOI:10.1145/2304496
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 21 May 2012

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. ground truth data
  2. object detection
  3. object tracking
  4. video labeling

Qualifiers

  • Research-article

Funding Sources

Conference

VIGTA '12

Acceptance Rates

Overall Acceptance Rate 8 of 15 submissions, 53%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)1
Reflects downloads up to 26 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media