skip to main content
10.1145/1873951.1874161acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
short-paper

Refining video annotation by exploiting inter-shot context

Published: 25 October 2010 Publication History

Abstract

This paper proposes a new approach to refine video annotation by exploiting the inter-shot context. Our method is mainly novel in two ways. On one hand, to refine annotation result of the target concept, we model the sequence of shots in video as a conditional random field with chain structure. In this way, we can capture different kinds of concept relationships in inter-shot context to improve the annotation accuracy. On the other hand, to exploit inter-shot context for the target concept, we classify shots into different types according to their correlation to the target concept, which will be used to represent different kinds of concept relationships in inter-shot context. Experiments on the widely used TRECVID 2006 data set show that our method is effective for refining video annotation, achieving a significant performance improvement over several state of the art methods.

References

[1]
Y. G. Jiang, C. W. Ngo, and J. Yang. Towards optimal bag-of-features for object categorization and semantic video retrieval. CIVR, 2007.
[2]
J. Lafferty, A. McCallum, and F. Pereira. Conditional random fields: probabilistic models for segmenting and labeling sequence data. ICML, 2001.
[3]
K. H. Liu, M. F. Weng, and et al. Association and temporal rule mining for post-processing of semantic concept detection in video. IEEE Trans. on MM, 10:240--251, 2008.
[4]
G. J. Qi, X. S. Hua, and et al. Correlative multi-label video annotation with temporal kernels. TOMCCAP, 2008.
[5]
A. F. Smeaton, P. Over, and W. Kraaij. Evaluation campaigns and trecvid. MIR, 2006.
[6]
C. Sutton and et al. An introduction to conditional random fields for relational learning. In L.Getoor and B. Taskar (Eds.), Introduction to statistical relational learning. MIT Press, 2007.
[7]
M. F. Weng and Y. Y. Chuang. Multi-cue fusion for semantic video indexing. ACM MM, 2008.
[8]
A. Yanagawa and et al. Columbia university's baseline detectors for 374 lscom semantic visual concepts. Technical report, Columbia University, 2007.
[9]
E. Yilmaz and J. A. Aslam. Estimating average precision with incomplete and imperfect judgments. CIKM, 2006.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MM '10: Proceedings of the 18th ACM international conference on Multimedia
October 2010
1836 pages
ISBN:9781605589336
DOI:10.1145/1873951
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 October 2010

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. conditional random field
  2. inter-shot context
  3. video annotation

Qualifiers

  • Short-paper

Conference

MM '10
Sponsor:
MM '10: ACM Multimedia Conference
October 25 - 29, 2010
Firenze, Italy

Acceptance Rates

Overall Acceptance Rate 995 of 4,171 submissions, 24%

Upcoming Conference

MM '24
The 32nd ACM International Conference on Multimedia
October 28 - November 1, 2024
Melbourne , VIC , Australia

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 14 Sep 2024

Other Metrics

Citations

Cited By

View all

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media