skip to main content
10.1145/3582197.3582210acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicitConference Proceedingsconference-collections
research-article

Video Inter-frame Tampering Detection Based on SN-VGG+BiLSTM-AE Composite Model

Published: 30 March 2023 Publication History

Abstract

Video inter-frame tampering detection is the most common type of forensics in video forensics. The traditional detection method is to detect tampering by extracting digital image features of video frames, such as SIFT, HOG, and ORB. The accuracy of frame discrimination and localization is limited. This paper introduces deep learning into the problem of tampering detection, and proposes a composite network model structure using the Siamese network Siamese and the bidirectional long short-term memory network autoencoder BiLSTM AutoEncoder to detect tampered frames. Among them, Siamese calculates the inter-frame distance by calculating the depth features of the frames extracted by VGG-16, and inputs them into BiLSTM AutoEncoder for frame sequence anomaly detection and localization. The model is experimented on two different datasets with good results, validating the model generalization performance. Compared with the classical method, this model obtains higher precision(93.7%) of tamper points, which verifies the superiority of this deep learning model.

References

[1]
W. Wang and H. Farid, “Exposing digital forgeries in video by detecting double MPEG compression,” in Proceedings of the 8th workshop on Multimedia and security, 2006, pp. 37–47.
[2]
J. A. Aghamaleki and A. Behrad, “Inter-frame video forgery detection and localization using intrinsic effects of double compression on quantization errors of video coding,” Signal Process. Image Commun., vol. 47, pp. 289–302, 2016.
[3]
S. Fadl, Q. Han, and L. Qiong, “Exposing video inter-frame forgery via histogram of oriented gradients and motion energy image,” Multidimens. Syst. Signal Process., vol. 31, no. 4, pp. 1365–1384, 2020.
[4]
W. Wang, X. Jiang, S. Wang, M. Wan, and T. Sun, “Identifying video forgery process using optical flow,” in International Workshop on Digital Watermarking, 2013, pp. 244–257.
[5]
Y. Wu, X. Jiang, T. Sun, and W. Wang, “Exposing video inter-frame forgery based on velocity field consistency,” in 2014 IEEE international conference on acoustics, speech and signal processing (ICASSP), 2014, pp. 2674–2678.
[6]
T. Huang, Z. Chen, L. Su, Z. Zheng, and X. Yuan, “Digital video forgeries detection based on content continuity,” J. Nanjing Univ. Nat. Sci., vol. 47, no. 5, pp. 493–503, 2011.
[7]
Y. Liu and T. Huang, “Exposing video inter-frame forgery by Zernike opponent chromaticity moments and coarseness analysis,” Multimed. Syst., vol. 23, no. 2, pp. 223–238, 2017.
[8]
Z. Zhen-zhen, H. Jian-jun, L. Zhao-Hong, and G. Sheng, “Video-frame insertion and deletion detection based on consistency of quotients of MSSIM,” J. Beijing Univ. Posts Telecommun., vol. 38, no. 4, p. 86, 2015.
[9]
L. Zheng, T. Sun, and Y.-Q. Shi, “Inter-frame video forgery detection based on block-wise brightness variance descriptor,” in International workshop on digital watermarking, 2014, pp. 18–30.
[10]
R. Parmani, S. Butala, A. Khanvilkar, S. Pawar, and N. Pulgam, “Inter Frame Video Forgery Detection using Normalized Multi Scale One Level Subtraction,” in Proceedings of International Conference on Communication and Information Processing (ICCIP), 2019.
[11]
V. Kumar and M. Gaur, “Multiple forgery detection in video using inter-frame correlation distance with dual-threshold,” Multimed. Tools Appl., pp. 1–20, 2022.
[12]
K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” ArXiv Prepr. ArXiv14091556, 2014.
[13]
J. Bromley, I. Guyon, Y. LeCun, E. Säckinger, and R. Shah, “Signature verification using a" siamese" time delay neural network,” Adv. Neural Inf. Process. Syst., vol. 6, 1993.
[14]
S. Zagoruyko and N. Komodakis, “Learning to compare image patches via convolutional neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 4353–4361.
[15]
R. C. Staudemeyer and E. R. Morris, “Understanding LSTM–a tutorial into long short-term memory recurrent neural networks,” ArXiv Prepr. ArXiv190909586, 2019.
[16]
S. Siami-Namini, N. Tavakoli, and A. S. Namin, “The performance of LSTM and BiLSTM in forecasting time series,” in 2019 IEEE International Conference on Big Data (Big Data), 2019, pp. 3285–3292.
[17]
N. Srivastava, E. Mansimov, and R. Salakhudinov, “Unsupervised learning of video representations using lstms,” in International conference on machine learning, 2015, pp. 843–852.
[18]
J. Li, “Expansion of video forgery detection database and validation of its effectiveness,” J Appl Sci, vol. 36, no. 2, pp. 347–361, 2018.
[19]
Xiph.org, “Xiph.org Video Test Media [derf's collection],” Nov. 10, 2022. https://media.xiph.org/video/derf/
[20]
D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” ArXiv Prepr. ArXiv14126980, 2014.
[21]
J. Chao, X. Jiang, and T. Sun, “A novel video inter-frame forgery model detection scheme based on optical flow consistency,” in International workshop on digital watermarking, 2012, pp. 267–281.
[22]
M. Z. Tan, “Research on video forgery passive detection based on spatial-temporal feature,” Shanghai Jiao Tong University, 2012.
[23]
R. Girshick, “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448.

Cited By

View all

Index Terms

  1. Video Inter-frame Tampering Detection Based on SN-VGG+BiLSTM-AE Composite Model

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICIT '22: Proceedings of the 2022 10th International Conference on Information Technology: IoT and Smart City
    December 2022
    385 pages
    ISBN:9781450397438
    DOI:10.1145/3582197
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 30 March 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. BiLSTM
    2. Siamese
    3. autoencoder
    4. deep learning
    5. video forensics
    6. video inter-frame tampering

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICIT 2022
    ICIT 2022: IoT and Smart City
    December 23 - 25, 2022
    Shanghai, China

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)20
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 03 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media