skip to main content
10.1145/3343031.3350888acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

LiveSense: Contextual Advertising in Live Streaming Videos

Published: 15 October 2019 Publication History

Abstract

Live streaming has become a new form of entertainment, which attracts hundreds of millions of users worldwide. The huge amount of multimedia data in live streaming platforms creates tremendous opportunities for online advertising. However, existing state-of-the-art video advertising strategies (e.g., pre-roll and contextual mid-roll advertising) that rely on analyzing the whole video, are not applicable to live streaming videos. This paper describes a novel monetization framework, named LiveSense, for live streaming videos, which is able to display a contextually relevant ad at a suitable timestamp in a non-intrusive way. Specifically, given a live streaming video, we first employ a deep neural network to determine whether the current moment is appropriate for displaying an ad using the historical streaming data. Then, we detect a set of candidate ad insertion areas by incorporating image saliency, background map, and location priorities, so that the ad is displayed over the non-important area. We introduce three types of relevance metrics including textual relevance, global visual relevance and local visual relevance to select the contextually relevant ad. To minimize user intrusiveness, we initially display the ad at a non-important area. If the user is interested in the ad, we will show the ad in an overlaid window with a translucent background. Empirical evaluation on a real-world dataset demonstrates that our proposed framework is able to effectively display ads in live streaming videos while maintaining users' online experience.

References

[1]
Yusuf Aytar, Carl Vondrick, and Antonio Torralba. 2016. Soundnet: Learning sound representations from unlabeled video. In Advances in Neural Information Processing Systems. 892--900.
[2]
Lamberto Ballan, Marco Bertini, and Arjun Jain. 2008. A system for automatic detection and recognition of advertising trademarks in sports videos. In Proceedings of the 16th ACM international conference on Multimedia. 991--992.
[3]
Chiahu Chang, Kueiyi Hsieh, Mingche Chung, and Jaling Wu. 2008. ViSA: virtual spotlighted advertising. In Proceedings of the 16th ACM international conference on Multimedia. 837--840.
[4]
Xiang Chen, Bowei Chen, and Mohan Kankanhalli. 2017. Optimizing trade-offs among stakeholders in real-time bidding by incorporating multimedia metrics. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, 205--214.
[5]
Xiang Chen, Tam V Nguyen, and Mohan Kankanhalli. 2015. SalAd: A Multimodal Approach for Contextual Video Advertising. In 2015 IEEE International Symposium on Multimedia. 211--216.
[6]
Zhiqi Cheng, Yang Liu, Xiao Wu, and Xiansheng Hua. 2016. Video ecommerce: Towards online video advertising. In Proceedings of the 2016 ACM on Multimedia Conference. 1365--1374.
[7]
Jie Deng, Felix Cuadrado, Gareth Tyson, and Steve Uhlig. 2015. Behind the game: Exploring the twitch streaming platform. In 2015 International Workshop on Network and Systems Support for Games (NetGames). 1--6.
[8]
Jamesdouglas Hamilton. 1994. Time series analysis . Vol. 2. Princeton university press Princeton, NJ.
[9]
William Hamilton, Oliver Garretson, and Andruid Kerne. 2014. Streaming on twitch: fostering participatory communities of play within live mixed media. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 1315--1324.
[10]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770--778.
[11]
Zorah Hilvert, James Neill, Max Sjöblom, and Juho Hamari. 2018. Social motivations of live-streaming viewer engagement on Twitch. Computers in Human Behavior, Vol. 84 (2018), 58--67.
[12]
IAB. 2017. Live Video Streaming -- A global Perspective. https://www.iab.com/wp-content/uploads/2018/06/IAB-Live-Video-Streaming-Trends.pdf
[13]
Mehdi Kaytoue, Arlei Silva, Lo"ic Cerf, Wagner Meirajr, and Chedy Ra"issi. 2012. Watch me playing, i am a professional: a first study on video game live streaming. In Proceedings of the 21st International Conference on World Wide Web . 1181--1188.
[14]
Jan Kostka, Yvonne Anne Oswald, and Roger Wattenhofer. 2008. Word of mouth: Rumor dissemination in social networks. In International Colloquium on Structural Information and Communication Complexity . 185--196.
[15]
Zhiwei Li, Lei Zhang, and Weiying Ma. 2008. Delivering online advertisements inside images. In Proceedings of the 16th ACM international conference on Multimedia. 1051--1060.
[16]
Mathias Lux, Michael Riegler, Duc-Tien Dang-Nguyen, Marcus Larson, Martin Potthast, and Pål Halvorsen. 2018. GameStory Task at MediaEval 2018. In Working Notes Proceedings of the MediaEval 2018 Workshop .
[17]
Tao Mei, Jinlian Guo, Xiansheng Hua, and Falin Liu. 2010. AdOn: Toward contextual overlay in-video advertising. Multimedia systems, Vol. 16, 4--5 (2010), 335--344.
[18]
Tao Mei, Xiansheng Hua, Linjun Yang, and Shipeng Li. 2007. VideoSense: towards effective online video advertising. In Proceedings of the 15th ACM international conference on Multimedia. 1075--1084.
[19]
Tao Mei, Lusong Li, Xiansheng Hua, and Shipeng Li. 2012. ImageSense: Towards contextual image advertising. ACM Transactions on Multimedia Computing, Communications, and Applications, Vol. 8, 1 (2012), 6.
[20]
Tao Mei, Lusong Li, Xinmei Tian, Dacheng Tao, and Chongwah Ngo. 2016. PageSense: Towards Style-wise Contextual Advertising via Visual Analysis of Webpages. IEEE Transactions on Circuits and Systems for Video Technology (2016).
[21]
Lindasalwa Muda, Mumtaj Begam, and Irraivan Elamvazuthi. 2010. Voice recognition algorithms using mel frequency cepstral coefficient (MFCC) and dynamic time warping (DTW) techniques. arXiv preprint arXiv:1003.4083 (2010).
[22]
Tam V Nguyen and Mohan Kankanhalli. 2017. As-similar-as-possible saliency fusion. Multimedia Tools and Applications, Vol. 76, 8 (2017), 10501--10519.
[23]
Karla Okada, Edleno de Moura, Marco Cristo, David Fernandes, Marcos André Goncc alves, and Klessius Berlt. 2012. Advertisement selection for online videos. In Proceedings of the 18th Brazilian symposium on Multimedia and the web. ACM, 367--374.
[24]
Karine Pires and Gwendal Simon. 2015. YouTube live and Twitch: a tour of user-generated live streaming systems. In Proceedings of the 6th ACM Multimedia Systems Conference. 225--230.
[25]
Berthier Ribeiro-Neto, Marco Cristo, Paulo Golgher, and Edleno Silva de Moura. 2005. Impedance coupling in content-targeted advertising. In Proceedings of the 28th annual international ACM SIGIR conference on Research and development in information retrieval. 496--503.
[26]
Srinivasan Sengamedu, Neela Sawant, and Smita Wadhwa. 2007. vADeo: video advertising system. In Proceedings of the 15th ACM international conference on Multimedia. 455--456.
[27]
Mennatullah Siam, Sepehr Valipour, Martin Jagersand, and Nilanjan Ray. 2017. Convolutional gated recurrent networks for video segmentation. In Proceedings of the 2017 IEEE International Conference on Image Processing. 3090--3094.
[28]
Karen Simonyan and Andrew Zisserman. 2014. Two-stream convolutional networks for action recognition in videos. In Advances in neural information processing systems. 568--576.
[29]
Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. 2015. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE international conference on computer vision. 4489--4497.
[30]
Subhashini Venugopalan, Marcus Rohrbach, Jeffrey Donahue, Raymond Mooney, Trevor Darrell, and Kate Saenko. 2015. Sequence to sequence-video to text. In Proceedings of the IEEE international conference on computer vision. 4534--4542.
[31]
Wenguan Wang, Jianbing Shen, and Ling Shao. 2018. Video salient object detection via fully convolutional networks. IEEE Transactions on Image Processing, Vol. 27, 1 (2018), 38--49.
[32]
R William Soukoreff and I Scott Mackenzie. 1995. Theoretical upper and lower bounds on typing speed using a stylus and a soft keyboard. Behaviour & Information Technology, Vol. 14, 6 (1995), 370--379.
[33]
Zuxuan Wu, Yugang Jiang, Xi Wang, Hao Ye, and Xiangyang Xue. 2016. Multi-stream multi-class fusion of deep networks for video classification. In Proceedings of the 2016 ACM on Multimedia Conference. 791--800.
[34]
Karthik Yadati, Harish Katti, and Mohan Kankanhalli. 2014. CAVVA: Computational affective video-in-video advertising. IEEE Transactions on Multimedia, Vol. 16, 1 (2014), 15--23.
[35]
Jianming Zhang, Stan Sclaroff, Zhe Lin, Xiaohui Shen, Brian Price, and Radomir Mech. 2015. Minimum barrier salient object detection at 80 fps. In Proceedings of the IEEE International Conference on Computer Vision. 1404--1412.

Cited By

View all

Index Terms

  1. LiveSense: Contextual Advertising in Live Streaming Videos

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      MM '19: Proceedings of the 27th ACM International Conference on Multimedia
      October 2019
      2794 pages
      ISBN:9781450368896
      DOI:10.1145/3343031
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 15 October 2019

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. interactive display
      2. live streaming
      3. online advertising
      4. user engagement

      Qualifiers

      • Research-article

      Conference

      MM '19
      Sponsor:

      Acceptance Rates

      MM '19 Paper Acceptance Rate 252 of 936 submissions, 27%;
      Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)111
      • Downloads (Last 6 weeks)36
      Reflects downloads up to 23 Dec 2024

      Other Metrics

      Citations

      Cited By

      View all

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media