CN111918077A - Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition - Google Patents

Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition Download PDF

Info

Publication number
CN111918077A
CN111918077A CN202010718614.9A CN202010718614A CN111918077A CN 111918077 A CN111918077 A CN 111918077A CN 202010718614 A CN202010718614 A CN 202010718614A CN 111918077 A CN111918077 A CN 111918077A
Authority
CN
China
Prior art keywords
video
data packet
audio
desktop
video image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010718614.9A
Other languages
Chinese (zh)
Inventor
虞焰兴
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Semxum Information Technology Co ltd
Original Assignee
Anhui Semxum Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Semxum Information Technology Co ltd filed Critical Anhui Semxum Information Technology Co ltd
Priority to CN202010718614.9A priority Critical patent/CN111918077A/en
Publication of CN111918077A publication Critical patent/CN111918077A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/2368Multiplexing of audio and video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/242Synchronization processes, e.g. processing of PCR [Program Clock References]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention discloses a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, and relates to the technical field of streaming media synthesis. The invention comprises the following steps: acquiring audio/position parameters of at least one display area pre-selected on a desktop; capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet; the camera collects the video of the teacher in class and binds the video into a second data packet; analyzing and merging the first data packet and the second data packet; the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image; and covering the video data successfully matched at the position of the designated area. The invention is convenient for students to remotely watch courseware for teaching and the synchronous explanation of teachers, and simultaneously, a plurality of video frames are randomly switched, thereby improving the learning effect of remote teaching of students.

Description

Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition
Technical Field
The invention belongs to the technical field of streaming media synthesis, and particularly relates to a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition.
Background
Multimedia (Multimedia) is a composite of multiple media, generally including multiple media forms such as text, sound, and images. Multimedia is embodied in informatization and is a trend of social development, multimedia education is a part of informatization, and the great promotion of multimedia education becomes a trend of educational development.
Due to the inherent characteristics of multimedia in various media forms such as text, sound and images, multimedia technology is often used in teaching, which is beneficial to improving the professional level of teachers and is beneficial to teachers to integrate teaching resources. The multimedia teaching technology can make up the defects in the traditional teaching, and is a novel scientific teaching technology.
Currently, in remote training adopting multimedia technology, a lecturer needs to share courseware (PPT/or other documents) and a computer desktop to students, but students in class cannot see expressions, actions and body languages of the teacher when watching a computer, so that body expression information of the teacher can be omitted, and besides, interactivity and interestingness are poor; in the live broadcasting process, the problem of video blocking or unsynchronized sound and pictures is caused because the data frames of the streaming media are easy to lose frames, and the teaching quality of teachers is seriously influenced.
Disclosure of Invention
The invention aims to provide a synchronous synthesis method for video live broadcast image sound acquisition and computer desktop acquisition, which is characterized in that desktop video information and teacher teaching video information are acquired, analyzed and combined, the combined video is subjected to region matching by a server, and after the matching is successful, the combined video is covered in a specified region, so that the problems that the teacher in a classroom cannot be synchronized with the computer desktop in the prior remote training, the teaching efficiency is low, and live broadcast video is easy to block are solved.
In order to solve the technical problems, the invention is realized by the following technical scheme:
the invention relates to a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, which comprises the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
Preferably, in step S1, the desktop video and the lesson video captured by the camera may be switched in the display area.
Preferably, in step S2, the client intercepts the video image of the whole desktop, and performs encoding and compression to obtain a video encoding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
Preferably, in step S4, the desktop video is subjected to an audio flag identification process, and a regional audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the regional audio/position parameter set are bound to a corresponding first network packet, and sent to the target server.
Preferably, the target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one group of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain the video image if the second network data packet is the second network data packet.
Preferably, the target server detects whether an audio mark exists in a display area corresponding to the position parameter, if so, matches audio data with a video image corresponding to the position parameter according to an area audio/position parameter set corresponding to the audio mark, and shares a corresponding client desktop area, and if not, shares a corresponding client desktop area with the video image corresponding to the position parameter.
Preferably, in S5, if the first packet and the second packet are analyzed and merged to include the audio flag, the merged audio data needs to be mixed.
The invention has the following beneficial effects:
according to the invention, the desktop video information and the teacher teaching video information are collected, analyzed and combined, the combined videos are subjected to region matching by the server, and the matched videos are covered in the designated region, so that students can conveniently watch courseware of teaching and synchronous explanation of teachers remotely, and meanwhile, a plurality of video frames are randomly switched, so that the students can conveniently watch and study, and the remote teaching effect of the students is improved.
Of course, it is not necessary for any product in which the invention is practiced to achieve all of the above-described advantages at the same time.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a step diagram of a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, the present invention is a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, comprising the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
In step S1, the desktop video and the lesson video captured by the camera may be switched in the display area.
In step S2, the client intercepts the video image of the entire desktop, and performs encoding and compression to obtain a video encoding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
In step S4, the desktop video is marked with an audio flag, and an area audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the area audio/position parameter set are bound to a corresponding first network packet, and sent to the target server.
The target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one group of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain a video image if the second network data packet is the second network data packet.
And if not, the video image corresponding to the position parameter is subjected to corresponding client desktop area sharing.
In S5, if the first packet and the second packet are analyzed and merged to include the audio flag, the merged audio data needs to be mixed.
It should be noted that, in the above system embodiment, each included unit is only divided according to functional logic, but is not limited to the above division as long as the corresponding function can be implemented; in addition, specific names of the functional units are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present invention.
In addition, it is understood by those skilled in the art that all or part of the steps in the method for implementing the embodiments described above may be implemented by a program instructing associated hardware, and the corresponding program may be stored in a computer-readable storage medium.
The preferred embodiments of the invention disclosed above are intended to be illustrative only. The preferred embodiments are not intended to be exhaustive or to limit the invention to the precise embodiments disclosed. Obviously, many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and the practical application, to thereby enable others skilled in the art to best utilize the invention. The invention is limited only by the claims and their full scope and equivalents.

Claims (7)

1. A method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition is characterized by comprising the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
2. The method as claimed in claim 1, wherein in step S1, the desktop video and the class video captured by the camera can be switched in the display area.
3. The method as claimed in claim 1, wherein in step S2, the client intercepts the video image of the whole desktop, and performs coding compression to obtain the video coding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
4. The method as claimed in claim 1, wherein in step S4, the desktop video is identified by an audio marker, and a regional audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the regional audio/position parameter set are bound to form a corresponding first network packet, and sent to the target server.
5. The method as claimed in claim 1 or 4, wherein the target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one set of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain the video image if the second network data packet is the second network data packet.
6. The method as claimed in claim 4, wherein the target server detects whether there is an audio flag in a display area corresponding to the location parameter, if so, matches the audio data with the video image corresponding to the location parameter according to a regional audio/location parameter set corresponding to the audio flag, and shares a corresponding client desktop area, and if not, shares a corresponding client desktop area with the video image corresponding to the location parameter.
7. The method as claimed in claim 1, wherein in S5, if the first data packet and the second data packet are analyzed and merged to include an audio flag, the merged audio data needs to be mixed.
CN202010718614.9A 2020-07-23 2020-07-23 Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition Pending CN111918077A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010718614.9A CN111918077A (en) 2020-07-23 2020-07-23 Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010718614.9A CN111918077A (en) 2020-07-23 2020-07-23 Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition

Publications (1)

Publication Number Publication Date
CN111918077A true CN111918077A (en) 2020-11-10

Family

ID=73281404

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010718614.9A Pending CN111918077A (en) 2020-07-23 2020-07-23 Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition

Country Status (1)

Country Link
CN (1) CN111918077A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140038781A1 (en) * 2012-07-31 2014-02-06 John Paul Foley Exercise system and method
CN105141883A (en) * 2015-08-18 2015-12-09 广东威创视讯科技股份有限公司 Desktop audio sharing method and system
US20160049082A1 (en) * 2014-08-14 2016-02-18 Albert Roy Leatherman, III System for Interactive Online Instruction
CN108769721A (en) * 2018-05-23 2018-11-06 福建掌搜科技有限公司 A kind of live scene intelligent switching system and its method
CN109168017A (en) * 2018-10-16 2019-01-08 深圳市三叶虫科技有限公司 A kind of net cast interaction systems and living broadcast interactive mode based on intelligent glasses
CN110213609A (en) * 2019-06-12 2019-09-06 珠海读书郎网络教育有限公司 The method, apparatus and storage medium of the company of progress wheat live streaming in Web education live streaming

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140038781A1 (en) * 2012-07-31 2014-02-06 John Paul Foley Exercise system and method
US20160049082A1 (en) * 2014-08-14 2016-02-18 Albert Roy Leatherman, III System for Interactive Online Instruction
CN105141883A (en) * 2015-08-18 2015-12-09 广东威创视讯科技股份有限公司 Desktop audio sharing method and system
CN108769721A (en) * 2018-05-23 2018-11-06 福建掌搜科技有限公司 A kind of live scene intelligent switching system and its method
CN109168017A (en) * 2018-10-16 2019-01-08 深圳市三叶虫科技有限公司 A kind of net cast interaction systems and living broadcast interactive mode based on intelligent glasses
CN110213609A (en) * 2019-06-12 2019-09-06 珠海读书郎网络教育有限公司 The method, apparatus and storage medium of the company of progress wheat live streaming in Web education live streaming

Similar Documents

Publication Publication Date Title
CN100535959C (en) System for multi media real-time synchronous teaching based on network
CN103646573B (en) A kind of generation method of professional format file of panning mode tutoring system
CN111611434B (en) Online course interaction method and interaction platform
CN106781770A (en) A kind of normalization interaction recording and broadcasting system
CN104539436A (en) Lesson content real-time live broadcasting method and system
CN102209080A (en) Terminal system for synchronous teaching or conferences and control method thereof
CN112468822A (en) Multimedia recording and broadcasting course interaction method based on video SEI message
CN110072137A (en) A kind of data transmission method and transmitting device of net cast
Moe et al. Media institutions as a research field: Three phases of Norwegian broadcasting research
CN106454200A (en) Video interaction method and video interaction system based on scene switching
CN109862375B (en) Cloud recording and broadcasting system
CN110933350A (en) Electronic cloud mirror recording and broadcasting system, method and device
CN110276999A (en) A kind of remote interactive teaching system and method with synchronous writing on the blackboard and direct broadcast function
CN111918077A (en) Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition
Samčović Accessibility of services in digital television for hearing impaired consumers
Yu et al. A near-reality approach to improve the e-learning open courseware
CN203193773U (en) Multimedia panoramic recording system
CN103944888B (en) The method, apparatus and system of a kind of resource-sharing
CN112837574B (en) Interactive classroom system and method thereof
CN109889916B (en) Application system of recorded broadcast data
CN114938443A (en) Real-time scoring method for experimental practice test based on streaming media
CN101883270A (en) Method for inserting related media in independent video streams
WO2017149447A1 (en) A system and method for providing real time media recommendations based on audio-visual analytics
KR20110124048A (en) System and method for providing on-line eduacational portal service
CN111614869A (en) 4K high definition camera double-circuit picture collection system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20201110