CN111918077A - Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition - Google Patents
Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition Download PDFInfo
- Publication number
- CN111918077A CN111918077A CN202010718614.9A CN202010718614A CN111918077A CN 111918077 A CN111918077 A CN 111918077A CN 202010718614 A CN202010718614 A CN 202010718614A CN 111918077 A CN111918077 A CN 111918077A
- Authority
- CN
- China
- Prior art keywords
- video
- data packet
- audio
- desktop
- video image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000001360 synchronised effect Effects 0.000 title abstract description 6
- 238000001308 synthesis method Methods 0.000 title description 3
- 238000000034 method Methods 0.000 claims abstract description 15
- 230000002194 synthesizing effect Effects 0.000 claims abstract description 6
- 230000006835 compression Effects 0.000 claims description 3
- 238000007906 compression Methods 0.000 claims description 3
- 239000003550 marker Substances 0.000 claims 1
- 230000015572 biosynthetic process Effects 0.000 abstract description 2
- 230000000694 effects Effects 0.000 abstract description 2
- 238000003786 synthesis reaction Methods 0.000 abstract description 2
- 230000009286 beneficial effect Effects 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2368—Multiplexing of audio and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/242—Synchronization processes, e.g. processing of PCR [Program Clock References]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The invention discloses a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, and relates to the technical field of streaming media synthesis. The invention comprises the following steps: acquiring audio/position parameters of at least one display area pre-selected on a desktop; capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet; the camera collects the video of the teacher in class and binds the video into a second data packet; analyzing and merging the first data packet and the second data packet; the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image; and covering the video data successfully matched at the position of the designated area. The invention is convenient for students to remotely watch courseware for teaching and the synchronous explanation of teachers, and simultaneously, a plurality of video frames are randomly switched, thereby improving the learning effect of remote teaching of students.
Description
Technical Field
The invention belongs to the technical field of streaming media synthesis, and particularly relates to a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition.
Background
Multimedia (Multimedia) is a composite of multiple media, generally including multiple media forms such as text, sound, and images. Multimedia is embodied in informatization and is a trend of social development, multimedia education is a part of informatization, and the great promotion of multimedia education becomes a trend of educational development.
Due to the inherent characteristics of multimedia in various media forms such as text, sound and images, multimedia technology is often used in teaching, which is beneficial to improving the professional level of teachers and is beneficial to teachers to integrate teaching resources. The multimedia teaching technology can make up the defects in the traditional teaching, and is a novel scientific teaching technology.
Currently, in remote training adopting multimedia technology, a lecturer needs to share courseware (PPT/or other documents) and a computer desktop to students, but students in class cannot see expressions, actions and body languages of the teacher when watching a computer, so that body expression information of the teacher can be omitted, and besides, interactivity and interestingness are poor; in the live broadcasting process, the problem of video blocking or unsynchronized sound and pictures is caused because the data frames of the streaming media are easy to lose frames, and the teaching quality of teachers is seriously influenced.
Disclosure of Invention
The invention aims to provide a synchronous synthesis method for video live broadcast image sound acquisition and computer desktop acquisition, which is characterized in that desktop video information and teacher teaching video information are acquired, analyzed and combined, the combined video is subjected to region matching by a server, and after the matching is successful, the combined video is covered in a specified region, so that the problems that the teacher in a classroom cannot be synchronized with the computer desktop in the prior remote training, the teaching efficiency is low, and live broadcast video is easy to block are solved.
In order to solve the technical problems, the invention is realized by the following technical scheme:
the invention relates to a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, which comprises the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
Preferably, in step S1, the desktop video and the lesson video captured by the camera may be switched in the display area.
Preferably, in step S2, the client intercepts the video image of the whole desktop, and performs encoding and compression to obtain a video encoding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
Preferably, in step S4, the desktop video is subjected to an audio flag identification process, and a regional audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the regional audio/position parameter set are bound to a corresponding first network packet, and sent to the target server.
Preferably, the target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one group of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain the video image if the second network data packet is the second network data packet.
Preferably, the target server detects whether an audio mark exists in a display area corresponding to the position parameter, if so, matches audio data with a video image corresponding to the position parameter according to an area audio/position parameter set corresponding to the audio mark, and shares a corresponding client desktop area, and if not, shares a corresponding client desktop area with the video image corresponding to the position parameter.
Preferably, in S5, if the first packet and the second packet are analyzed and merged to include the audio flag, the merged audio data needs to be mixed.
The invention has the following beneficial effects:
according to the invention, the desktop video information and the teacher teaching video information are collected, analyzed and combined, the combined videos are subjected to region matching by the server, and the matched videos are covered in the designated region, so that students can conveniently watch courseware of teaching and synchronous explanation of teachers remotely, and meanwhile, a plurality of video frames are randomly switched, so that the students can conveniently watch and study, and the remote teaching effect of the students is improved.
Of course, it is not necessary for any product in which the invention is practiced to achieve all of the above-described advantages at the same time.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a step diagram of a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition according to the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, the present invention is a method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition, comprising the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
In step S1, the desktop video and the lesson video captured by the camera may be switched in the display area.
In step S2, the client intercepts the video image of the entire desktop, and performs encoding and compression to obtain a video encoding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
In step S4, the desktop video is marked with an audio flag, and an area audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the area audio/position parameter set are bound to a corresponding first network packet, and sent to the target server.
The target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one group of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain a video image if the second network data packet is the second network data packet.
And if not, the video image corresponding to the position parameter is subjected to corresponding client desktop area sharing.
In S5, if the first packet and the second packet are analyzed and merged to include the audio flag, the merged audio data needs to be mixed.
It should be noted that, in the above system embodiment, each included unit is only divided according to functional logic, but is not limited to the above division as long as the corresponding function can be implemented; in addition, specific names of the functional units are only for convenience of distinguishing from each other, and are not used for limiting the protection scope of the present invention.
In addition, it is understood by those skilled in the art that all or part of the steps in the method for implementing the embodiments described above may be implemented by a program instructing associated hardware, and the corresponding program may be stored in a computer-readable storage medium.
The preferred embodiments of the invention disclosed above are intended to be illustrative only. The preferred embodiments are not intended to be exhaustive or to limit the invention to the precise embodiments disclosed. Obviously, many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and the practical application, to thereby enable others skilled in the art to best utilize the invention. The invention is limited only by the claims and their full scope and equivalents.
Claims (7)
1. A method for synchronously synthesizing live video image and sound acquisition and computer desktop acquisition is characterized by comprising the following steps:
step S1: acquiring audio/position parameters of at least one display area pre-selected on a desktop;
step S2: capturing current whole desktop video information, and binding the current whole desktop video information into a first data packet;
step S3: the camera collects the video of the teacher in class and binds the video into a second data packet;
step S4: analyzing and merging the first data packet and the second data packet;
step S5: the merged video is sent to a server side to match audio data of the regional audio/position parameter set with the video image;
step S6: and covering the video data successfully matched at the position of the designated area.
2. The method as claimed in claim 1, wherein in step S1, the desktop video and the class video captured by the camera can be switched in the display area.
3. The method as claimed in claim 1, wherein in step S2, the client intercepts the video image of the whole desktop, and performs coding compression to obtain the video coding data packet; the video image comprises at least one display area; and binding the video coding data packet and at least one position parameter into a corresponding second network data packet, and sending the second network data packet to the target server.
4. The method as claimed in claim 1, wherein in step S4, the desktop video is identified by an audio marker, and a regional audio/position parameter set is formed, and the extracted, encoded and compressed audio data and the regional audio/position parameter set are bound to form a corresponding first network packet, and sent to the target server.
5. The method as claimed in claim 1 or 4, wherein the target server performs type analysis on a first network data packet or a second network data packet which is sent or sent, decodes the encoded and compressed audio data to obtain at least one set of audio data if the first network data packet is the first network data packet, and decodes the encoded and compressed audio data to obtain the video image if the second network data packet is the second network data packet.
6. The method as claimed in claim 4, wherein the target server detects whether there is an audio flag in a display area corresponding to the location parameter, if so, matches the audio data with the video image corresponding to the location parameter according to a regional audio/location parameter set corresponding to the audio flag, and shares a corresponding client desktop area, and if not, shares a corresponding client desktop area with the video image corresponding to the location parameter.
7. The method as claimed in claim 1, wherein in S5, if the first data packet and the second data packet are analyzed and merged to include an audio flag, the merged audio data needs to be mixed.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010718614.9A CN111918077A (en) | 2020-07-23 | 2020-07-23 | Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010718614.9A CN111918077A (en) | 2020-07-23 | 2020-07-23 | Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111918077A true CN111918077A (en) | 2020-11-10 |
Family
ID=73281404
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010718614.9A Pending CN111918077A (en) | 2020-07-23 | 2020-07-23 | Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111918077A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140038781A1 (en) * | 2012-07-31 | 2014-02-06 | John Paul Foley | Exercise system and method |
CN105141883A (en) * | 2015-08-18 | 2015-12-09 | 广东威创视讯科技股份有限公司 | Desktop audio sharing method and system |
US20160049082A1 (en) * | 2014-08-14 | 2016-02-18 | Albert Roy Leatherman, III | System for Interactive Online Instruction |
CN108769721A (en) * | 2018-05-23 | 2018-11-06 | 福建掌搜科技有限公司 | A kind of live scene intelligent switching system and its method |
CN109168017A (en) * | 2018-10-16 | 2019-01-08 | 深圳市三叶虫科技有限公司 | A kind of net cast interaction systems and living broadcast interactive mode based on intelligent glasses |
CN110213609A (en) * | 2019-06-12 | 2019-09-06 | 珠海读书郎网络教育有限公司 | The method, apparatus and storage medium of the company of progress wheat live streaming in Web education live streaming |
-
2020
- 2020-07-23 CN CN202010718614.9A patent/CN111918077A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140038781A1 (en) * | 2012-07-31 | 2014-02-06 | John Paul Foley | Exercise system and method |
US20160049082A1 (en) * | 2014-08-14 | 2016-02-18 | Albert Roy Leatherman, III | System for Interactive Online Instruction |
CN105141883A (en) * | 2015-08-18 | 2015-12-09 | 广东威创视讯科技股份有限公司 | Desktop audio sharing method and system |
CN108769721A (en) * | 2018-05-23 | 2018-11-06 | 福建掌搜科技有限公司 | A kind of live scene intelligent switching system and its method |
CN109168017A (en) * | 2018-10-16 | 2019-01-08 | 深圳市三叶虫科技有限公司 | A kind of net cast interaction systems and living broadcast interactive mode based on intelligent glasses |
CN110213609A (en) * | 2019-06-12 | 2019-09-06 | 珠海读书郎网络教育有限公司 | The method, apparatus and storage medium of the company of progress wheat live streaming in Web education live streaming |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN100535959C (en) | System for multi media real-time synchronous teaching based on network | |
CN103646573B (en) | A kind of generation method of professional format file of panning mode tutoring system | |
CN111611434B (en) | Online course interaction method and interaction platform | |
CN106781770A (en) | A kind of normalization interaction recording and broadcasting system | |
CN104539436A (en) | Lesson content real-time live broadcasting method and system | |
CN102209080A (en) | Terminal system for synchronous teaching or conferences and control method thereof | |
CN112468822A (en) | Multimedia recording and broadcasting course interaction method based on video SEI message | |
CN110072137A (en) | A kind of data transmission method and transmitting device of net cast | |
Moe et al. | Media institutions as a research field: Three phases of Norwegian broadcasting research | |
CN106454200A (en) | Video interaction method and video interaction system based on scene switching | |
CN109862375B (en) | Cloud recording and broadcasting system | |
CN110933350A (en) | Electronic cloud mirror recording and broadcasting system, method and device | |
CN110276999A (en) | A kind of remote interactive teaching system and method with synchronous writing on the blackboard and direct broadcast function | |
CN111918077A (en) | Synchronous synthesis method for video live broadcast image and sound acquisition and computer desktop acquisition | |
Samčović | Accessibility of services in digital television for hearing impaired consumers | |
Yu et al. | A near-reality approach to improve the e-learning open courseware | |
CN203193773U (en) | Multimedia panoramic recording system | |
CN103944888B (en) | The method, apparatus and system of a kind of resource-sharing | |
CN112837574B (en) | Interactive classroom system and method thereof | |
CN109889916B (en) | Application system of recorded broadcast data | |
CN114938443A (en) | Real-time scoring method for experimental practice test based on streaming media | |
CN101883270A (en) | Method for inserting related media in independent video streams | |
WO2017149447A1 (en) | A system and method for providing real time media recommendations based on audio-visual analytics | |
KR20110124048A (en) | System and method for providing on-line eduacational portal service | |
CN111614869A (en) | 4K high definition camera double-circuit picture collection system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201110 |