CN108965757B - Video recording method, device, terminal and storage medium - Google Patents

Video recording method, device, terminal and storage medium Download PDF

Info

Publication number
CN108965757B
CN108965757B CN201810871649.9A CN201810871649A CN108965757B CN 108965757 B CN108965757 B CN 108965757B CN 201810871649 A CN201810871649 A CN 201810871649A CN 108965757 B CN108965757 B CN 108965757B
Authority
CN
China
Prior art keywords
audio file
sound
special effect
file
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810871649.9A
Other languages
Chinese (zh)
Other versions
CN108965757A (en
Inventor
陈宙炜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN201810871649.9A priority Critical patent/CN108965757B/en
Publication of CN108965757A publication Critical patent/CN108965757A/en
Application granted granted Critical
Publication of CN108965757B publication Critical patent/CN108965757B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)

Abstract

The invention discloses a video recording method, a video recording device, a video recording terminal and a video recording storage medium, and belongs to the technical field of the Internet. The method comprises the following steps: when a recording instruction is received, acquiring audio data and images of a target object, wherein the recording instruction is used for indicating the audio data and images of the target object when the target object sing a target song; generating a first audio file in a first file format according to the acquired audio data; when a sound special effect instruction is received, converting the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing; adding a target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file; and generating a video file according to the image and the third audio file. The invention realizes the processing of the audio in the video file and improves the flexibility of video processing.

Description

Video recording method, device, terminal and storage medium
Technical Field
The present invention relates to the field of internet technologies, and in particular, to a video recording method, apparatus, terminal, and storage medium.
Background
With the development of internet technology, a plurality of video applications support the recording of songs, and during the process of singing a certain song, the video applications can record the user and share the recorded videos to a network platform of the video applications through a network.
In the related art, the terminal records audio data and images of a user singing a target song through the video application, and in addition, the terminal can also perform beautification processing on the collected images, for example, whitening processing, buffing processing and the like on the images in the video. And then, the terminal sends the processed video to a server, and the server shares the video to a network platform of the video application.
The recording process is mainly used for recording songs sung by a user, however, the process only beautifies images in a video and only performs video processing from the perspective of a video picture, and the effect of beautifying the video is poor; moreover, the audio data in the video file is compressed audio data, and the sound quality is damaged, so that the audio in the video file cannot be processed, which results in poor flexibility of video processing in the video recording process.
Disclosure of Invention
The embodiment of the invention provides a video recording method, a video recording device, a video recording terminal and a video recording storage medium, which can solve the problem that audio data in a video cannot be processed in the related art. The technical scheme is as follows:
in one aspect, a video recording method is provided, and the method includes:
when a recording instruction is received, acquiring audio data and images of a target object, wherein the recording instruction is used for indicating the audio data and images of the target object when the target object sing a target song;
generating a first audio file in a first file format according to the acquired audio data;
when a sound special effect instruction is received, converting the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing;
adding a target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file;
and generating a video file according to the image and the third audio file.
Optionally, when the sound special effect instruction is received, before the first audio file is converted into the second audio file, the method further includes:
displaying a plurality of sound effect options on a recording interface, wherein each sound effect option corresponds to one type of sound special effect;
and when any sound effect option in the plurality of sound effect options is detected to be selected, receiving the sound special effect instruction, and determining the sound special effect of the type corresponding to the selected sound effect option as the target sound special effect.
Optionally, the adding a target sound special effect to the second audio file according to the sound special effect instruction to obtain a third audio file includes:
displaying the playing time length of the audio data in the second audio file in a sound effect adding interface;
determining target audio data in the second audio file based on the selected target time period in the playing time length;
and adding the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain a third audio file.
Optionally, the generating a first audio file in a first file format according to the collected audio data includes:
carrying out noise reduction processing on the acquired audio data through a preset noise reduction algorithm;
and writing the audio data subjected to noise reduction processing into the first audio file.
Optionally, the first audio file is an MP3 audio file, and the second audio file is a pulse code modulation PCM audio file.
In one aspect, a video recording apparatus is provided, the apparatus comprising:
the acquisition module is used for acquiring the audio data and the image of the target object when receiving a recording instruction, wherein the recording instruction is used for indicating the audio data and the image when the target object is recorded to sing the target song;
the generating module is used for generating a first audio file in a first file format according to the acquired audio data;
the conversion module is used for converting the first audio file into a second audio file when a sound special effect instruction is received, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing;
the adding module is used for adding a target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file;
the generating module is further configured to generate a video file according to the image and the third audio file.
Optionally, the apparatus further comprises:
the display module is used for displaying a plurality of sound effect options on the recording interface, and each sound effect option corresponds to one type of sound special effect;
and the receiving module is used for receiving the sound special effect instruction when any one sound effect option in the plurality of sound effect options is detected to be selected, and determining the sound special effect of the type corresponding to the selected sound effect option as the target sound special effect.
Optionally, the adding module is further configured to display a playing time length of the audio data in the second audio file in a sound effect adding interface; determining target audio data in the second audio file based on the selected target time period in the playing time length; and adding the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain a third audio file.
Optionally, the generating module is further configured to perform noise reduction processing on the acquired audio data through a preset noise reduction algorithm; and writing the audio data subjected to noise reduction processing into the first audio file.
Optionally, the first audio file is an MP3 audio file, and the second audio file is a pulse code modulation PCM audio file.
In one aspect, a terminal is provided, and the terminal includes a processor and a memory, where at least one instruction is stored in the memory, and the instruction is loaded and executed by the processor to implement the operation performed by the video recording method as described above.
In one aspect, a computer-readable storage medium is provided, in which at least one instruction is stored, and the instruction is loaded and executed by a processor to implement the operations performed by the video recording method as described above.
The technical scheme provided by the embodiment of the invention has the following beneficial effects:
according to the method and the device provided by the embodiment of the invention, when the terminal receives the recording instruction, the terminal collects the audio data and the image of the target object, and the terminal generates the first audio file in the first file format according to the collected audio data; when a sound special effect instruction is received, the terminal converts the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing; the terminal adds a target sound special effect in the second audio file according to the sound special effect instruction to obtain a third audio file; and the terminal generates a video file according to the image and the third audio file. According to the method, the second audio file in the converted second file format is subjected to sound special effect processing, so that the processing of the audio data in the video file is realized, the blank of the processing of the audio data in the video is filled, and the flexibility of the video processing is improved. And by adding the sound special effect to the second audio file, the interestingness of the video file is improved, and the user activity of the video application is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a flowchart of a video recording method according to an embodiment of the present invention;
fig. 2 is a flowchart of a video recording method according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a terminal interface according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a terminal interface according to an embodiment of the present invention;
fig. 5 is a flow chart of video recording according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a video recording apparatus according to an embodiment of the present invention;
fig. 7 is a schematic structural diagram of a terminal according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Fig. 1 is a flowchart of a video recording method according to an embodiment of the present invention, where an execution subject of the method may be a terminal, as shown in fig. 1, the method includes:
101. when a recording instruction is received, acquiring audio data and images of a target object, wherein the recording instruction is used for indicating the audio data and images of the target object when the target object sing a target song;
102. generating a first audio file in a first file format according to the acquired audio data;
103. when a sound special effect instruction is received, converting the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing;
104. adding a target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file;
105. and generating a video file according to the image and the third audio file.
Optionally, when the sound special effect instruction is received, before the first audio file is converted into the second audio file, the method further includes:
displaying a plurality of sound effect options on a recording interface, wherein each sound effect option corresponds to one type of sound special effect;
when any sound effect option in the plurality of sound effect options is detected to be selected, the sound effect instruction is received based on the target sound effect of the type corresponding to the selected sound effect option.
Optionally, the adding a target sound special effect to the second audio file according to the sound special effect instruction to obtain a third audio file includes:
displaying the playing time length of the audio data in the second audio file in a sound effect adding interface;
determining target audio data in the second audio file based on the selected target time period in the playing time length;
and adding the target sound special effect to the target audio data according to the sound special effect instruction to obtain the third audio file.
Optionally, the generating a first audio file in a first file format according to the collected audio data includes:
carrying out noise reduction processing on the acquired audio data through a preset noise reduction algorithm;
and writing the audio data subjected to noise reduction processing into the first audio file.
Optionally, the first audio file is an MP3 audio file, and the second audio file is a pulse code modulation PCM audio file.
In the embodiment of the invention, when a terminal receives a recording instruction, the terminal acquires audio data and images of a target object, and the terminal generates a first audio file in a first file format according to the acquired audio data; when a sound special effect instruction is received, the terminal converts the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing; the terminal adds a target sound special effect in the second audio file according to the sound special effect instruction to obtain a third audio file; and the terminal generates a video file according to the image and the third audio file. According to the method, the second audio file in the converted second file format is subjected to sound special effect processing, so that the processing of the audio data in the video file is realized, the blank of the processing of the audio data in the video is filled, and the flexibility of the video processing is improved. And by adding the sound special effect to the second audio file, the interestingness of the video file is improved, and the user activity of the video application is improved.
Fig. 2 is a flowchart of a video recording method according to an embodiment of the present invention. The execution subject of the embodiment of the present invention is a terminal, and referring to fig. 2, the method includes:
201. and when receiving the recording instruction, the terminal collects the audio data and the image of the target object in real time.
Wherein, the recording instruction is used for indicating audio data and images when the recording target object sing the target song; in this step, when the terminal receives the recording instruction, the terminal starts a camera and an audio acquisition device to start to acquire the image and audio data of the target object in real time. The target object can trigger a recording instruction of the terminal in the video application: when the video application is opened, the terminal may display a plurality of song identifications for a plurality of alternative songs on the current interface. When the terminal detects that any song identification is selected, the terminal receives the recording instruction.
When the terminal receives the recording instruction, the terminal can also acquire the accompaniment audio of the target song and play the accompaniment audio of the target song, the target object can sing the target song along with the accompaniment audio, and the terminal collects the audio data and the image of the target object in real time in the process of singing the target song by the target object. In addition, the terminal can also display the lyrics of the target song in the current interface.
202. The terminal generates a first audio file in a first file format according to the collected audio data.
In this step, the terminal creates a first audio file, encodes the acquired audio data according to a first file format of the first audio file, and writes the encoded audio data into the first audio file.
The first Audio file may be an MP3(Moving Picture Experts Group Audio Layer III) Audio file.
In a possible implementation, the terminal may also filter out noise in the audio data. Correspondingly, the steps can also be as follows: the terminal carries out noise reduction processing on the acquired audio data through a preset noise reduction algorithm, and writes the audio data subjected to noise reduction processing into the first audio file.
It should be noted that, when the terminal collects the audio data, the MP3 audio file is generally generated first, and since the MP3 audio file has a small compression on the audio data, when the MP3 audio file is played, the sound quality is not significantly reduced, and the storage space occupied by the MP3 audio file is small, so that on the premise of ensuring the playing fluency, the storage space occupied by the file is saved to the maximum extent, and the system resource utilization rate of the terminal is improved. In addition, the terminal can also perform noise reduction processing on the acquired audio data, so that noise, noise and the like in the acquired voice are eliminated, and the recorded audio data is further ensured to have higher tone quality.
203. And when a sound special effect instruction is received, the terminal converts the first audio file into a second audio file.
The file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing. The sound special effect method is a method for adding a sound special effect to the second audio file subsequently. In this step, the terminal converts the first audio file into a second audio file in a second file format according to the second file format.
The terminal can display a sound special effect adding button on a recording interface, when the terminal detects that the sound special effect adding button is triggered, the terminal can provide a plurality of alternative sound special effects, and then the sound special effects are added based on the sound special effects selected by a user. Wherein, the process can be as follows: the terminal displays a plurality of sound effect options on a recording interface, wherein each sound effect option corresponds to one type of sound special effect; when any sound effect option in the plurality of sound effect options is detected to be selected, the sound effect instruction is received, and the sound effect corresponding to the selected sound effect option is determined as the target sound effect.
Wherein, the second audio file is a PCM (Pulse Code Modulation) audio file.
As shown in fig. 3, the terminal may display a "sound effect" button in the upper right corner of the recording interface, and when the "sound effect" button is triggered, a plurality of sound effect options are displayed on the recording interface of the terminal, as shown in fig. 4, the plurality of sound effect options may include: electric sound effect, minions sound effect, phonograph sound effect, magic beasts sound effect, and the like. Of course, the sound effect option may also include an option to indicate that no sound effect is to be added.
It should be noted that, the terminal converts the first audio file into the second audio file, so that the terminal can perform subsequent processing of the sound special effect based on the audio file in the second file format. In addition, the second audio file can be a PCM audio file, audio data in the PCM audio file is not subjected to any coding and compression processing, compared with audio files in other formats, the tone quality of the audio data in the PCM audio file is higher, subsequent sound effect processing is performed on the basis of the PCM audio file, an audio file with better tone quality is obtained, and the accuracy of sound effect processing is improved.
204. And the terminal adds a target sound special effect in the second audio file according to the sound special effect instruction to obtain a third audio file.
In this step, the terminal performs sound special effect processing on the audio data in the second audio file through a target sound effect algorithm according to the sound special effect instruction, and generates the third audio file according to the audio data after the sound special effect processing.
The terminal can display a plurality of sound effect options on the interface, each sound effect option corresponds to one sound effect algorithm, and the terminal can store the sound effect algorithm corresponding to each sound effect option in advance. The target sound effect algorithm is a sound effect algorithm corresponding to the selected sound effect option in the sound effect options. And when a sound special effect instruction is received, the terminal takes a sound effect algorithm corresponding to the selected sound effect option as a target sound effect algorithm based on the selected sound effect option, and adds a target sound characteristic to the second audio file through the target sound effect algorithm to obtain a third audio file. Wherein the third audio file is also a PCM audio file.
The terminal can adjust the frequency, the tone and the like of the sound signal corresponding to the audio data through a target sound effect algorithm, so that the adjusted audio data are the sound meeting the personalized requirements of the user. For example, the user can add a sound special effect to make the sound of singing have the effects of a minions special effect, a telephone sound special effect and the like after being processed. In addition, the sound effect algorithm corresponding to each sound effect option may be any algorithm for realizing addition of a sound effect of a type corresponding to the sound effect option, and this is not specifically limited in the embodiment of the present invention. For example, the sound effect algorithm may be SOLA (Synchronized overlap-add method).
As shown in fig. 4, the terminal may display a plurality of sound effect options, and the user may select among the plurality of sound effect options, so that the terminal performs sound effect processing based on the personalized selection of the user.
In a possible implementation manner, the terminal may further perform sound effect processing on only a part of the audio data in the second audio file based on a user requirement, where the terminal may select a part of the audio data to be processed based on a time period corresponding to the audio data. Specifically, the steps may also be: the terminal displays the playing time length of the audio data in the second audio file in a sound effect adding interface, and the terminal determines the target audio data in the second audio file based on the selected target time interval in the playing time length. And the terminal adds the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain the third audio file.
As shown in fig. 4, the terminal may display the playing time length of the audio data in the second audio file, the user may select the target audio data in the second audio file based on the playing time length, and the terminal performs sound effect processing based on the target audio data selected by the user.
It should be noted that the terminal can directly perform sound special effect processing on the PCM audio file, which solves the problem that the sound effect algorithm is not matched with the format of the audio file, so that the terminal can perform sound effect processing based on the PCM audio file with higher sound quality, and the applicability of the video recording method is improved.
205. And the terminal generates a video file according to the image and the third audio file.
In this step, the third audio file is also a PCM audio file. The terminal may convert the third audio file into a fourth audio file, and then generate the video file according to the acquired image and the fourth audio file. Wherein the fourth audio file is an MP3 audio file.
Furthermore, the terminal can also send the video file to a server, and the server sends the video file to a platform of the video application, so that more users can browse the video file, information exchange among multiple users is promoted, and the activity of the users on the application platform is improved.
It should be noted that the PCM audio file is a lossless compressed file, and generally occupies a large storage space, and is inefficient when being transmitted or played. In the step, when audio and video are stored, the audio data are stored as MP3 audio files, so that the size of the files is reduced to the maximum extent on the premise of guaranteeing smooth playing of the audio files, the storage space occupied by the files is reduced, and the cost required in the process of recording the video is saved.
To describe the technical process of the embodiment of the present invention more clearly, taking the whole technical process shown in fig. 5 as an example, a video recording process is introduced, as shown in fig. 5, when a terminal receives a recording instruction, the terminal collects audio data and images, and encodes the audio data into an MP3 audio file, when a user needs to add a sound special effect, the terminal converts an MP3 audio file into a PCM audio file based on the sound special effect instruction, and adds a target sound special effect selected by the user in the PCM audio file to generate a PCM audio file with the target sound special effect, and finally, the terminal also needs to convert the PCM audio file into an MP3 audio file, so as to facilitate subsequent file transmission and playing based on the MP3 audio file.
In the embodiment of the invention, when a terminal receives a recording instruction, the terminal acquires audio data and images of a target object, and the terminal generates a first audio file in a first file format according to the acquired audio data; when a sound special effect instruction is received, the terminal converts the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing; the terminal adds a target sound special effect in the second audio file according to the sound special effect instruction to obtain a third audio file; and the terminal generates a video file according to the image and the third audio file. According to the method, the second audio file in the converted second file format is subjected to sound special effect processing, so that the processing of the audio data in the video file is realized, the blank of the processing of the audio data in the video is filled, and the flexibility of the video processing is improved. And by adding the sound special effect to the second audio file, the interestingness of the video file is improved, and the user activity of the video application is improved.
Fig. 6 is a schematic structural diagram of a video recording apparatus according to an embodiment of the present invention. Referring to fig. 6, the apparatus includes: the system comprises an acquisition module 601, a generation module 602, a conversion module 603 and an addition module 604.
The acquisition module 601 is configured to acquire audio data and images of a target object when a recording instruction is received, where the recording instruction is used to instruct the recording of the audio data and images of the target object when a target song is sung;
a generating module 602, configured to generate a first audio file in a first file format according to the acquired audio data;
a conversion module 603, configured to convert, when a sound special effect instruction is received, the first audio file into a second audio file, where a file format of the second audio file is a second file format that is matched with a sound special effect method adopted by sound special effect processing;
an adding module 604, configured to add a target sound special effect to the second audio file according to the sound special effect instruction, to obtain a third audio file;
the generating module 602 is further configured to generate a video file according to the image and the third audio file.
Optionally, the apparatus further comprises:
the display module is used for displaying a plurality of sound effect options on the recording interface, and each sound effect option corresponds to one type of sound special effect;
and the receiving module is used for receiving the sound special effect instruction when any one sound effect option in the plurality of sound effect options is detected to be selected, and determining the sound special effect of the type corresponding to the selected sound effect option as the target sound special effect.
Optionally, the adding module 604 is further configured to display the playing time of the audio data in the second audio file in a sound effect adding interface; determining target audio data in the second audio file based on the selected target time period in the playing time length; and adding the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain the third audio file.
Optionally, the generating module 602 is further configured to perform noise reduction processing on the acquired audio data through a preset noise reduction algorithm; and writing the audio data subjected to noise reduction processing into the first audio file.
Optionally, the first audio file is an MP3 audio file, and the second audio file is a pulse code modulation PCM audio file.
In the embodiment of the invention, when a terminal receives a recording instruction, the terminal acquires audio data and images of a target object, and the terminal generates a first audio file in a first file format according to the acquired audio data; when a sound special effect instruction is received, the terminal converts the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing; the terminal adds a target sound special effect in the second audio file according to the sound special effect instruction to obtain a third audio file; and the terminal generates a video file according to the image and the third audio file. According to the method, the second audio file in the converted second file format is subjected to sound special effect processing, so that the processing of the audio data in the video file is realized, the blank of the processing of the audio data in the video is filled, and the flexibility of the video processing is improved. And by adding the sound special effect to the second audio file, the interestingness of the video file is improved, and the user activity of the video application is improved.
All the above optional technical solutions may be combined arbitrarily to form the optional embodiments of the present disclosure, and are not described herein again.
It should be noted that: in the video recording apparatus provided in the foregoing embodiment, when recording a video, only the division of the functional modules is described as an example, and in practical applications, the function distribution may be completed by different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules to complete all or part of the functions described above. In addition, the video recording apparatus and the video recording method provided by the above embodiments belong to the same concept, and specific implementation processes thereof are described in the method embodiments in detail and are not described herein again.
Fig. 7 is a block diagram illustrating a terminal 700 according to an exemplary embodiment of the present invention. The terminal 700 may be: a smart phone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, motion video Experts compression standard Audio Layer 3), an MP4 player (Moving Picture Experts Group Audio Layer IV, motion video Experts compression standard Audio Layer 4), a notebook computer, or a desktop computer. Terminal 700 may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, and so on.
In general, terminal 700 includes: a processor 701 and a memory 702.
The processor 701 may include one or more processing cores, such as a 4-core processor, a 7-core processor, and so on. The processor 701 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 701 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 701 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed on the display screen. In some embodiments, the processor 701 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 702 may include one or more computer-readable storage media, which may be non-transitory. Memory 702 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 702 is used to store at least one instruction for execution by processor 701 to implement a video recording method as provided by method embodiments herein.
In some embodiments, the terminal 700 may further optionally include: a peripheral interface 703 and at least one peripheral. The processor 701, the memory 702, and the peripheral interface 703 may be connected by buses or signal lines. Various peripheral devices may be connected to peripheral interface 703 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of radio frequency circuitry 704, touch screen display 705, camera 706, audio circuitry 707, positioning components 708, and power source 709.
The peripheral interface 703 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 701 and the memory 702. In some embodiments, processor 701, memory 702, and peripheral interface 703 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 701, the memory 702, and the peripheral interface 703 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 704 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 704 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 704 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 704 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 704 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 704 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 705 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 705 is a touch display screen, the display screen 705 also has the ability to capture touch signals on or over the surface of the display screen 705. The touch signal may be input to the processor 701 as a control signal for processing. At this point, the display 705 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 705 may be one, providing the front panel of the terminal 700; in other embodiments, the display 705 can be at least two, respectively disposed on different surfaces of the terminal 700 or in a folded design; in still other embodiments, the display 705 may be a flexible display disposed on a curved surface or on a folded surface of the terminal 700. Even more, the display 705 may be arranged in a non-rectangular irregular pattern, i.e. a shaped screen. The Display 705 may be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), or the like.
The camera assembly 706 is used to capture images or video. Optionally, camera assembly 706 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 706 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuitry 707 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 701 for processing or inputting the electric signals to the radio frequency circuit 704 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 700. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 701 or the radio frequency circuit 704 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 707 may also include a headphone jack.
The positioning component 708 is used to locate the current geographic Location of the terminal 700 for navigation or LBS (Location Based Service). The Positioning component 708 can be a Positioning component based on the GPS (Global Positioning System) in the united states, the beidou System in china, the graves System in russia, or the galileo System in the european union.
Power supply 709 is provided to supply power to various components of terminal 700. The power source 709 may be alternating current, direct current, disposable batteries, or rechargeable batteries. When power source 709 includes a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 700 also includes one or more sensors 710. The one or more sensors 710 include, but are not limited to: acceleration sensor 711, gyro sensor 712, pressure sensor 713, fingerprint sensor 714, optical sensor 715, and proximity sensor 716.
The acceleration sensor 711 can detect the magnitude of acceleration in three coordinate axes of a coordinate system established with the terminal 700. For example, the acceleration sensor 711 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 701 may control the touch screen 705 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 711. The acceleration sensor 711 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 712 may detect a body direction and a rotation angle of the terminal 700, and the gyro sensor 712 may cooperate with the acceleration sensor 711 to acquire a 3D motion of the terminal 700 by the user. From the data collected by the gyro sensor 712, the processor 701 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 713 may be disposed on a side bezel of terminal 700 and/or an underlying layer of touch display 705. When the pressure sensor 713 is disposed on a side frame of the terminal 700, a user's grip signal on the terminal 700 may be detected, and the processor 701 performs right-left hand recognition or shortcut operation according to the grip signal collected by the pressure sensor 713. When the pressure sensor 713 is disposed at a lower layer of the touch display 705, the processor 701 controls the operability control on the UI interface according to the pressure operation of the user on the touch display 705. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 714 is used for collecting a fingerprint of a user, and the processor 701 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 714, or the fingerprint sensor 714 identifies the identity of the user according to the collected fingerprint. When the user identity is identified as a trusted identity, the processor 701 authorizes the user to perform relevant sensitive operations, including unlocking a screen, viewing encrypted information, downloading software, paying, changing settings, and the like. The fingerprint sensor 714 may be disposed on the front, back, or side of the terminal 700. When a physical button or a vendor Logo is provided on the terminal 700, the fingerprint sensor 714 may be integrated with the physical button or the vendor Logo.
The optical sensor 715 is used to collect the ambient light intensity. In one embodiment, the processor 701 may control the display brightness of the touch display 705 based on the ambient light intensity collected by the optical sensor 715. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 705 is increased; when the ambient light intensity is low, the display brightness of the touch display 705 is turned down. In another embodiment, processor 701 may also dynamically adjust the shooting parameters of camera assembly 706 based on the ambient light intensity collected by optical sensor 715.
A proximity sensor 716, also referred to as a distance sensor, is typically disposed on a front panel of the terminal 700. The proximity sensor 716 is used to collect the distance between the user and the front surface of the terminal 700. In one embodiment, when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually decreases, the processor 701 controls the touch display 705 to switch from the bright screen state to the dark screen state; when the proximity sensor 716 detects that the distance between the user and the front surface of the terminal 700 gradually becomes larger, the processor 701 controls the touch display 705 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 7 is not intended to be limiting of terminal 700 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
In an exemplary embodiment, a computer-readable storage medium, such as a memory, including instructions executable by a processor in a terminal to perform the video recording method in the above embodiments is also provided. For example, the computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (12)

1. A method for video recording, the method comprising:
when a recording instruction is received, acquiring audio data and images of a target object, wherein the recording instruction is used for indicating the audio data and images of the target object when the target object sing a target song;
generating a first audio file in a first file format according to the acquired audio data;
when a sound special effect instruction is received, converting the first audio file into a second audio file, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing, and the sound special effect method is a method for adding a selected target sound special effect to the second audio file;
adding the target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file;
and generating a video file according to the image and the third audio file.
2. The method of claim 1, wherein before converting the first audio file to a second audio file when the sound effect instruction is received, the method further comprises:
displaying a plurality of sound effect options on a recording interface, wherein each sound effect option corresponds to one type of sound special effect;
and when any sound effect option in the plurality of sound effect options is detected to be selected, receiving the sound special effect instruction, and determining the sound special effect of the type corresponding to the selected sound effect option as the target sound special effect.
3. The method of claim 1, wherein adding a target sound effect to the second audio file according to the sound effect instruction to obtain a third audio file comprises:
displaying the playing time length of the audio data in the second audio file in a sound effect adding interface;
determining target audio data in the second audio file based on the selected target time period in the playing time length;
and adding the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain a third audio file.
4. The method of claim 1, wherein generating the first audio file in the first file format from the captured audio data comprises:
carrying out noise reduction processing on the acquired audio data through a preset noise reduction algorithm;
and writing the audio data subjected to noise reduction processing into the first audio file.
5. The method of claim 1, wherein the first audio file is an MP3 audio file, and wherein the second audio file is a pulse code modulation, PCM, audio file.
6. A video recording apparatus, characterized in that the apparatus comprises:
the acquisition module is used for acquiring the audio data and the image of the target object when receiving a recording instruction, wherein the recording instruction is used for indicating the audio data and the image when the target object is recorded to sing the target song;
the generating module is used for generating a first audio file in a first file format according to the acquired audio data;
the conversion module is used for converting the first audio file into a second audio file when a sound special effect instruction is received, wherein the file format of the second audio file is a second file format matched with a sound special effect method adopted by sound special effect processing, and the sound special effect method is a method for adding a selected target sound special effect to the second audio file;
the adding module is used for adding the target sound special effect into the second audio file according to the sound special effect instruction to obtain a third audio file;
the generating module is further configured to generate a video file according to the image and the third audio file.
7. The apparatus of claim 6, further comprising:
the display module is used for displaying a plurality of sound effect options on the recording interface, and each sound effect option corresponds to one type of sound special effect;
and the receiving module is used for receiving the sound special effect instruction when any one sound effect option in the plurality of sound effect options is detected to be selected, and determining the sound special effect of the type corresponding to the selected sound effect option as the target sound special effect.
8. The apparatus of claim 6,
the adding module is also used for displaying the playing time length of the audio data in the second audio file in a sound effect adding interface; determining target audio data in the second audio file based on the selected target time period in the playing time length; and adding the target sound special effect to the target audio data in the second audio file according to the sound special effect instruction to obtain a third audio file.
9. The apparatus of claim 6,
the generation module is further used for carrying out noise reduction processing on the acquired audio data through a preset noise reduction algorithm; and writing the audio data subjected to noise reduction processing into the first audio file.
10. The apparatus of claim 6,
the first audio file is an MP3 audio file, and the second audio file is a Pulse Code Modulation (PCM) audio file.
11. A terminal, characterized in that the terminal comprises a processor and a memory, wherein at least one instruction is stored in the memory, and the instruction is loaded and executed by the processor to implement the operations performed by the video recording method according to any one of claims 1 to 5.
12. A computer-readable storage medium having stored therein at least one instruction which is loaded and executed by a processor to perform operations performed by a video recording method according to any one of claims 1 to 5.
CN201810871649.9A 2018-08-02 2018-08-02 Video recording method, device, terminal and storage medium Active CN108965757B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810871649.9A CN108965757B (en) 2018-08-02 2018-08-02 Video recording method, device, terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810871649.9A CN108965757B (en) 2018-08-02 2018-08-02 Video recording method, device, terminal and storage medium

Publications (2)

Publication Number Publication Date
CN108965757A CN108965757A (en) 2018-12-07
CN108965757B true CN108965757B (en) 2021-04-06

Family

ID=64467134

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810871649.9A Active CN108965757B (en) 2018-08-02 2018-08-02 Video recording method, device, terminal and storage medium

Country Status (1)

Country Link
CN (1) CN108965757B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109920457A (en) * 2019-03-01 2019-06-21 广州酷狗计算机科技有限公司 Play method, apparatus, equipment and the storage medium of song
CN109819314B (en) * 2019-03-05 2022-07-12 广州酷狗计算机科技有限公司 Audio and video processing method and device, terminal and storage medium
CN112397102B (en) * 2019-08-14 2022-07-08 腾讯科技(深圳)有限公司 Audio processing method and device and terminal
CN111935395A (en) * 2020-06-29 2020-11-13 联想(北京)有限公司 Video file generation method and electronic equipment
CN112165647B (en) * 2020-08-26 2022-06-17 北京字节跳动网络技术有限公司 Audio data processing method, device, equipment and storage medium
CN112165591B (en) * 2020-09-30 2022-05-31 联想(北京)有限公司 Audio data processing method and device and electronic equipment
CN113038053A (en) * 2021-03-30 2021-06-25 北京乐学帮网络技术有限公司 Data synthesis method and device, electronic equipment and storage medium
CN113593608B (en) * 2021-06-29 2022-06-24 北京荣耀终端有限公司 Object recognition-based voice beautifying method, electronic equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101888475A (en) * 2009-05-12 2010-11-17 虹软(杭州)多媒体信息技术有限公司 Photographic electronic device
CN103208298A (en) * 2012-01-11 2013-07-17 三星电子(中国)研发中心 Video shooting method and system
CN103268763A (en) * 2013-06-05 2013-08-28 广州市花都区中山大学国光电子与通信研究院 Wireless media system based on synchronous audio extraction and real-time transmission
CN104836961A (en) * 2015-05-13 2015-08-12 广州市久邦数码科技有限公司 Implementation method of real-time filter shooting based on Android system and system thereof
CN106157986A (en) * 2016-03-29 2016-11-23 联想(北京)有限公司 A kind of information processing method and device, electronic equipment
CN107172377A (en) * 2017-06-30 2017-09-15 福州瑞芯微电子股份有限公司 A kind of data processing method and device of video calling
CN107295284A (en) * 2017-08-03 2017-10-24 浙江大学 A kind of generation of video file being made up of audio and picture and index playing method, device

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9210352B2 (en) * 2012-09-28 2015-12-08 Key Digital Systems, Inc. Audio video instruction format switcher
CN104023192B (en) * 2014-06-27 2016-04-13 努比亚技术有限公司 A kind of method and apparatus of recorded video
EP3251116A4 (en) * 2015-01-30 2018-07-25 DTS, Inc. System and method for capturing, encoding, distributing, and decoding immersive audio
CN104967900B (en) * 2015-05-04 2018-08-07 腾讯科技(深圳)有限公司 A kind of method and apparatus generating video
CN105321526B (en) * 2015-09-23 2020-07-24 联想(北京)有限公司 Audio processing method and electronic equipment
CN105872253B (en) * 2016-05-31 2020-07-07 腾讯科技(深圳)有限公司 Live broadcast sound processing method and mobile terminal
US10116878B2 (en) * 2016-12-21 2018-10-30 Samsung Electronics Co., Ltd. Method for producing media file and electronic device thereof

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101888475A (en) * 2009-05-12 2010-11-17 虹软(杭州)多媒体信息技术有限公司 Photographic electronic device
CN103208298A (en) * 2012-01-11 2013-07-17 三星电子(中国)研发中心 Video shooting method and system
CN103268763A (en) * 2013-06-05 2013-08-28 广州市花都区中山大学国光电子与通信研究院 Wireless media system based on synchronous audio extraction and real-time transmission
CN104836961A (en) * 2015-05-13 2015-08-12 广州市久邦数码科技有限公司 Implementation method of real-time filter shooting based on Android system and system thereof
CN106157986A (en) * 2016-03-29 2016-11-23 联想(北京)有限公司 A kind of information processing method and device, electronic equipment
CN107172377A (en) * 2017-06-30 2017-09-15 福州瑞芯微电子股份有限公司 A kind of data processing method and device of video calling
CN107295284A (en) * 2017-08-03 2017-10-24 浙江大学 A kind of generation of video file being made up of audio and picture and index playing method, device

Also Published As

Publication number Publication date
CN108965757A (en) 2018-12-07

Similar Documents

Publication Publication Date Title
CN108965757B (en) Video recording method, device, terminal and storage medium
CN108401124B (en) Video recording method and device
CN109033335B (en) Audio recording method, device, terminal and storage medium
CN108965922B (en) Video cover generation method and device and storage medium
CN110688082B (en) Method, device, equipment and storage medium for determining adjustment proportion information of volume
CN109451343A (en) Video sharing method, apparatus, terminal and storage medium
CN109346111B (en) Data processing method, device, terminal and storage medium
CN108922506A (en) Song audio generation method, device and computer readable storage medium
CN111061405B (en) Method, device and equipment for recording song audio and storage medium
CN109192218B (en) Method and apparatus for audio processing
CN109144346B (en) Song sharing method and device and storage medium
CN109635133B (en) Visual audio playing method and device, electronic equipment and storage medium
CN111402844B (en) Song chorus method, device and system
CN108848394A (en) Net cast method, apparatus, terminal and storage medium
CN111083526B (en) Video transition method and device, computer equipment and storage medium
CN111142838A (en) Audio playing method and device, computer equipment and storage medium
CN110266982B (en) Method and system for providing songs while recording video
CN109743461B (en) Audio data processing method, device, terminal and storage medium
CN110808021B (en) Audio playing method, device, terminal and storage medium
CN111092991B (en) Lyric display method and device and computer storage medium
CN109819314B (en) Audio and video processing method and device, terminal and storage medium
CN109448676B (en) Audio processing method, device and storage medium
CN112118482A (en) Audio file playing method and device, terminal and storage medium
CN108966026B (en) Method and device for making video file
CN111294626A (en) Lyric display method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant