CN108055490B - Video processing method and device, mobile terminal and storage medium - Google Patents

Video processing method and device, mobile terminal and storage medium Download PDF

Info

Publication number
CN108055490B
CN108055490B CN201711009668.2A CN201711009668A CN108055490B CN 108055490 B CN108055490 B CN 108055490B CN 201711009668 A CN201711009668 A CN 201711009668A CN 108055490 B CN108055490 B CN 108055490B
Authority
CN
China
Prior art keywords
video data
audio
source
data
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711009668.2A
Other languages
Chinese (zh)
Other versions
CN108055490A (en
Inventor
刘飞跃
田东渭
贾松辉
郭伟
王程博
张志刚
杨玉奇
周朗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing environment and Wind Technology Co., Ltd.
Original Assignee
Beijing Mijinghefeng Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Mijinghefeng Technology Co ltd filed Critical Beijing Mijinghefeng Technology Co ltd
Priority to CN201711009668.2A priority Critical patent/CN108055490B/en
Publication of CN108055490A publication Critical patent/CN108055490A/en
Application granted granted Critical
Publication of CN108055490B publication Critical patent/CN108055490B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The embodiment of the invention provides a video processing method, a video processing device, a mobile terminal and a storage medium, and relates to the technical field of mobile communication. The video processing method is applied to the mobile terminal and comprises the following steps: acquiring source audio and video data; recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data; and comparing and displaying the target audio data and/or the target video data with the source audio and video data. The embodiment of the invention realizes the audio and video comparison function of the mobile terminal and meets the requirements of users.

Description

Video processing method and device, mobile terminal and storage medium
Technical Field
The present invention relates to the field of mobile communications technologies, and in particular, to a video processing method and apparatus, a terminal device, and a storage medium.
Background
With the development of mobile communication technology, mobile terminals such as mobile phones are more and more popular, and great convenience is brought to life, study and work of people.
These mobile terminals are typically provided with a camera so that the user can take pictures, record videos, etc. using the mobile terminal. In addition, these mobile terminals may also be installed with various applications, so that users may use the various applications in the mobile terminals to perform required operations, such as playing games through a game application, publishing or playing audio and video through a video application, and the like.
Disclosure of Invention
The invention provides a video processing method, and a corresponding video processing device, a mobile terminal and a storage medium, so as to realize the audio and video comparison function of the mobile terminal and meet the user requirements.
According to an aspect of the present invention, there is provided a video processing method applied to a mobile terminal, the method including: acquiring source audio and video data; recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data; and comparing and displaying the target audio data and/or the target video data with the source audio and video data.
Optionally, recording target audio data and/or target video data according to source audio data and/or source video data in the source audio/video data, including: separating source audio data from the source audio and video data, playing the source audio data, and recording target video data through a camera; and/or separating source video data from the source audio and video data, playing the source video data in an interface, and recording target audio data through a microphone.
Optionally, the recording target video data according to source audio data and/or source video data in the source audio and video data includes: and playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera.
Optionally, the playing source audio data and source video data in the source audio and video data in the interface, and recording target video data through a camera includes: dividing the interface into a first area and a second area; playing the source video data in the first area; and recording target video data through a camera, and displaying the recorded target video data in the second area.
Optionally, the comparing and displaying the target audio data and/or the target video data with the source audio and video data includes: synthesizing source audio data in the source audio and video data and recorded target video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data.
Optionally, the comparing and displaying the target audio data and/or the target video data with the source audio and video data includes: synthesizing the recorded target audio data and the target video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data.
Optionally, the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data includes: acquiring source image data of each frame corresponding to at least one segmentation time from the source video data; acquiring target image data of each frame corresponding to at least one segmentation time from the target video data; and synthesizing the source image data and the target image data of each frame according to the time sequence to obtain image data of each frame in the compared audio and video data.
Optionally, the method further comprises: and according to a preset rule, cross setting each section time corresponding to the source video data and each section time corresponding to the target video data.
Optionally, the comparing and displaying the target audio data and/or the target video data with the source audio and video data includes: synthesizing source video data and recorded target audio data in the source audio and video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data.
Optionally, the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data includes: acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data; acquiring a target audio fragment corresponding to at least one segmentation time from the target audio data; and synthesizing the source audio clips and the target audio clips according to the time sequence to obtain audio data in the comparison audio and video data.
Optionally, the comparing and displaying the target audio data and/or the target video data with the source audio and video data includes: providing a third region and a fourth region in the interface; and displaying the source video data in the third area and displaying the target video data in the fourth area to perform the display of the comparison video.
Optionally, the method further comprises: and according to a preset rule, cross setting each section time corresponding to the source audio data and each section time corresponding to the target audio data.
According to another aspect of the present invention, there is provided a video processing apparatus applied to a mobile terminal, the apparatus including: the acquisition module is used for acquiring source audio and video data; the recording module is used for recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data; and the comparison display module is used for comparing and displaying the target audio data and/or the target video data with the source audio and video data.
Optionally, the recording module includes: the video recording submodule is used for separating source audio data from the source audio and video data, playing the source audio data and recording target video data through a camera; and the audio recording submodule is used for separating source video data from the source audio and video data, playing the source video data in an interface and recording target audio data through a microphone.
Optionally, the recording module is specifically configured to play source audio data and source video data in the source audio and video data in an interface, and record target video data through a camera.
Optionally, the recording module includes: the interface dividing submodule is used for dividing the interface into a first area and a second area; a source video playing sub-module, configured to play the source video data in the first area; and the video recording submodule is used for recording target video data through a camera and displaying the recorded target video data in the second area.
Optionally, the comparative display module comprises: the target synthesis submodule is used for synthesizing the source audio data in the source audio and video data and the recorded target video data into target audio and video data; the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and the comparison display submodule is used for displaying the comparison audio and video data.
Optionally, the comparative display module comprises: the target synthesis submodule is used for synthesizing the recorded target audio data and the target video data into target audio and video data; the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and the comparison display submodule is used for displaying the comparison audio and video data.
Optionally, the contrast synthesis sub-module comprises: the source video acquisition unit is used for acquiring source image data of each frame corresponding to at least one segmentation time from the source video data; the target video acquisition unit is used for acquiring target image data of each frame corresponding to at least one segmentation time from the target video data; and the video synthesis unit is used for synthesizing the source image data and the target image data of each frame according to the time sequence to obtain the image data of each frame in the compared audio and video data.
Optionally, the method further comprises: and the setting module is used for alternately setting each section time corresponding to the source video data and each section time corresponding to the target video data according to a preset rule.
Optionally, the comparative display module comprises: the target synthesis submodule is used for synthesizing the source video data in the source audio and video data and the recorded target audio data into target audio and video data; the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and the comparison display submodule is used for displaying the comparison audio and video data.
Optionally, the contrast synthesis sub-module comprises: the source audio acquisition unit is used for acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data; the target audio acquiring unit is used for acquiring a target audio fragment corresponding to at least one segmentation time from the target video data; and the audio synthesis unit synthesizes the audio segments and the target audio segments according to the time sequence to obtain audio data in the comparison audio and video data.
Optionally, the comparative display module comprises: the area setting submodule is used for setting a third area and a fourth area in the interface; and the comparison display submodule is used for displaying the source image data of each frame in the source video data in the third area and displaying the target video data in the fourth area to display the comparison video.
Optionally, the method further comprises: and the setting module is used for alternately setting each section time corresponding to the source audio data and each section time corresponding to the target audio data according to a preset rule.
According to still another aspect of the present invention, there is provided a mobile terminal including: one or more processors; and one or more machine readable media having instructions stored thereon, which when executed by the one or more processors, cause the mobile terminal to perform a video processing method as described in one or more of the embodiments of the invention.
Embodiments of the present invention also provide a machine-readable medium having stored thereon instructions, which, when executed by one or more processors, cause a mobile terminal to perform a video processing method as described in one or more of the embodiments of the present invention.
According to the video processing method and device provided by the invention, the video processing method and device are applied to the mobile terminal, the target audio data and/or the target video data can be recorded according to the obtained source audio and video data, and then the recorded target audio data and/or the target video data can be compared with the source audio data for display, so that the audio and video comparison function of the mobile terminal is realized, a user can compare the recorded audio and/or video with the source audio and video by using the mobile terminal, and the beneficial effect of meeting the user requirements is obtained.
The foregoing description is only an overview of the technical solutions of the present invention, and the embodiments of the present invention are described below in order to make the technical means of the present invention more clearly understood and to make the above and other objects, features, and advantages of the present invention more clearly understandable.
Drawings
Various other advantages and benefits will become apparent to those of ordinary skill in the art upon reading the following detailed description of the preferred embodiments. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
FIG. 1 is a flow chart illustrating the steps of a video processing method according to one embodiment of the invention;
FIG. 2 is a flow chart illustrating the steps of a video processing method according to an alternative embodiment of the invention;
fig. 3 is a block diagram showing a configuration of a video processing apparatus according to an embodiment of the present invention;
fig. 4 is a block diagram showing a configuration of a video processing apparatus according to an alternative embodiment of the present invention;
FIG. 5 schematically shows a block diagram of a server for performing the method according to the invention;
fig. 6 schematically shows a storage unit for holding or carrying program code implementing a method according to the invention; and
fig. 7 is a block diagram illustrating a partial structure of a mobile terminal according to an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
The embodiment of the invention can be applied to the mobile terminal, and the mobile terminal can be provided with the video application, so that a user can process audio and video data through the video application. For example, a user may record a video, upload and distribute the video, etc. by using a video application in the mobile terminal, and may also obtain and play video data by using the video application, for example, obtain and play video data distributed by other users, etc.
The audio and video data can generally comprise video data and audio data; the video data can be used for playing video pictures corresponding to the audio and video data, and specifically can include one or more frames of image data, and the image data can be used for displaying the video pictures corresponding to the video data; the audio data may be used to play audio in video data, such as music in video, and the like.
In order to provide audio and video comparison service for a user, after acquiring source audio and video data, a mobile terminal in the implementation of the invention can record target audio data and/or target video data according to the source audio data and/or the source video data in the source audio and video data, and if the target video data can be recorded according to the source audio data in the source audio and video data, the user can record corresponding target audio and video according to the source audio data in the source audio and video data; and then, the recorded target audio data and/or target video data can be compared with the source audio and video data for display, so that the audio and video comparison function of the mobile terminal is realized, a user can compare the recorded target video and/or target audio with the obtained source audio and video by using the mobile terminal, and the audio and video comparison requirements of the user are met.
Referring to fig. 1, a flow diagram of the steps of a video processing method according to one embodiment of the invention is shown. The video processing method can be applied to a mobile terminal, and specifically comprises the following steps:
and 102, acquiring source audio and video data.
In the embodiment of the present invention, the source audio/video data may be used to represent the acquired audio/video, such as an audio/video pre-stored in the mobile terminal, an audio/video currently downloaded by the mobile terminal from the server, an audio/video searched by a video application in the mobile terminal, and the like, which is not limited in the implementation of the present invention.
And 104, recording target audio data and/or target video data according to the source audio data and/or the source video data in the source audio and video data.
In the embodiment of the invention, after the source audio and video data are acquired, the source video data and/or the source audio data in the source audio and video data can be played through the mobile terminal, so that a user can watch a video picture played corresponding to the source video data and/or can listen to the audio in the source audio and video data, for example, the mobile terminal or a video application in the mobile terminal can separate the source video data from the source audio and video data, and then the video data is played, so that the user can listen to the audio in the source audio and video data; similarly, source audio data can be separated from the source audio and video data, and then the audio data can be played; similarly, source audio data can be separated from the source audio and video data, and then the audio data is played, so that a user can watch a video picture played by the source video data correspondingly; for example, the source audio/video data can be directly played on the interface of the video application, and the like.
Specifically, during playing, a camera built in the mobile terminal and/or a camera connected to the mobile terminal can be started to record target video data through the started camera, so that a user can record target video to be shot aiming at currently played source audio data and/or source video data, and video comparison between the recorded target video data and source video data in source audio and video data is facilitated. The recorded target video data may be used to represent a video picture shot by a camera, and may specifically include target image data of each frame recorded in the playing process of the source audio/video data or the audio data in the source audio/video data.
Similarly, a microphone arranged in the mobile terminal and/or a microphone connected to the mobile terminal may also be started to record the target audio data through the started microphone, so that the user may record the target audio required to be recorded for the currently played source audio data and/or source video data, and the audio comparison between the recorded target audio data and the source audio data in the source audio data is facilitated.
And 106, comparing and displaying the target audio data and/or the target video data with the source audio and video data.
In the embodiment of the invention, the currently recorded target audio data and/or target video data can be compared with the source audio and video data to obtain corresponding comparison audio and video data, and if the source video data and the target video data can be compared, the corresponding comparison video data is generated; if the source audio data and the target audio data can be compared, corresponding comparison audio data is generated; for example, the target video data and/or the target audio/video data are synthesized with the source audio/video data to obtain the comparative audio/video data, and the like. And then, the comparison audio and video data can be displayed on an interface, so that the user can check the recorded target video and/or target audio and compare the result with the acquired source audio and video, and the audio and video comparison requirements of the user are met.
In an optional implementation manner, the embodiment of the present invention may synthesize currently recorded target video data and source audio data in source audio/video data to obtain synthesized target audio/video data. The target audio and video data can be used for representing a target audio and video recorded by a user based on source audio data in the source audio and video data, such as audio and video recorded by the user aiming at music in the source audio and video distributed by other users, and the like. And then, comparing the source audio and video data with the target audio and video data to generate corresponding comparison audio and video data, if the source audio and video data and the target audio and video data can be synthesized to obtain the comparison audio and video data, and then displaying the comparison audio and video data on an interface, so that a user can check the comparison result of the recorded target video and the obtained source video, and the video comparison requirement of the user is met.
In another optional implementation manner, the embodiment of the present invention may synthesize currently recorded target audio data and source video data in the source audio/video data to obtain synthesized target audio/video data. The target audio and video data can be used for representing a target audio and video recorded by a user based on source video data in the source audio and video data, such as audio and video recorded by a user aiming at video pictures in source audio and video published by other users, and the like. And then, comparing the source audio and video data with the target audio and video data to generate corresponding comparison audio and video data, if the source audio and video data and the target audio and video data can be synthesized to obtain the comparison audio and video data, and then displaying the comparison audio and video data on an interface, so that a user can obtain a comparison result of the recorded target audio and the obtained source audio, and the audio comparison requirement of the user is met.
In yet another optional implementation manner, the embodiment of the present invention may synthesize currently recorded target audio data and target video data to obtain synthesized target audio/video data. The target audio and video data can be used for representing a target audio and video recorded by the user aiming at the source audio and video data, such as a target audio and video recorded by the user aiming at source audio and video published by other users, and the like. And then, the source audio and video data and the target audio and video data can be compared to generate corresponding comparison audio and video data, and then the comparison audio and video data can be displayed on an interface, so that a user can obtain a comparison result of the recorded target audio and video and the obtained source audio and video, and the audio and video comparison requirements of the user are met.
In an optional embodiment of the present invention, the recording of the target video data and the target video data according to the source audio data and/or the source video data in the source audio/video data may include: separating source audio data from the source audio and video data, playing the source audio data, and recording target video data through a camera; and/or separating source video data from the source audio and video data, playing the source video data in an interface, and recording target audio data through a microphone.
Specifically, after the source audio and video data is acquired, the source audio data can be directly separated from the source audio and video data, then the separated source audio data is played, the source video data in the source audio and video data is not played, namely, the video picture corresponding to the source audio and video data is not displayed, so that a user can hear the audio corresponding to the audio data in the source audio and video data, the problem of blockage caused by playing the video picture corresponding to the source video data can be avoided, the smoothness of audio data playing is ensured, and the user experience is improved. When the source audio data is played, each frame of target image data corresponding to the video image currently shot by the user can be recorded through the camera, that is, the target video data correspondingly recorded aiming at the currently played source audio data can be obtained, so that the user can record the corresponding target video data in the audio data playing process, and the source audio data and the recorded target video data can be conveniently synthesized into the target audio and video data. For example, source audio data in source audio and video data can be synthesized with target image data of each frame recorded currently to obtain synthesized target audio and video data, that is, a target audio and video recorded by a user aiming at the source audio data in the source audio and video data is generated, so that the target audio and video recorded by the user can be compared with the source audio and video to meet the video comparison requirement of the user.
Similarly, after the source audio and video data is acquired, the source video data can be directly separated from the source audio and video data, and then the separated source video data is played without playing the source audio data in the source audio and video data, that is, the video picture corresponding to the source audio and video data is not displayed, so that a user can watch the video picture corresponding to the source video data on an interface, and meanwhile, the problem of unsmooth playing caused by playing the source audio data in the source audio and video data can be avoided, the smoothness of playing the video data is ensured, and the user experience is improved. When the source video data is played, each frame of target audio data corresponding to the target audio currently recorded by the user can be recorded through the microphone, that is, the target audio data corresponding to the currently played source video data can be obtained, so that the user can record the corresponding target audio data in the video data playing process, and the source video data and the recorded target audio data can be conveniently synthesized into the target audio and video data. For example, source video data in source audio and video data can be synthesized with target audio data of each frame recorded currently to obtain synthesized target audio and video data, that is, a target audio and video recorded by a user aiming at the source video data in the source audio and video data is generated, so that the target audio and video recorded by the user can be compared with the source audio and video to meet the audio comparison requirement of the user.
Of course, after the source video data and the source audio data are separated from the source audio and video data, corresponding target video data can be recorded when the source audio data are played, and corresponding target video data can be recorded when the source video data are played; or recording corresponding target video data and target audio data when playing the source audio data; the corresponding target video data and target audio data may also be recorded when the source video data is played, and the like, which is not limited in this embodiment of the present invention.
In the actual processing, optionally, after the source audio and video data is obtained, the source audio and video data may not be separated from the source audio and video data for playing, but the source audio and video data is directly played on an interface of a mobile terminal or a video application, so that a user can watch a corresponding video picture on the interface while listening to audio in the source audio and video data, and simultaneously can record each frame of target image data corresponding to a video picture currently shot by the user through a camera, so that the user can record corresponding target video data in the audio and video data playing process, and can record each frame of target audio data corresponding to a target audio currently recorded by the user through a microphone, so that the user can record corresponding target audio data in the audio and video data playing process.
Referring to fig. 2, a flowchart illustrating steps of a video processing method according to another embodiment of the present invention is shown, which may specifically include the following steps:
and step 202, acquiring source audio and video data.
And 204, playing source audio data and source video data in the source audio and video data in an interface, and recording target video data and/or target audio data.
In an optional embodiment of the present invention, recording target video data according to source audio data and/or source video data in the source audio/video data may include: and playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera. That is, the source audio/video data is played in the interface, and the target video data can be recorded through the camera, so that the target audio/video data can be synthesized according to the recorded target video data in the following process, for example, the audio data in the source audio/video data and the recorded target video data can be synthesized into the target video data.
Of course, in the process of playing the source audio-visual data, optionally, the embodiment of the present invention may also record the target audio-visual data through the microphone, so as to synthesize the target audio-visual data according to the recorded target audio-visual data in the following. In another optional embodiment of the present invention, recording target audio data according to source audio data and/or source video data in the source audio and video data may include: and playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a microphone.
For example, in the process of using the mobile terminal by the user, the video application in the mobile terminal can be started to search the video application for source audio and video data published or uploaded by other users. After the video application acquires the source audio and video data, the source audio and video data can be played on an interface, so that a user can watch the source audio and video data issued by other users; meanwhile, target video data can be recorded through a camera in the mobile terminal, so that a user can record each frame of image data corresponding to a video picture required to be shot in the playing process of source audio and video data, and can record target audio data through a microphone, and further the user can record target audio and target video simultaneously in the playing process of the source audio and video data.
In an optional implementation manner of the present invention, playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera may include: dividing the interface into a first area and a second area; playing the source audio and video data in the first area; and recording target video data through a camera, and displaying the recorded target video data in the second area. Specifically, the interface of the video application can be divided into two display areas, one of the display areas can be referred to as a first area to play the acquired source audio/video data in the first area, and the other display area can be referred to as a second area to display a video picture corresponding to the target video data recorded by the camera in the second area. For example, after the video application acquires the source audio and video data, the source audio and video data can be played in the first area, and simultaneously, the target video data recorded by the camera at present can be acquired, so that the video picture corresponding to the currently recorded target video data can be displayed in the second area, and a user can watch the source audio and video data and the video picture corresponding to the recorded target video data at the same time.
And step 206, comparing and displaying the target audio data and/or the target video data with the source audio and video data.
In an optional embodiment of the present invention, the comparing and displaying the target audio data and/or the target video data with the source audio-video data may include: synthesizing source audio data in the source audio and video data and recorded target video data into target video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data.
In the embodiment of the present invention, optionally, after the source audio data and the recorded target video data are synthesized into the target audio/video data, the source audio/video data and the target audio/video data may be compared, for example, the source audio/video data and the target audio/video data may be synthesized, or for example, the source audio/video data and the target video data may be synthesized by a video application in the mobile terminal to obtain the comparative audio/video data; and then displaying the comparative audio and video data, so that the user can check the comparative result between the recorded target video and the acquired source video. For example, each frame of source image data in the source video data and each frame of target image data in the target video data can be simultaneously checked on the interface, so that a user can simultaneously check a video picture corresponding to each frame of source image data and a video picture corresponding to each frame of target image data, and comparison between the source video and the target video is realized; as another example, source image data in source video data and destination image data in destination video data corresponding to the same audio segment may be presented across the interface in a chronological order, such that a user may compare video pictures of source video data and video pictures of destination video data corresponding to the same audio segment, and so on.
Of course, the target audio and video data may also be synthesized in other manners, and the synthesized target audio and video data is used for audio/video comparison, for example, the recorded target audio data and the target video data may be used to synthesize the target audio and video data, or the recorded target audio data and the source video data may be used to synthesize the target audio and video data, and the like, which is not limited in this embodiment of the present invention.
In another optional embodiment of the present invention, the comparing and displaying the target audio data and/or the target video data with the source audio-video data may include: synthesizing source video data and recorded target audio data in the source audio and video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data. For example, source audio data and target audio data corresponding to source image data of each frame in the same source video segment can be cross-played in the mobile terminal according to a time sequence, so that a user can compare source audio and target audio corresponding to the same video segment, thereby realizing audio comparison between the source audio and the target audio and further meeting the audio comparison requirement of the user.
In yet another optional embodiment of the present invention, the comparing and presenting the target audio data and/or the target video data with the source audio-video data may include: synthesizing the recorded target audio data and the target video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data. Therefore, the audio-video comparison between the source audio-video and the target audio-video is realized, so that the user can simultaneously compare the target audio and the target video recorded by the user with the source audio-video, namely, the audio-video comparison requirement of the user is met.
In an optional implementation manner, the comparative audio/video data in the embodiment of the present invention may include: the audio data and the source image data of each frame in the source audio and video data and the target image data of each frame in the target video data can be displayed on an interface in the playing process of comparing the audio and video data, so that the video picture corresponding to the image data of each frame in the source video data and the video picture corresponding to the target image data of each frame in the target video data can be displayed on the interface. For example, comparing audio-visual data may include: acquiring source video data, source audio data and target video data; as another example, comparing audio-visual data may include: the source audio and video data includes source video data, recorded target audio data, and target video data, among others.
In an optional embodiment of the present invention, the comparing and displaying the target audio data and/or the target video data with the source audio/video data may include: setting a third area and a fourth area on the interface; and displaying the source image data of each frame in the source video data in the third area and displaying the target video data in the fourth area to display the comparison video. For example, a third area and a fourth area may be set on an interface of a video application, so that in the process of displaying and comparing video data, a video picture corresponding to each frame of source image data in source video data is displayed in the third area, a video picture corresponding to each frame of target image data in target video data is displayed in the fourth area, and simultaneously audio data in source audio and video data may be played, so that a user may compare a source video and a target video corresponding to the same audio data.
Optionally, in the embodiment of the present invention, each segment time corresponding to the source video data and each segment time corresponding to the target video data may be set in a crossed manner according to a preset rule, so that in a process of synthesizing comparison audio/video data, each frame of source image data and each frame of target image data may be synthesized according to a time sequence to obtain each frame of image data in the comparison audio/video data. The preset rule may be set according to source audio data in the source audio/video data, or may also be set according to target audio data, which is not limited in this embodiment of the present invention.
In an optional example, the source audio data in the source audio and video data may be divided according to a preset rule, for example, according to the playing pause time of the audio data, and the playing time duration corresponding to each divided segment of audio data may be determined. For example, in the case of dividing the source audio data into 3 pieces of segmented audio data, the playing time periods of the 3 pieces of segmented audio data may be equal, such as the playing time periods of the 3 pieces of segmented audio data are all 30 seconds; the playing time periods of the 3 segments of audio data may also be unequal, such as the playing time period of the first segment of audio data may be 20 seconds, the playing time period of the second segment of audio data may be 25 seconds, the playing time period of the third segment of audio data may be 40 seconds, and so on. Therefore, based on the playing duration corresponding to each segmented audio data, each segmented time corresponding to the source video data and each segmented time corresponding to the target video data can be set in an intersecting manner, for example, in combination with the first example, the first segmented time corresponding to the source video data is set to be 0-30 seconds after the audio and video data is started, then the first segmented time corresponding to the target video data is set to be 30-60 seconds after the audio and video data is started, then the second segmented time corresponding to the source video data is set to be 60-90 seconds after the audio and video data is started, the second segmented time corresponding to the target video data is set to be 90-120 seconds after the audio and video data is started, and so on until the segmented times corresponding to the source video and the target video are set.
In an optional embodiment of the present invention, the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data may include: acquiring source image data of each frame corresponding to at least one segmentation time from the source video data; acquiring target image data of each frame corresponding to at least one segmentation time from the target video data; and synthesizing the source image data and the target image data of each frame according to the time sequence to obtain image data of each frame in the compared audio and video data. Then, each frame of image data in the comparison audio and video data can be synthesized with the audio data, for example, the comparison audio and video data can be synthesized with one or more corresponding source audio segments in the source audio and video data, can also be synthesized with one or more corresponding target audio segments in the target audio and video data, and the like, so as to obtain the corresponding comparison audio and video data.
According to the embodiment of the invention, according to the requirements of users, such as the playing duration of the comparative audio/video data set by the users, the source image data of each frame corresponding to one or more segment times is obtained from the source video data, the target image data of each frame corresponding to one or more segment times is obtained from the target video data, and then the source image data and the target image data of each frame are synthesized according to the time sequence to obtain the comparative audio/video data. Of course, source image data and target image data of each frame corresponding to the corresponding segment time may also be obtained from the source audio/video data and the target audio/video data respectively according to an audio data segment cut by the user in the source audio data of the source audio/video data, so as to perform video comparison.
Of course, the embodiment of the present invention may also set each segment time corresponding to the source audio data and each segment time corresponding to the target audio data in a crossed manner according to a preset rule, so that in the process of synthesizing the comparison audio/video data, each source audio clip and each target audio clip may be synthesized according to a time sequence to obtain the audio data in the comparison audio/video data. The preset rule may be set according to source video data in the source audio/video data, or may also be set according to target video data, which is not limited in this embodiment of the present invention.
In an optional example, the source video data in the source audio/video data may be divided according to a preset rule, for example, according to the playing time of the video data, and the playing time corresponding to each divided segment of video data is determined. For example, in the case of dividing the video data into 3 video segments, the playing time lengths of the 3 video segments may be equal, for example, the playing time lengths of the 3 video segments are all 30 seconds; the playing time lengths of the 3 video segments may also be unequal, for example, the playing time length of the first video segment may be 20 seconds, the playing time length of the second video segment may be 25 seconds, the playing time length of the third video segment may be 40 seconds, and so on. Therefore, based on the playing duration corresponding to each video segment, each segment time corresponding to the source audio data and each segment time corresponding to the target audio data can be set in an intersecting manner, for example, in combination with the first example, the first segment time corresponding to the source audio data is set to be 0-30 seconds after the audio and video data is compared and started, then the first segment time corresponding to the target audio data is set to be 30-60 seconds after the audio and video data is compared and started, then the second segment time corresponding to the source audio data is set to be 60-90 seconds after the audio and video data is compared and started, the second segment time corresponding to the target audio data is set to be 90-120 seconds after the audio and video data is compared and started, and so on until the segment times corresponding to the source audio and the target audio in the audio and video data are compared and set.
In an optional embodiment of the present invention, the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data may include: acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data; acquiring a target audio fragment corresponding to at least one segmentation time from the target audio data; and synthesizing the source audio clips and the target audio clips according to the time sequence to obtain audio data in the comparison audio and video data. Then, the audio data in the comparison audio and video data and each frame of image data can be synthesized, for example, the audio data can be synthesized with one or more corresponding source video segments in the source audio and video data, or can be synthesized with one or more corresponding target video segments in the target audio and video data, and the like, so as to obtain the corresponding comparison audio and video data.
According to the embodiment of the invention, according to the requirements of users, such as the playing time length of the comparative audio/video data set by the users, the source audio segments corresponding to one or more segment times are obtained from the source audio data, the target audio segments corresponding to one or more segment times are obtained from the target audio data, and then the obtained source audio segments and the target audio segments are synthesized according to the time sequence to obtain the comparative audio/video data. Of course, according to one or a pair of video segments cut by the user in the source audio data of the source audio/video data, the source audio segment and the target audio segment corresponding to the corresponding segment time may be obtained from the source audio/video data and the target audio/video data, respectively, so as to perform audio comparison.
According to the embodiment of the present invention, for example, according to the playing duration of the comparative audio/video data set by the user, the source audio/video clips (including the source audio clip and the source video clip) corresponding to one or more segment times are obtained from the source audio/video data, and the target audio/video clips (including the target audio clip and the target video clip) corresponding to one or more segment times are obtained from the target audio/video data, and then the obtained source audio/video clips and the obtained target audio/video clips are synthesized according to the time sequence to obtain the comparative audio/video data, so as to perform audio/video comparison according to the comparative audio/video data, which is not limited in this embodiment of the present invention.
For simplicity of explanation, the method embodiments are described as a series of acts or combinations, but those skilled in the art will appreciate that the embodiments are not limited by the order of acts described, as some steps may occur in other orders or concurrently with other steps in accordance with the embodiments of the invention. Further, those skilled in the art will appreciate that the embodiments described in the specification are presently preferred and that no particular act is required to implement the invention.
Referring to fig. 3, a block diagram of an embodiment of a video processing apparatus according to an embodiment of the present invention is shown. The video processing device can be applied to a mobile terminal, and specifically comprises the following modules:
an obtaining module 302, configured to obtain source audio and video data;
the recording module 304 is configured to record target audio data and/or target video data according to source audio data and/or source video data in the source audio/video data;
and the comparison display module 306 is configured to perform comparison display on the target audio data and/or the target video data and the source audio/video data.
To sum up, after the source audio and video data is acquired, the embodiment of the invention can record the target audio data and/or the target video data according to the acquired source audio and video data, and then can compare and display the recorded target audio data and/or the target video data with the source audio data, thereby realizing the audio and video comparison function of the mobile terminal, enabling a user to compare the recorded audio and/or video with the source audio and video by using the mobile terminal, and obtaining the beneficial effect of meeting the user requirements.
Referring to fig. 4, a block diagram of a video processing apparatus according to an alternative embodiment of the present invention is shown.
In this embodiment of the present invention, optionally, the recording module 304 may include the following sub-modules:
a video recording submodule 3042, configured to separate source audio data from the source audio and video data, play the source audio data, and record target video data through a camera;
the audio recording sub-module 3044 is configured to separate the source video data from the source audio/video data, play the source video data in an interface, and record the target audio data through a microphone.
In an optional embodiment of the present invention, the recording module 304 is specifically configured to play source audio data and source video data in the source audio and video data in an interface, and record target video data through a camera.
In an optional embodiment of the present invention, the recording module 304 may include the following sub-modules:
an interface dividing submodule 3046 for dividing the interface into a first area and a second area;
a source video playing sub-module 3048 configured to play the source video data in the first area;
the video recording sub-module 3042 is configured to record target video data through a camera and display the recorded target video data in the second area.
In an alternative embodiment of the present invention, the comparative display module 306 may include the following sub-modules:
the target synthesis submodule 3062 is configured to synthesize the source audio data in the source audio and video data and the recorded target video data into target audio and video data;
the comparison synthesis submodule 3064 is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule 3066 is used for displaying the comparison audio and video data.
In another alternative embodiment of the present invention, the comparative display module 306 may include the following sub-modules:
the target synthesis submodule 3062 is used for synthesizing the recorded target audio data and the target video data into target audio and video data;
the comparison synthesis submodule 3064 is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule 3066 is used for displaying the comparison audio and video data.
In the embodiment of the present invention, optionally, the comparison synthesis submodule 3064 may include the following units:
the source video acquisition unit is used for acquiring source image data of each frame corresponding to at least one segmentation time from the source video data;
the target video acquisition unit is used for acquiring target image data of each frame corresponding to at least one segmentation time from the target video data;
and the video synthesis unit is used for synthesizing the source image data and the target image data of each frame according to the time sequence to obtain the image data of each frame in the compared audio and video data.
In an optional embodiment of the present invention, further comprising: the setting module 308 is configured to set, according to a preset rule, each segment time corresponding to the source video data and each segment time corresponding to the target video data in a crossing manner.
In yet another alternative embodiment of the present invention, the comparative display module 306 may include the following sub-modules:
the target synthesis submodule 3062 is configured to synthesize the source video data and the recorded target audio data in the source audio and video data into target audio and video data;
the comparison synthesis submodule 3064 is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule 3066 is used for displaying the comparison audio and video data.
In the embodiment of the present invention, optionally, the comparison synthesis submodule 3064 may include the following units:
the source audio acquisition unit is used for acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data;
the target audio acquiring unit is used for acquiring a target audio fragment corresponding to at least one segmentation time from the target video data;
and the audio synthesis unit synthesizes the audio segments and the target audio segments according to the time sequence to obtain audio data in the comparison audio and video data.
In an alternative embodiment of the present invention, the comparative display module 306 may include the following sub-modules:
a region setting submodule 3068 for setting a third region and a fourth region in the interface;
the comparison and display submodule 3066 is configured to display a comparison video by displaying the source image data of each frame in the source video data in the third area and displaying the target video data in the fourth area.
In the embodiment of the present invention, optionally, the method further includes: the setting module 308 is configured to set, according to a preset rule, each segment time corresponding to the source audio data and each segment time corresponding to the target audio data in a crossing manner.
In summary, in the implementation of the present invention, after the source audio/video data is acquired, the target audio data and/or the target video data can be recorded according to the audio data and/or the source video data in the source audio/video data, so that a user can record a corresponding target audio/video for the source audio/video data; and then, the recorded target audio data and/or target video data can be compared with the source audio and video data for display, so that the audio and video comparison function of the mobile terminal is realized, a user can compare the recorded target video and/or target audio with the obtained source audio and video by using the mobile terminal, and the audio and video comparison requirements of the user are met.
For the device embodiment, since it is basically similar to the method embodiment, the description is simple, and for the relevant points, refer to the partial description of the method embodiment.
The various component embodiments of the invention may be implemented in hardware, or in software modules running on one or more processors, or in a combination thereof. Those skilled in the art will appreciate that a microprocessor or Digital Signal Processor (DSP) may be used in practice to implement some or all of the functionality of some or all of the components in an electronic device according to embodiments of the present invention. The present invention may also be embodied as apparatus or device programs (e.g., computer programs and computer program products) for performing a portion or all of the methods described herein. Such programs implementing the present invention may be stored on computer-readable media or may be in the form of one or more signals. Such a signal may be downloaded from an internet website or provided on a carrier signal or in any other form. The electronic devices may include servers (clusters), mobile terminals, etc.
An embodiment of the present invention provides a server, including: one or more processors; and one or more machine readable media having instructions stored thereon that, when executed by the one or more processors, cause the server to perform a plug-in backup method as described in one or more of the embodiments of the invention.
Embodiments of the present invention provide one or more machine-readable media having instructions stored thereon, which when executed by one or more processors, cause a server to perform a video processing method as described in one or more of embodiments of the present invention.
For example, fig. 5 shows a server, such as a management server, a storage server, an application server, a cloud control service server cluster, and the like, which can implement the method according to the present invention. The server conventionally includes a processor 510 and a computer program product or computer-readable medium in the form of a memory 520. The memory 520 may be an electronic memory such as a flash memory, an EEPROM (electrically erasable programmable read only memory), an EPROM, a hard disk, or a ROM. The memory 520 has a memory space 530 for program code 531 for performing any of the method steps in the method described above. For example, the storage space 530 for the program code may include respective program codes 531 for implementing various steps in the above method, respectively. The program code can be read from or written to one or more computer program products. These computer program products comprise a program code carrier such as a hard disk, a Compact Disc (CD), a memory card or a floppy disk. Such a computer program product is typically a portable or fixed storage unit as described with reference to fig. 6. The storage unit may have a storage section, a storage space, and the like arranged similarly to the memory 520 in the server of fig. 5. The program code may be compressed, for example, in a suitable form. Typically, the storage unit comprises computer readable code 531', i.e. code that can be read by a processor, such as 510, for example, which when executed by a server causes the server to perform the steps of the method described above.
An embodiment of the present invention further provides a mobile terminal, including: one or more processors; and one or more machine readable media having instructions stored thereon, which when executed by the one or more processors, cause the mobile terminal to perform a video processing method as described in one or more of the embodiments of the invention.
Also provided in an embodiment of the present invention is a machine-readable medium having stored thereon instructions, which, when executed by one or more processors, cause a mobile terminal to perform a video processing method as described in one or more of embodiments of the present invention.
As shown in fig. 7, for convenience of description, only the parts related to the embodiment of the present invention are shown, and details of the specific technology are not disclosed, please refer to the method part in the embodiment of the present invention. The mobile terminal may be any device including a mobile phone, a tablet computer, a PDA (Personal Digital Assistant), a POS (Point of Sales), a vehicle-mounted computer, and the like.
Fig. 7 is a block diagram illustrating a partial structure of a mobile terminal according to an embodiment of the present invention. Referring to fig. 7, the mobile terminal includes: radio Frequency (RF) circuit 710, memory 720, input unit 730, display unit 740, sensor 750, audio circuit 760, wireless fidelity (WiFi) module 770, processor 780, power supply 790 and camera 7110. Those skilled in the art will appreciate that the mobile terminal architecture shown in fig. 7 is not intended to be limiting of mobile terminals and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components.
The following describes each component of the mobile terminal in detail with reference to fig. 7:
the RF circuit 710 may be used for receiving and transmitting signals during information transmission and reception or during a call, and in particular, receives downlink information of a base station and then processes the received downlink information to the processor 780; in addition, the data for designing uplink is transmitted to the base station. In general, the RF circuit 710 includes, but is not limited to, an antenna, at least one Amplifier, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like. In addition, the RF circuit 710 may also communicate with networks and other devices via wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to Global System for Mobile communication (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), email, Short Messaging Service (SMS), and the like.
The memory 720 may be used to store software programs and modules, and the processor 780 performs various functional applications and data processing of the mobile terminal by operating the software programs and modules stored in the memory 720. The memory 720 may mainly include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required by at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the mobile terminal, and the like. Further, the memory 720 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
The input unit 730 may be used to receive input numeric or character information and generate key signal inputs related to user settings and function control of the mobile terminal. Specifically, the input unit 730 may include a touch panel 731 and other input devices 732. The touch panel 731, also referred to as a touch screen, can collect touch operations of a user (e.g. operations of the user on or near the touch panel 731 by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a preset program. Alternatively, the touch panel 731 may include two portions of a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts it to touch point coordinates, and sends the touch point coordinates to the processor 780, and can receive and execute commands from the processor 780. In addition, the touch panel 731 may be implemented by various types, such as a resistive type, a capacitive type, an infrared ray, and a surface acoustic wave. The input unit 730 may include other input devices 732 in addition to the touch panel 731. In particular, other input devices 732 may include, but are not limited to, one or more of a physical keyboard, function keys (such as volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like.
The display unit 740 may be used to display information input by the user or information provided to the user and various menus of the mobile terminal. The Display unit 740 may include a Display panel 741, and optionally, the Display panel 741 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like. Further, the touch panel 731 can cover the display panel 741, and when the touch panel 731 detects a touch operation on or near the touch panel 731, the touch operation is transmitted to the processor 780 to determine the type of the touch event, and then the processor 780 provides a corresponding visual output on the display panel 741 according to the type of the touch event. Although the touch panel 731 and the display panel 741 are two separate components to implement the input and output functions of the mobile terminal in fig. 7, in some embodiments, the touch panel 731 and the display panel 741 may be integrated to implement the input and output functions of the mobile terminal.
The mobile terminal may also include at least one sensor 750, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor may include an ambient light sensor that may adjust the brightness of the display panel 741 according to the brightness of ambient light, and a proximity sensor that may turn off the display panel 741 and/or a backlight when the mobile terminal is moved to the ear. As one of the motion sensors, the accelerometer sensor can detect the magnitude of acceleration in each direction (generally, three axes), detect the magnitude and direction of gravity when stationary, and can be used for applications (such as horizontal and vertical screen switching, related games, magnetometer attitude calibration) for recognizing the attitude of the mobile terminal, and related functions (such as pedometer and tapping) for vibration recognition; as for other sensors such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured on the mobile terminal, further description is omitted here.
The audio circuit 760, speaker 761, and microphone 762 may provide an audio interface between a user and the mobile terminal. The audio circuit 760 can transmit the electrical signal converted from the received audio data to the speaker 761, and the electrical signal is converted into a sound signal by the speaker 761 and output; on the other hand, the microphone 762 converts the collected sound signal into an electric signal, converts the electric signal into audio data after being received by the audio circuit 760, processes the audio data by the audio data output processor 780, and transmits the processed audio data to, for example, another mobile terminal via the RF circuit 710, or outputs the audio data to the memory 720 for further processing.
WiFi belongs to short-distance wireless transmission technology, and the mobile terminal can help a user to receive and send e-mails, browse webpages, access streaming media and the like through the WiFi module 770, and provides wireless broadband Internet access for the user. Although fig. 7 shows the WiFi module 770, it is understood that it does not belong to the essential constitution of the mobile terminal, and can be omitted entirely as needed within the scope not changing the essence of the invention.
The processor 780 is a control center of the mobile terminal, connects various parts of the entire mobile terminal using various interfaces and lines, and performs various functions of the mobile terminal and processes data by operating or executing software programs and/or modules stored in the memory 720 and calling data stored in the memory 720, thereby integrally monitoring the mobile terminal. Optionally, processor 780 may include one or more processing units; preferably, the processor 780 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 780.
The mobile terminal also includes a power supply 790 (e.g., a battery) for powering the various components, which may preferably be logically coupled to the processor 780 through a power management system that may enable management of charging, discharging, and power consumption management functions.
The camera 7110 may perform a photographing function.
Although not shown, the mobile terminal may further include a bluetooth module, etc., which will not be described herein.
The algorithms and displays presented herein are not inherently related to any particular computer, virtual machine, or other apparatus. Various general purpose systems may also be used with the teachings herein. The required structure for constructing such a system will be apparent from the description above. Moreover, the present invention is not directed to any particular programming language. It is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any descriptions of specific languages are provided above to disclose the best mode of the invention.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the detailed description are hereby expressly incorporated into this detailed description, with each claim standing on its own as a separate embodiment of this invention.
Those skilled in the art will appreciate that the modules in the device in an embodiment may be adaptively changed and disposed in one or more devices different from the embodiment. The modules or units or components of the embodiments may be combined into one module or unit or component, and furthermore they may be divided into a plurality of sub-modules or sub-units or sub-components. All of the features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or elements of any method or apparatus so disclosed, may be combined in any combination, except combinations where at least some of such features and/or processes or elements are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings) may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
Furthermore, those skilled in the art will appreciate that while some embodiments described herein include some features included in other embodiments, rather than other features, combinations of features of different embodiments are meant to be within the scope of the invention and form different embodiments. For example, in the following claims, any of the claimed embodiments may be used in any combination.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The usage of the words first, second and third, etcetera do not indicate any ordering. These words may be interpreted as names.
The invention discloses A1 and a video processing method, which is applied to a mobile terminal and comprises the following steps:
acquiring source audio and video data;
recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data;
and comparing and displaying the target audio data and/or the target video data with the source audio and video data.
A2, the method as in a1, recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data, including:
separating source audio data from the source audio and video data, playing the source audio data, and recording target video data through a camera; and/or the presence of a gas in the gas,
separating source video data from the source audio and video data, playing the source video data in an interface, and recording target audio data through a microphone.
A3, the method as in a1, wherein the recording target video data according to source audio data and/or source video data in the source audio/video data includes:
and playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera.
A4, the method as in A3, where playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera, includes:
dividing the interface into a first area and a second area;
playing the source video data in the first area;
and recording target video data through a camera, and displaying the recorded target video data in the second area.
A5, the method as recited in a1, wherein the displaying the target audio data and/or the target video data in comparison with the source audio-video data includes:
synthesizing source audio data in the source audio and video data and recorded target video data into target audio and video data;
synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and displaying the comparative audio and video data.
A6, the method as recited in a1, wherein the displaying the target audio data and/or the target video data in comparison with the source audio-video data includes:
synthesizing the recorded target audio data and the target video data into target audio and video data;
synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and displaying the comparative audio and video data.
A7, the method described in a5 or a6, where the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data includes:
acquiring source image data of each frame corresponding to at least one segmentation time from the source video data;
acquiring target image data of each frame corresponding to at least one segmentation time from the target video data;
and synthesizing the source image data and the target image data of each frame according to the time sequence to obtain image data of each frame in the compared audio and video data.
A8, the method of a7, further comprising:
and according to a preset rule, cross setting each section time corresponding to the source video data and each section time corresponding to the target video data.
A9, the method as recited in a1, wherein the displaying the target audio data and/or the target video data in comparison with the source audio-video data includes:
synthesizing source video data and recorded target audio data in the source audio and video data into target audio and video data;
synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and displaying the comparative audio and video data.
A10, the method described in a6 or a9, where the synthesizing the source audio/video data and the target audio/video data to obtain the comparison audio/video data includes:
acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data;
acquiring a target audio fragment corresponding to at least one segmentation time from the target audio data;
and synthesizing the source audio clips and the target audio clips according to the time sequence to obtain audio data in the comparison audio and video data.
A11, the method as recited in a1, wherein the displaying the target audio data and/or the target video data in comparison with the source audio-video data includes:
providing a third region and a fourth region in the interface;
and displaying the source video data in the third area and displaying the target video data in the fourth area to perform the display of the comparison video.
A12, the method of a10, further comprising:
and according to a preset rule, cross setting each section time corresponding to the source audio data and each section time corresponding to the target audio data.
The invention also discloses B13 and a video processing device, which is applied to a mobile terminal, and the device comprises:
the acquisition module is used for acquiring source audio and video data;
the recording module is used for recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data;
and the comparison display module is used for comparing and displaying the target audio data and/or the target video data with the source audio and video data.
B14, the apparatus of B13, the recording module comprising:
the video recording submodule is used for separating source audio data from the source audio and video data, playing the source audio data and recording target video data through a camera;
and the audio recording submodule is used for separating source video data from the source audio and video data, playing the source video data in an interface and recording target audio data through a microphone.
B15, device according to B13,
the recording module is specifically configured to play source audio data and source video data in the source audio and video data in an interface, and record target video data through a camera.
B16, the apparatus of B15, the recording module comprising:
the interface dividing submodule is used for dividing the interface into a first area and a second area;
a source video playing sub-module, configured to play the source video data in the first area;
and the video recording submodule is used for recording target video data through a camera and displaying the recorded target video data in the second area.
B17, the device of B13, the comparative display module comprising:
the target synthesis submodule is used for synthesizing the source audio data in the source audio and video data and the recorded target video data into target audio and video data;
the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule is used for displaying the comparison audio and video data.
B18, the device of B13, the comparative display module comprising:
the target synthesis submodule is used for synthesizing the recorded target audio data and the target video data into target audio and video data;
the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule is used for displaying the comparison audio and video data.
B19, the device of B17 or B18, the contrast synthesis submodule comprising:
the source video acquisition unit is used for acquiring source image data of each frame corresponding to at least one segmentation time from the source video data;
the target video acquisition unit is used for acquiring target image data of each frame corresponding to at least one segmentation time from the target video data;
and the video synthesis unit is used for synthesizing the source image data and the target image data of each frame according to the time sequence to obtain the image data of each frame in the compared audio and video data.
B20, the apparatus of B19, further comprising:
and the setting module is used for alternately setting each section time corresponding to the source video data and each section time corresponding to the target video data according to a preset rule.
B21, the device of B13, the comparative display module comprising:
the target synthesis submodule is used for synthesizing the source video data in the source audio and video data and the recorded target audio data into target audio and video data;
the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule is used for displaying the comparison audio and video data.
B22, the device of B18 or B21, the contrast synthesis submodule comprising:
the source audio acquisition unit is used for acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data;
the target audio acquiring unit is used for acquiring a target audio fragment corresponding to at least one segmentation time from the target video data;
and the audio synthesis unit synthesizes the audio segments and the target audio segments according to the time sequence to obtain audio data in the comparison audio and video data.
B23, the device of B13, the comparative display module comprising:
the area setting submodule is used for setting a third area and a fourth area in the interface;
and the comparison display submodule is used for displaying the source image data of each frame in the source video data in the third area and displaying the target video data in the fourth area to display the comparison video.
B24, the apparatus of B22, further comprising:
and the setting module is used for alternately setting each section time corresponding to the source audio data and each section time corresponding to the target audio data according to a preset rule.
The invention also discloses C25, a mobile terminal, comprising: one or more processors; and
one or more machine readable media having instructions stored thereon that, when executed by the one or more processors, cause the mobile terminal to perform a video processing method as described in one or more of a1-a 12.
The present invention also discloses D26, a machine-readable medium having instructions stored thereon, which when executed by one or more processors, cause a mobile terminal to perform a video processing method as described in one or more of a1-a 12.

Claims (20)

1. A video processing method is applied to a mobile terminal, and the method comprises the following steps:
acquiring source audio and video data;
recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data;
comparing and displaying the target audio data and/or the target video data with the source audio and video data, specifically: synthesizing source audio data and recorded target video data in the source audio and video data into target audio and video data, or synthesizing source video data and recorded target audio data in the source audio and video data into target audio and video data, or synthesizing recorded target audio data and target video data into target audio and video data; synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data; and displaying the comparative audio and video data.
2. The method of claim 1, wherein recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data comprises:
separating source audio data from the source audio and video data, playing the source audio data, and recording target video data through a camera; and/or the presence of a gas in the gas,
separating source video data from the source audio and video data, playing the source video data in an interface, and recording target audio data through a microphone.
3. The method of claim 1, wherein recording target video data according to source audio data and/or source video data in the source audio and video data comprises:
and playing source audio data and source video data in the source audio and video data in an interface, and recording target video data through a camera.
4. The method of claim 3, wherein playing source audio data and source video data in the source audio and video data in the interface and recording target video data through a camera comprises:
dividing the interface into a first area and a second area;
playing the source video data in the first area;
and recording target video data through a camera, and displaying the recorded target video data in the second area.
5. The method of claim 1, wherein the synthesizing of the source audio-video data and the target audio-video data to obtain the comparison audio-video data comprises:
acquiring source image data of each frame corresponding to at least one segmentation time from the source video data;
acquiring target image data of each frame corresponding to at least one segmentation time from the target video data;
and synthesizing the source image data and the target image data of each frame according to the time sequence to obtain image data of each frame in the compared audio and video data.
6. The method of claim 5, further comprising:
and according to a preset rule, cross setting each section time corresponding to the source video data and each section time corresponding to the target video data.
7. The method of claim 1, wherein the synthesizing of the source audio-video data and the target audio-video data to obtain the comparison audio-video data comprises:
acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data;
acquiring a target audio fragment corresponding to at least one segmentation time from the target audio data;
and synthesizing the source audio clips and the target audio clips according to the time sequence to obtain audio data in the comparison audio and video data.
8. The method of claim 1, wherein said presenting said target audio data and/or target video data in comparison with said source audio-visual data comprises:
setting a third area and a fourth area in the interface;
and displaying the source video data in the third area and displaying the target video data in the fourth area to perform the display of the comparison video.
9. The method of claim 7, further comprising:
and according to a preset rule, cross setting each section time corresponding to the source audio data and each section time corresponding to the target audio data.
10. A video processing device applied to a mobile terminal comprises:
the acquisition module is used for acquiring source audio and video data;
the recording module is used for recording target audio data and/or target video data according to source audio data and/or source video data in the source audio and video data;
the comparison display module is used for comparing and displaying the target audio data and/or the target video data with the source audio and video data;
the comparison display module comprises:
the target synthesis submodule is used for synthesizing source audio data and recorded target video data in the source audio and video data into target audio and video data, or synthesizing source video data and recorded target audio data in the source audio and video data into target audio and video data, or synthesizing recorded target audio data and target video data into target audio and video data;
the comparison and synthesis submodule is used for synthesizing the source audio and video data and the target audio and video data to obtain comparison audio and video data;
and the comparison display submodule is used for displaying the comparison audio and video data.
11. The apparatus of claim 10, wherein the recording module comprises:
the video recording submodule is used for separating source audio data from the source audio and video data, playing the source audio data and recording target video data through a camera;
and the audio recording submodule is used for separating source video data from the source audio and video data, playing the source video data in an interface and recording target audio data through a microphone.
12. The apparatus of claim 10,
the recording module is specifically configured to play source audio data and source video data in the source audio and video data in an interface, and record target video data through a camera.
13. The apparatus of claim 12, wherein the recording module comprises:
the interface dividing submodule is used for dividing the interface into a first area and a second area;
a source video playing sub-module, configured to play the source video data in the first area;
and the video recording submodule is used for recording target video data through a camera and displaying the recorded target video data in the second area.
14. The apparatus of claim 10, wherein the contrast synthesis sub-module comprises:
the source video acquisition unit is used for acquiring source image data of each frame corresponding to at least one segmentation time from the source video data;
the target video acquisition unit is used for acquiring target image data of each frame corresponding to at least one segmentation time from the target video data;
and the video synthesis unit is used for synthesizing the source image data and the target image data of each frame according to the time sequence to obtain the image data of each frame in the compared audio and video data.
15. The apparatus of claim 14, further comprising:
and the setting module is used for alternately setting each section time corresponding to the source video data and each section time corresponding to the target video data according to a preset rule.
16. The apparatus of claim 10, wherein the contrast synthesis sub-module comprises:
the source audio acquisition unit is used for acquiring at least one source audio fragment corresponding to the segmentation time from the source audio data;
the target audio acquiring unit is used for acquiring a target audio fragment corresponding to at least one segmentation time from the target video data;
and the audio synthesis unit synthesizes the audio segments and the target audio segments according to the time sequence to obtain audio data in the comparison audio and video data.
17. The apparatus of claim 10, wherein the comparative display module comprises:
the area setting submodule is used for setting a third area and a fourth area in the interface;
and the comparison display submodule is used for displaying the source image data of each frame in the source video data in the third area and displaying the target video data in the fourth area to display the comparison video.
18. The apparatus of claim 17, further comprising:
and the setting module is used for alternately setting each section time corresponding to the source audio data and each section time corresponding to the target audio data according to a preset rule.
19. A mobile terminal, comprising:
one or more processors; and
one or more machine readable media having instructions stored thereon that, when executed by the one or more processors, cause the mobile terminal to perform a video processing method as recited in one or more of claims 1-9.
20. A machine-readable medium having stored thereon instructions, which when executed by one or more processors, cause a mobile terminal to perform a video processing method according to one or more of claims 1-9.
CN201711009668.2A 2017-10-25 2017-10-25 Video processing method and device, mobile terminal and storage medium Active CN108055490B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711009668.2A CN108055490B (en) 2017-10-25 2017-10-25 Video processing method and device, mobile terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711009668.2A CN108055490B (en) 2017-10-25 2017-10-25 Video processing method and device, mobile terminal and storage medium

Publications (2)

Publication Number Publication Date
CN108055490A CN108055490A (en) 2018-05-18
CN108055490B true CN108055490B (en) 2021-04-13

Family

ID=62119658

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711009668.2A Active CN108055490B (en) 2017-10-25 2017-10-25 Video processing method and device, mobile terminal and storage medium

Country Status (1)

Country Link
CN (1) CN108055490B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108600825B (en) * 2018-07-12 2019-10-25 北京微播视界科技有限公司 Select method, apparatus, terminal device and the medium of background music shooting video
CN108668164A (en) 2018-07-12 2018-10-16 北京微播视界科技有限公司 Select method, apparatus, terminal device and the medium of background music shooting video
CN109274900A (en) * 2018-09-05 2019-01-25 浙江工业大学 A kind of video dubbing method
CN109151356A (en) * 2018-09-05 2019-01-04 传线网络科技(上海)有限公司 video recording method and device
CN109168076B (en) * 2018-11-02 2021-03-19 北京字节跳动网络技术有限公司 Online course recording method, device, server and medium
CN109348155A (en) * 2018-11-08 2019-02-15 北京微播视界科技有限公司 Video recording method, device, computer equipment and storage medium
CN109379633B (en) * 2018-11-08 2020-01-10 北京微播视界科技有限公司 Video editing method and device, computer equipment and readable storage medium
CN109587549B (en) * 2018-12-05 2021-08-13 广州酷狗计算机科技有限公司 Video recording method, device, terminal and storage medium
CN109788308B (en) * 2019-02-01 2022-07-15 腾讯音乐娱乐科技(深圳)有限公司 Audio and video processing method and device, electronic equipment and storage medium
CN111583972B (en) * 2020-05-28 2022-03-25 北京达佳互联信息技术有限公司 Singing work generation method and device and electronic equipment

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101261864B (en) * 2008-04-21 2010-08-18 中兴通讯股份有限公司 A method and system for mixing recording voice at a mobile terminal
CN104967900B (en) * 2015-05-04 2018-08-07 腾讯科技(深圳)有限公司 A kind of method and apparatus generating video
CN105959773B (en) * 2016-04-29 2019-06-18 魔方天空科技(北京)有限公司 The treating method and apparatus of multimedia file
CN106060388B (en) * 2016-06-24 2019-09-27 广东紫旭科技有限公司 A kind of automatically micro- class recording control method and system

Also Published As

Publication number Publication date
CN108055490A (en) 2018-05-18

Similar Documents

Publication Publication Date Title
CN108055490B (en) Video processing method and device, mobile terminal and storage medium
US11355157B2 (en) Special effect synchronization method and apparatus, and mobile terminal
CN108055567B (en) Video processing method and device, terminal equipment and storage medium
CN108022279B (en) Video special effect adding method and device and intelligent mobile terminal
CN107943683B (en) Test script generation method and device, electronic equipment and storage medium
CN107948748B (en) Method, device, mobile terminal and computer storage medium for recommending videos
CN109194973A (en) A kind of more main broadcaster's direct broadcasting rooms give the methods of exhibiting, device and equipment of virtual present
CN105979312B (en) Information sharing method and device
CN106921791B (en) Multimedia file storage and viewing method and device and mobile terminal
CN106303733B (en) Method and device for playing live special effect information
AU2018273505B2 (en) Method for capturing fingerprint and associated products
CN108156508B (en) Barrage information processing method and device, mobile terminal, server and system
CN107943685B (en) Traversal path generation method and device, server and storage medium
CN106231433B (en) A kind of methods, devices and systems playing network video
CN108012090A (en) A kind of method for processing video frequency, device, mobile terminal and storage medium
CN107103074B (en) Processing method of shared information and mobile terminal
CN109660911A (en) Recording sound effect treatment method, device, mobile terminal and storage medium
CN104602100A (en) Method and device for recording video and audio in applications
WO2017215661A1 (en) Scenario-based sound effect control method and electronic device
WO2015172705A1 (en) Method and system for collecting statistics on streaming media data, and related apparatus
CN106791916B (en) Method, device and system for recommending audio data
CN107330867B (en) Image synthesis method, image synthesis device, computer-readable storage medium and computer equipment
CN107396178B (en) Method and device for editing video
CN106777204B (en) Picture data processing method and device and mobile terminal
CN106713636B (en) Loading method, device and the mobile terminal of image data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20180917

Address after: 100015, 15 floor, 3 building, 10 Jiuxianqiao Road, Chaoyang District, Beijing, 17 story 1701-48A

Applicant after: Beijing environment and Wind Technology Co., Ltd.

Address before: 100012 No. 28 building, No. 27 building, Lai Chun Yuan, Chaoyang District, Beijing, No. 28, 2, 201, No. 112, No. 28.

Applicant before: Beijing Chuan Shang Technology Co., Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant