CN109547847B - Method and device for adding video information and computer readable storage medium - Google Patents

Method and device for adding video information and computer readable storage medium Download PDF

Info

Publication number
CN109547847B
CN109547847B CN201811400882.5A CN201811400882A CN109547847B CN 109547847 B CN109547847 B CN 109547847B CN 201811400882 A CN201811400882 A CN 201811400882A CN 109547847 B CN109547847 B CN 109547847B
Authority
CN
China
Prior art keywords
video
song
audio data
page
song name
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811400882.5A
Other languages
Chinese (zh)
Other versions
CN109547847A (en
Inventor
龚巧燕
吴露霜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN201811400882.5A priority Critical patent/CN109547847B/en
Publication of CN109547847A publication Critical patent/CN109547847A/en
Application granted granted Critical
Publication of CN109547847B publication Critical patent/CN109547847B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/437Interfacing the upstream path of the transmission network, e.g. for transmitting client requests to a VOD server
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/475End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • H04N21/8113Monomedia components thereof involving special audio data, e.g. different tracks for different languages comprising music, e.g. song in MP3 format

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The invention discloses a method and a device for adding video information and a computer readable storage medium, and belongs to the technical field of internet. The method comprises the following steps: acquiring audio data in a video; performing song name identification on the audio data, and identifying to obtain a song name corresponding to the song audio data if the audio data comprises song audio data; displaying a confirmation page of the song name, and adding the song name in the video information of the video when receiving an adding instruction triggered by the confirmation page; and sending the video and the video information to a server. By adopting the invention, the adding efficiency of the video information can be improved.

Description

Method and device for adding video information and computer readable storage medium
Technical Field
The present invention relates to the field of internet technologies, and in particular, to a method and an apparatus for adding video information, and a computer-readable storage medium.
Background
With the development of internet technology, users can upload self-made videos to a server through a video application program so as to be watched by other users.
When uploading a video, a user can edit a video picture on a video editing page, and of course, operations such as naming the video, adding a tag and the like can also be performed, and the video name and the tag are added as video information. After the user uploads the video, other users can search for the video name, the label and other search terms of the video, so that the video is exposed.
In the process of implementing the invention, the inventor finds that the prior art has at least the following problems:
if a song is also used in the video as background music, the user may also need to add the song name to the video information. At this time, the user may manually input the song title. However, if the user is unsure of the song title used, it may take time to recall or launch another music application for a query, making it inefficient to add the song title to the video information.
Disclosure of Invention
The embodiment of the invention provides a method and a device for adding video information and a computer readable storage medium, which can improve the adding efficiency of the video information. The technical scheme is as follows:
in a first aspect, a method for adding video information is provided, the method comprising:
acquiring audio data in a video;
performing song name identification on the audio data, and identifying to obtain a song name corresponding to the song audio data if the audio data comprises song audio data;
displaying a confirmation page of the song name, and adding the song name in the video information of the video when receiving an adding instruction triggered by the confirmation page;
and sending the video and the video information to a server.
Optionally, the method further includes:
when a modification instruction triggered by the confirmation page is received, displaying a modification page of the song name, and acquiring the modified song name through the modification page;
and when an adding instruction triggered by the modification page is received, adding the modified song name in the video information of the video.
Optionally, the obtaining the modified song name through the modification page includes:
acquiring a keyword input by a user in an input box of the modification page;
associating the keywords to obtain associated words of the keywords, and displaying the associated words in the modification page, wherein the associated words comprise song names;
modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received;
determining the words in the input box as the modified song names.
Optionally, the acquiring the audio data in the video includes:
when a video editing page of the video is entered, acquiring audio data in the video; or
And when a video publishing and editing page of the video is entered, acquiring audio data in the video.
Optionally, the acquiring the audio data in the video includes:
when the video is a locally imported video, acquiring audio data in the video; or
Before recording the video, when the song is not selected, acquiring audio data in the video; or
And before the video is recorded, when a locally imported song is selected, acquiring audio data in the video.
In a second aspect, there is provided an apparatus for adding video information, the apparatus comprising:
the acquisition module is used for acquiring audio data in the video;
the identification module is used for identifying song names of the audio data, and if the audio data comprises song audio data, identifying to obtain song names corresponding to the song audio data;
the adding module is used for displaying a confirmation page of the song name, and adding the song name in the video information of the video when an adding instruction triggered by the confirmation page is received;
and the sending module is used for sending the video and the video information to a server.
Optionally, the apparatus further includes a modification module, where the modification module is configured to:
when a modification instruction triggered by the confirmation page is received, displaying a modification page of the song name, and acquiring the modified song name through the modification page;
and when an adding instruction triggered by the modification page is received, adding the modified song name in the video information of the video.
Optionally, the modification module is configured to:
acquiring a keyword input by a user in an input box of the modification page;
associating the keywords to obtain associated words of the keywords, and displaying the associated words in the modification page, wherein the associated words comprise song names;
modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received;
determining the words in the input box as the modified song names.
Optionally, the obtaining module is configured to:
when a video editing page of the video is entered, acquiring audio data in the video; or
And when a video publishing and editing page of the video is entered, acquiring audio data in the video.
Optionally, the obtaining module is configured to:
when the video is a locally imported video, acquiring audio data in the video; or
Before recording the video, when the song is not selected, acquiring audio data in the video; or
And before the video is recorded, when a locally imported song is selected, acquiring audio data in the video.
In a third aspect, a terminal is provided, which includes a processor and a memory, where the memory stores at least one instruction, and the at least one instruction is loaded and executed by the processor to implement the method for adding video information according to the first aspect.
In a fourth aspect, a computer-readable storage medium is provided, in which at least one instruction is stored, and the at least one instruction is loaded and executed by a processor of a terminal to implement the method for adding video information according to the first aspect.
The technical scheme provided by the embodiment of the invention has the following beneficial effects:
in this embodiment, before the user uploads the video, the terminal may identify audio data of the video. If the video contains songs, the terminal can identify the songs from the audio data and obtain corresponding song names. Furthermore, the terminal can add the song name to the video information, and the adding efficiency of the video information is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a diagram of an implementation environment provided by an embodiment of the invention;
fig. 2 is a flowchart of a method for adding video information according to an embodiment of the present invention;
fig. 3 is a flowchart of a method for adding video information according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a confirmation page provided by an embodiment of the invention;
FIG. 5 is a schematic diagram of a modified page according to an embodiment of the present invention;
FIG. 6 is a diagram illustrating a display of associative words according to an embodiment of the present invention;
fig. 7 is a schematic diagram of an apparatus for adding video information according to an embodiment of the present invention;
fig. 8 is a block diagram of a terminal according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in detail with reference to the accompanying drawings.
Fig. 1 is a diagram of an implementation environment provided by an embodiment of the invention. The implementation environment may include a plurality of terminals 101, a server 102 for providing services to the plurality of terminals 101. A plurality of terminals 101 are connected to the server 102 through a wireless or wired network, and the plurality of terminals 101 may be computer devices or intelligent terminals or the like capable of accessing the server 102. The terminal 101 may have a video application installed therein, and the server 102 may provide a background service for the video application. The server 102 may also have at least one database for storing videos uploaded by users, and the like.
The embodiment of the invention provides a method for adding video information, which can be realized by a terminal. As shown in fig. 2, the processing flow of the method for adding video information may include the following steps:
in step 201, the terminal acquires audio data in a video.
The user may upload the video to the server through the video application. Before uploading, the terminal may extract audio data from the video to be uploaded.
In step 202, the terminal performs song name identification on the audio data, and if the audio data includes song audio data, a song name corresponding to the song audio data is identified.
The video application may be provided with audio recognition functions, such as listening to songs, humming recognition. The terminal may also have a song library stored thereon for the video application including song audio data for each song.
After the terminal acquires the audio data, the audio data can be compared with the audio data in the song library one by one, and the matching rate of the audio data and each song is determined. When the audio data of the songs are included in the audio data, the terminal may determine one or more songs having a matching rate greater than a threshold and acquire corresponding song names.
Of course, if the terminal does not store the song library of the video application program, or the data volume of the song library stored by the terminal is small, the terminal may also upload the acquired audio data to the server, and the server executes the processing for identifying the audio data. And when the server sends the identification result to the terminal, the terminal can obtain the corresponding song name.
In step 203, the terminal displays a confirmation page of the song title, and when an adding instruction triggered by the confirmation page is received, the song title is added in the video information of the video.
The video information may include a video name, a tag, an audio name, and the like, and the specific video information is not limited in this embodiment.
After the terminal acquires the song name, a confirmation page of the song name can be displayed, wherein the song name and the confirmation option can be included. The user can see the identified song title and confirm that it is correct. If the song title is correct, the user may click on the confirm option, triggering an add instruction to the song title. Further, the terminal may add the confirmed song title as an audio title to the video information.
In step 204, the terminal sends the video and the video information to the server.
The user can click an upload option in the video application program to trigger an upload instruction of the video. Furthermore, the terminal can upload the video and the video information to the server accordingly. The server may store the video and the video information correspondingly.
When a user searches for a video, a search word can be input in a search box provided by a video application program, and a search instruction for the search word is triggered after a search option is clicked. Further, the terminal may send a search request carrying the search term to the server. The server may search for the matched video information from the stored video information according to the search term, and transmit a search result to the terminal of the user, where the search result may include at least one video and corresponding video information.
When the terminal plays the video, the video information of the video can be displayed at the set position. On this basis, the user can see the video information. If the user is interested in the songs in the video, the song names of the songs can be obtained from the video information, additional searching for the songs is not needed, and the efficiency of obtaining the songs in the video by the user is improved.
In this embodiment, before the user uploads the video, the terminal may identify audio data of the video. If the video contains songs, the terminal can identify the songs from the audio data and obtain corresponding song names. Furthermore, the terminal can add the song name to the video information, and the adding efficiency of the video information is improved.
Based on the same technical concept, the embodiment also provides a method for adding video information, which can be implemented by a terminal. As shown in fig. 3, the processing flow of the method for adding video information may include the following steps:
in step 301, the terminal acquires a video.
In implementation, a user may record a video through a video application, or may import a local video into a video application. When the video recording is finished or the local video importing is finished, the terminal may acquire data of the video, where the data may include video image data and audio data.
In step 302, the terminal enters a video editing page of the video.
When the video recording is finished or the local video importing is finished, the video application program of the terminal may enter a video editing page so that the user can edit the video, such as video clip, filter addition, and the like.
In step 303, when a video editing page of a video is entered, the terminal acquires audio data in the video.
When entering a video editing page of the video, the terminal can trigger an instruction for acquiring the audio data and extract the audio data of the video.
Of course, the terminal may also trigger an instruction to acquire audio data after the video editing is completed, that is, the processing in step 303 may also be as follows: when a video publishing and editing page of a video is entered, the terminal acquires audio data in the video. After the video editing is completed, the video application program of the terminal can enter a video publishing editing page so that the user can confirm the uploaded video. At this time, the terminal may trigger an instruction to acquire audio data, and extract audio data of the video.
For the above two ways of triggering the instruction to acquire audio data, a technician may select one way to trigger the instruction to acquire audio data when designing the process flow. Besides, in addition to the above two modes, a technician may set an audio identification option in a video editing page or a video publishing editing page of the video application program, and on this basis, the terminal may also trigger an instruction to acquire audio data when the user clicks the audio identification option. The present embodiment does not limit the manner of triggering the instruction to acquire audio data.
In one possible implementation, a user may select a song provided by a video application as background music when recording a video. In this case, the song name of the song is known to the video application, the audio data may not be identified, and the audio data need not be acquired. Therefore, the present embodiment provides three application scenarios for acquiring audio data in video, which are described below separately.
First, when a video is a locally imported video, a terminal acquires audio data in the video.
When a user imports a local video into a video application, if the video includes a song, the song is unknown to the video application. Therefore, for the video, the terminal may trigger an instruction to acquire audio data, and perform the processing of step 303 described above.
Secondly, before recording the video, when the song is not selected, the terminal acquires the audio data in the video.
Before recording a video, if the user does not select any song as background music, the user may hum a song during the recording of the video, or take a song played by another device. Likewise, the songs described above are also unknown to the video application. Therefore, for the video, the terminal may also trigger an instruction to acquire audio data, and perform the processing of step 303.
Thirdly, before recording the video, when selecting the locally imported song, the terminal acquires the audio data in the video.
The user may import a local song into the video application before recording the video and select the song as background music for the video. Likewise, the songs described above are also unknown to the video application. Therefore, for the video, the terminal may also trigger an instruction to acquire audio data, and perform the processing of step 303.
Optionally, when the user selects a song provided by the video application before recording the video, the terminal may not trigger the instruction for acquiring the audio data, may add the song name of the song to the video information, and end the method flow for adding the video information.
In step 304, the terminal performs song name identification on the audio data, and if the audio data includes song audio data, a song name corresponding to the song audio data is identified.
The processing of step 304 is the same as that of step 202, and is not described herein again. As shown in the schematic diagram of the confirmation page shown in fig. 4, when the terminal recognizes the song name, the confirmation page for the song name may be displayed in the video application, wherein a confirmation option and a denial option may be provided.
When the terminal does not recognize the song title, the audio title in the video information may be set to null or to an acoustic flag, which may be used to indicate that the song is not included in the video.
In step 305, whether the terminal has received an instruction to add a song title.
When the user judges that the song name is correct, the user can click a confirmation option to trigger an adding instruction of the song name. When the user judges that the song name is wrong, the user can click the denial option to trigger a modification instruction of the song name.
In step 306, the terminal displays a confirmation page of the song title, and when an adding instruction triggered by the confirmation page is received, the song title is added to the video information of the video.
When receiving an addition instruction for a song title, the terminal may add the confirmed song title as an audio title to the video information.
In step 307, when receiving a modification instruction triggered by the confirmation page, the terminal displays a modification page of the song title, and acquires the modified song title through the modification page.
As shown in the schematic diagram of the modification page of fig. 5, when a modification instruction for a song title is received, the video application of the terminal may enter the modification page in which an input box for the song title and a confirmation option for the input box may be displayed. The user may enter the song title he or she believes to be correct in the input box and may click on the confirmation option, triggering an add instruction to the song title in the input box. At this time, the terminal may acquire the song name in the input box and determine the song name as the modified song name.
In one possible implementation, when the user inputs a keyword in the input box, the page may display an associated word of the keyword. At this time, the process of acquiring the modified song title of the above step 307 may be as follows: acquiring a keyword input by a user in an input box of a modified page; performing association processing on the keywords to obtain associated words of the keywords, and displaying the associated words in a modification page; modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received; the words in the input box are determined as the modified song title.
The keyword input by the user may be a song title, a singer title or other words. The associative word may be a song title.
When a user inputs a keyword in the input box, the terminal can acquire the keyword, search the keyword in each song name in the song library, determine the matching rate of the keyword and each song name, and perform association processing. The present embodiment does not limit the specific search method, and may be, for example, fuzzy search.
As shown in fig. 6, the terminal may display the preset number of song titles with the highest matching rate as the association words in the modification page. In this embodiment, a specific display mode of the associated word is not limited, and for example, the associated word may be displayed in a manner of being pulled down below the input box.
The user may or may not select any song title among the displayed associated words. When the user clicks and selects any associative word, a selection instruction of the associative word can be triggered. At this time, the terminal may fill the associated word in the input box to replace the original keyword in the input box.
When the user clicks the confirmation option, an adding instruction of a word in the input box can be triggered, wherein the word can be an association word selected by the user or a keyword input by the user. At this time, the terminal may acquire a word in the input box and determine the word as the modified song title.
In step 308, when receiving an adding instruction triggered by modifying the page, the terminal adds the modified song name in the video information of the video.
When receiving an addition instruction for the modified song title, the terminal may add the modified song title to the video information as an audio title.
In step 309, the terminal transmits the video and the video information to the server.
The processing of step 309 is the same as that of step 204, and is not described herein again.
In this embodiment, before the user uploads the video, the terminal may identify audio data of the video. If the video contains songs, the terminal can identify the songs from the audio data and obtain corresponding song names. Furthermore, the terminal can add the song name to the video information, and the adding efficiency of the video information is improved. After the terminal uploads the video and the video information to the server correspondingly, other users can search the video through keywords such as video names and labels and can also search through song names, the convenience of searching the video is improved, and the exposure of the video can also be improved.
And, generally speaking, the user is clear about the song he or she selected, so that the user can participate in the confirmation of the song name and can improve the accuracy of the song name when actively modifying.
Some video songs are time-sensitive and if they are not identified by the server or manually after uploading, the correct song name cannot be added at the first time. In this embodiment, the user can determine the song name before uploading the video by controlling the uploading time, so that the timeliness of the song name can be ensured.
Based on the same technical concept, an embodiment of the present invention further provides an apparatus for adding video information, which is applied to a terminal in the foregoing embodiment, and as shown in fig. 7, the apparatus for adding video information includes:
an obtaining module 710, configured to obtain audio data in a video;
the identification module 720 is configured to perform song name identification on the audio data, and if the audio data includes song audio data, identify to obtain a song name corresponding to the song audio data;
an adding module 730, configured to display a confirmation page of the song name, and add the song name to the video information of the video when an adding instruction triggered by the confirmation page is received;
a sending module 740, configured to send the video and the video information to a server.
Optionally, the apparatus further includes a modification module, where the modification module is configured to:
when a modification instruction triggered by the confirmation page is received, displaying a modification page of the song name, and acquiring the modified song name through the modification page;
and when an adding instruction triggered by the modification page is received, adding the modified song name in the video information of the video.
Optionally, the modification module is configured to:
acquiring a keyword input by a user in an input box of the modification page;
associating the keywords to obtain associated words of the keywords, and displaying the associated words in the modification page, wherein the associated words comprise song names;
modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received;
determining the words in the input box as the modified song names.
Optionally, the obtaining module 710 is configured to:
when a video editing page of the video is entered, acquiring audio data in the video; or
And when a video publishing and editing page of the video is entered, acquiring audio data in the video.
Optionally, the obtaining module 710 is configured to:
when the video is a locally imported video, acquiring audio data in the video; or
Before recording the video, when the song is not selected, acquiring audio data in the video; or
And before the video is recorded, when a locally imported song is selected, acquiring audio data in the video.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
In this embodiment, before the user uploads the video, the terminal may identify audio data of the video. If the video contains songs, the terminal can identify the songs from the audio data and obtain corresponding song names. Furthermore, the terminal can add the song name to the video information, and the adding efficiency of the video information is improved.
It should be noted that: in the above-mentioned embodiment, when adding video information, the apparatus for adding video information is illustrated by only dividing the functional modules, and in practical applications, the above-mentioned function distribution may be completed by different functional modules according to needs, that is, the internal structure of the terminal is divided into different functional modules to complete all or part of the above-mentioned functions. In addition, the apparatus for adding video information and the method for adding video information provided by the above embodiments belong to the same concept, and specific implementation processes thereof are detailed in the method embodiments and are not described herein again.
Fig. 8 is a block diagram illustrating a terminal 800 according to an exemplary embodiment of the present invention. The terminal 800 may be: a smart phone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, motion video Experts compression standard Audio Layer 3), an MP4 player (Moving Picture Experts Group Audio Layer IV, motion video Experts compression standard Audio Layer 4), a notebook computer, or a desktop computer. The terminal 800 may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, etc.
In general, the terminal 800 includes: a processor 801 and a memory 802.
The processor 801 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so forth. The processor 801 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 801 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 801 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed on the display screen. In some embodiments, the processor 801 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 802 may include one or more computer-readable storage media, which may be non-transitory. Memory 802 may also include high speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 802 is used to store at least one instruction for execution by processor 801 to implement the method of adding video information provided by the method embodiments herein.
In some embodiments, the terminal 800 may further include: a peripheral interface 803 and at least one peripheral. The processor 801, memory 802 and peripheral interface 803 may be connected by bus or signal lines. Various peripheral devices may be connected to peripheral interface 803 by a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 804, a touch screen display 805, a camera 806, an audio circuit 807, a positioning component 808, and a power supply 809.
The peripheral interface 803 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 801 and the memory 802. In some embodiments, the processor 801, memory 802, and peripheral interface 803 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 801, the memory 802, and the peripheral interface 803 may be implemented on separate chips or circuit boards, which are not limited by this embodiment.
The Radio Frequency circuit 804 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 804 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 804 converts an electrical signal into an electromagnetic signal to be transmitted, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 804 includes: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuit 804 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 804 may further include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 805 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display 805 is a touch display, the display 805 also has the ability to capture touch signals on or above the surface of the display 805. The touch signal may be input to the processor 801 as a control signal for processing. At this point, the display 805 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 805 may be one, providing the front panel of the terminal 800; in other embodiments, the display 805 may be at least two, respectively disposed on different surfaces of the terminal 800 or in a folded design; in still other embodiments, the display 805 may be a flexible display disposed on a curved surface or a folded surface of the terminal 800. Even further, the display 805 may be arranged in a non-rectangular irregular pattern, i.e., a shaped screen. The Display 805 can be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), and other materials.
The camera assembly 806 is used to capture images or video. Optionally, camera assembly 806 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 806 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
The audio circuit 807 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 801 for processing or inputting the electric signals to the radio frequency circuit 804 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 800. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 801 or the radio frequency circuit 804 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, the audio circuitry 807 may also include a headphone jack.
The positioning component 808 is used to locate the current geographic position of the terminal 800 for navigation or LBS (Location Based Service). The Positioning component 808 may be a Positioning component based on the GPS (Global Positioning System) in the united states, the beidou System in china, the graves System in russia, or the galileo System in the european union.
Power supply 809 is used to provide power to various components in terminal 800. The power supply 809 can be ac, dc, disposable or rechargeable. When the power source 809 comprises a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 800 also includes one or more sensors 810. The one or more sensors 810 include, but are not limited to: acceleration sensor 811, gyro sensor 812, pressure sensor 813, fingerprint sensor 814, optical sensor 815 and proximity sensor 816.
The acceleration sensor 811 may detect the magnitude of acceleration in three coordinate axes of the coordinate system established with the terminal 800. For example, the acceleration sensor 811 may be used to detect the components of the gravitational acceleration in three coordinate axes. The processor 801 may control the touch screen 805 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 811. The acceleration sensor 811 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 812 may detect a body direction and a rotation angle of the terminal 800, and the gyro sensor 812 may cooperate with the acceleration sensor 811 to acquire a 3D motion of the user with respect to the terminal 800. From the data collected by the gyro sensor 812, the processor 801 may implement the following functions: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
Pressure sensors 813 may be disposed on the side bezel of terminal 800 and/or underneath touch display 805. When the pressure sensor 813 is disposed on the side frame of the terminal 800, the holding signal of the user to the terminal 800 can be detected, and the processor 801 performs left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 813. When the pressure sensor 813 is disposed at a lower layer of the touch display screen 805, the processor 801 controls the operability control on the UI interface according to the pressure operation of the user on the touch display screen 805. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 814 is used for collecting a fingerprint of the user, and the processor 801 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 814, or the fingerprint sensor 814 identifies the identity of the user according to the collected fingerprint. Upon identifying that the user's identity is a trusted identity, the processor 801 authorizes the user to perform relevant sensitive operations including unlocking a screen, viewing encrypted information, downloading software, paying for and changing settings, etc. Fingerprint sensor 814 may be disposed on the front, back, or side of terminal 800. When a physical button or a vendor Logo is provided on the terminal 800, the fingerprint sensor 814 may be integrated with the physical button or the vendor Logo.
The optical sensor 815 is used to collect the ambient light intensity. In one embodiment, the processor 801 may control the display brightness of the touch screen 805 based on the ambient light intensity collected by the optical sensor 815. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 805 is increased; when the ambient light intensity is low, the display brightness of the touch display 805 is turned down. In another embodiment, the processor 801 may also dynamically adjust the shooting parameters of the camera assembly 806 based on the ambient light intensity collected by the optical sensor 815.
A proximity sensor 816, also known as a distance sensor, is typically provided on the front panel of the terminal 800. The proximity sensor 816 is used to collect the distance between the user and the front surface of the terminal 800. In one embodiment, when the proximity sensor 816 detects that the distance between the user and the front surface of the terminal 800 gradually decreases, the processor 801 controls the touch display 805 to switch from the bright screen state to the dark screen state; when the proximity sensor 816 detects that the distance between the user and the front surface of the terminal 800 becomes gradually larger, the processor 801 controls the touch display 805 to switch from the screen-on state to the screen-on state.
Those skilled in the art will appreciate that the configuration shown in fig. 8 is not intended to be limiting of terminal 800 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
In an exemplary embodiment, a computer-readable storage medium, such as a memory, including instructions executable by a processor in a terminal to perform the above-described method of adding video information is also provided. For example, the computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (10)

1. A method of adding video information, the method comprising:
when a video is a locally imported video, or before the video is recorded, a song is not selected, or before the video is recorded, a locally imported song is selected, and audio data in the video is acquired;
performing song name identification on the audio data, and identifying to obtain a song name corresponding to the song audio data if the audio data comprises song audio data;
displaying a confirmation page of the song name, and adding the song name in the video information of the video when receiving an adding instruction triggered by the confirmation page; if before the video recording, when a song provided by a video application program is selected, instructions for acquiring audio data, identifying the song name of the audio data and displaying a confirmation page of the song name are not triggered, and the song name of the song is directly added to the video information;
and when an uploading instruction of the video is triggered, sending the video and the video information added with the song name to a server, so that other users can search the video through the song name.
2. The method of claim 1, further comprising:
when a modification instruction triggered by the confirmation page is received, displaying a modification page of the song name, and acquiring the modified song name through the modification page;
and when an adding instruction triggered by the modification page is received, adding the modified song name in the video information of the video.
3. The method of claim 2, wherein the obtaining of the modified song title through the modification page comprises:
acquiring a keyword input by a user in an input box of the modification page;
associating the keywords to obtain associated words of the keywords, and displaying the associated words in the modification page, wherein the associated words comprise song names;
modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received;
determining the words in the input box as the modified song names.
4. The method of claim 1, wherein the obtaining audio data in the video comprises:
when a video editing page of the video is entered, acquiring audio data in the video; or
And when a video publishing and editing page of the video is entered, acquiring audio data in the video.
5. An apparatus for adding video information, the apparatus comprising:
the acquisition module is used for acquiring audio data in a video when the video is a locally imported video, or before the video is recorded, a song is not selected, or before the video is recorded, a locally imported song is selected;
the identification module is used for identifying song names of the audio data, and if the audio data comprises song audio data, identifying to obtain song names corresponding to the song audio data;
the adding module is used for displaying a confirmation page of the song name, and adding the song name in the video information of the video when an adding instruction triggered by the confirmation page is received; if before the video recording, when a song provided by a video application program is selected, instructions for acquiring audio data, identifying the song name of the audio data and displaying a confirmation page of the song name are not triggered, and the song name of the song is directly added to the video information;
and the sending module is used for sending the video and the video information added with the song name to a server when an uploading instruction of the video is triggered, so that other users can search the video through the song name.
6. The apparatus of claim 5, further comprising a modification module to:
when a modification instruction triggered by the confirmation page is received, displaying a modification page of the song name, and acquiring the modified song name through the modification page;
and when an adding instruction triggered by the modification page is received, adding the modified song name in the video information of the video.
7. The apparatus of claim 6, wherein the modification module is configured to:
acquiring a keyword input by a user in an input box of the modification page;
associating the keywords to obtain associated words of the keywords, and displaying the associated words in the modification page, wherein the associated words comprise song names;
modifying the keywords in the input box into the associated words when a selection instruction of the associated words is received;
determining the words in the input box as the modified song names.
8. The apparatus of claim 5, wherein the obtaining module is configured to:
when a video editing page of the video is entered, acquiring audio data in the video; or
And when a video publishing and editing page of the video is entered, acquiring audio data in the video.
9. A terminal, characterized in that the terminal comprises a processor and a memory, wherein the memory stores at least one instruction, and the at least one instruction is loaded and executed by the processor to realize the method for adding video information according to any one of claims 1 to 4.
10. A computer-readable storage medium, wherein at least one instruction is stored in the storage medium, and the at least one instruction is loaded and executed by a processor of a terminal to implement the method for adding video information according to any one of claims 1 to 4.
CN201811400882.5A 2018-11-22 2018-11-22 Method and device for adding video information and computer readable storage medium Active CN109547847B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811400882.5A CN109547847B (en) 2018-11-22 2018-11-22 Method and device for adding video information and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811400882.5A CN109547847B (en) 2018-11-22 2018-11-22 Method and device for adding video information and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN109547847A CN109547847A (en) 2019-03-29
CN109547847B true CN109547847B (en) 2021-10-22

Family

ID=65849986

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811400882.5A Active CN109547847B (en) 2018-11-22 2018-11-22 Method and device for adding video information and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN109547847B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111491211B (en) * 2020-04-17 2022-01-28 维沃移动通信有限公司 Video processing method, video processing device and electronic equipment
CN112135182B (en) * 2020-09-22 2023-02-03 杭州网易云音乐科技有限公司 List processing method, list processing apparatus, storage medium, and electronic device
CN114390341B (en) * 2020-10-22 2023-06-06 华为技术有限公司 Video recording method, electronic equipment, storage medium and chip

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102074235A (en) * 2010-12-20 2011-05-25 上海华勤通讯技术有限公司 Method of video speech recognition and search
KR101770551B1 (en) * 2016-06-03 2017-09-05 티제이미디어 주식회사 Appratus and method for matching video

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009100093A1 (en) * 2008-02-05 2009-08-13 Dolby Laboratories Licensing Corporation Associating information with media content
CN103377203A (en) * 2012-04-18 2013-10-30 宇龙计算机通信科技(深圳)有限公司 Terminal and sound record management method
CN105913838B (en) * 2016-05-19 2019-11-05 努比亚技术有限公司 Audio frequency controller device and method
CN105872855A (en) * 2016-05-26 2016-08-17 广州酷狗计算机科技有限公司 Labeling method and device for video files
CN106294618A (en) * 2016-08-01 2017-01-04 北京百度网讯科技有限公司 Searching method and device
CN106375782B (en) * 2016-08-31 2020-12-18 北京小米移动软件有限公司 Video playing method and device
US10075451B1 (en) * 2017-03-08 2018-09-11 Venpath, Inc. Methods and systems for user opt-in to data privacy agreements
CN106804005B (en) * 2017-03-27 2019-05-17 维沃移动通信有限公司 A kind of production method and mobile terminal of video
CN106940996A (en) * 2017-04-24 2017-07-11 维沃移动通信有限公司 The recognition methods of background music and mobile terminal in a kind of video
CN107135419A (en) * 2017-06-14 2017-09-05 北京奇虎科技有限公司 A kind of method and apparatus for editing video
CN108039184A (en) * 2017-12-28 2018-05-15 腾讯音乐娱乐科技(深圳)有限公司 Lyrics adding method and device
CN108282678B (en) * 2018-02-11 2021-01-05 孙新峰 Multimedia data playing method, device and system
CN108600825B (en) * 2018-07-12 2019-10-25 北京微播视界科技有限公司 Select method, apparatus, terminal device and the medium of background music shooting video

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102074235A (en) * 2010-12-20 2011-05-25 上海华勤通讯技术有限公司 Method of video speech recognition and search
KR101770551B1 (en) * 2016-06-03 2017-09-05 티제이미디어 주식회사 Appratus and method for matching video

Also Published As

Publication number Publication date
CN109547847A (en) 2019-03-29

Similar Documents

Publication Publication Date Title
CN107885533B (en) Method and device for managing component codes
CN107908929B (en) Method and device for playing audio data
CN109168073B (en) Method and device for displaying cover of live broadcast room
CN108737897B (en) Video playing method, device, equipment and storage medium
CN109068160B (en) Method, device and system for linking videos
CN108717432B (en) Resource query method and device
CN109327608B (en) Song sharing method, terminal, server and system
CN110248236B (en) Video playing method, device, terminal and storage medium
CN109922356B (en) Video recommendation method and device and computer-readable storage medium
CN109144346B (en) Song sharing method and device and storage medium
CN113411680B (en) Multimedia resource playing method, device, terminal and storage medium
CN110266982B (en) Method and system for providing songs while recording video
CN110933468A (en) Playing method, playing device, electronic equipment and medium
CN111711838B (en) Video switching method, device, terminal, server and storage medium
CN112261491B (en) Video time sequence marking method and device, electronic equipment and storage medium
CN109618192B (en) Method, device, system and storage medium for playing video
CN109547847B (en) Method and device for adding video information and computer readable storage medium
CN112052354A (en) Video recommendation method, video display method and device and computer equipment
CN111880888A (en) Preview cover generation method and device, electronic equipment and storage medium
CN111818367A (en) Audio file playing method, device, terminal, server and storage medium
CN111402844A (en) Song chorusing method, device and system
CN109189978B (en) Method, device and storage medium for audio search based on voice message
CN108495183B (en) Method and device for displaying album information
CN111399796B (en) Voice message aggregation method and device, electronic equipment and storage medium
CN112069350A (en) Song recommendation method, device, equipment and computer storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant