WO2018095219A1 - 媒体信息处理方法和装置 - Google Patents

媒体信息处理方法和装置 Download PDF

Info

Publication number
WO2018095219A1
WO2018095219A1 PCT/CN2017/109927 CN2017109927W WO2018095219A1 WO 2018095219 A1 WO2018095219 A1 WO 2018095219A1 CN 2017109927 W CN2017109927 W CN 2017109927W WO 2018095219 A1 WO2018095219 A1 WO 2018095219A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
client
information
dubbing
request
Prior art date
Application number
PCT/CN2017/109927
Other languages
English (en)
French (fr)
Inventor
阳丹
谢奕
王贤骏
周莜
瞿佳
李锐
林�建
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Publication of WO2018095219A1 publication Critical patent/WO2018095219A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4113PC
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content

Definitions

  • the present application relates to the field of multimedia technologies, and in particular, to a media information processing method and apparatus.
  • the application example provides a media information processing method, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing method, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing method, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing server, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing client, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing client, which can improve dubbing efficiency and save system resources.
  • the application example provides a media information processing method, which is applied to a voiceover server, and includes:
  • the application example provides a media information processing method, including:
  • the first client triggers the dubbing trigger information displayed on the interface of the second client to play the first video to generate a dubbing request carrying the identifier ID of the first video;
  • the first client sends the voice-over request to the voice-over server, so that the voice-over server obtains the second video according to the ID of the first video carried in the voice-over request, where the second video is Obtaining the first audio information of the character in the first video;
  • the first client plays the second video, and collects second audio information for the character according to the played second video;
  • the first client sends the second audio information to the voiceover server.
  • the application example provides a media information processing method, which is applied to a media information processing client, and includes:
  • the dubbing server Transmitting the dubbing request to the dubbing server, so that the dubbing server acquires a second video according to the ID of the first video carried in the dubbing request, wherein the second video is removed Obtained from the first audio information of the character in the first video;
  • the application example provides a media information processing server, including: a processor and a memory, where the memory stores:
  • a first video sending module configured to send, to the first client, a first video and information indicating that the voiceover is allowed, so that the first client plays the first video, and the information that allows the voiceover is played according to the indication Dubbing trigger information is displayed on the interface of the first video;
  • a first dubbing request receiving module configured to receive a first dubbing request generated by the second client according to the triggering operation of the dubbing trigger information
  • a first video acquiring module configured to acquire a second video according to the identifier ID of the first video carried in the first voice-over request, where the second video is removed from the first video Obtained from the first audio information of the character;
  • a second video sending module configured to send the second video to the second client
  • a second video acquiring module configured to receive, from the second client, second audio information of the character collected according to the played second video, for the second video and the role
  • the second audio information performs a synthesis process to obtain a third video; or receives, from the second client, the third video obtained by the second client using the second video and the second audio information to perform a synthesis process.
  • the application example provides a media information processing client, including: a processor and a memory, where the memory stores:
  • a dubbing request generating module configured to perform a triggering operation on the dubbing trigger information displayed by another client on the interface for playing the first video, to generate a dubbing request carrying the identifier ID of the first video;
  • a dubbing request sending module configured to send the dubbing request to the dubbing server, so that the dubbing server acquires the second video according to the ID of the first video carried in the dubbing request, where the second video Is obtained by removing the first audio information of the character in the first video;
  • a video receiving module configured to receive a second video from the voice-over server
  • a video playing module configured to play the second video
  • An audio collection module configured to collect second audio information for the character according to the played second video
  • an audio sending module configured to send the second audio information to the voiceover server.
  • the application example provides a media information processing client, including: a processor and a memory, where the memory stores:
  • a video receiving module configured to receive a first video from the dubbing server and information indicating that the dubbing is allowed
  • a video playing module configured to play the first video, and display the dubbing trigger information on the interface that plays the first video according to the information that allows the dubbing according to the indication;
  • a dubbing request generating module configured to generate a dubbing request carrying an identifier ID of the first video in response to a triggering operation on the dubbing trigger information
  • a dubbing request sending module configured to send the dubbing request to the dubbing server, so that the dubbing server acquires the second video according to the ID of the first video carried in the dubbing request, where the second video Is obtained by removing the first audio information of the character in the first video;
  • the video receiving module is further configured to receive a second video from the voice-over server
  • the video playing module is further configured to play the second video
  • An audio collection module configured to collect second audio information for the character according to the played second video
  • an audio information sending module configured to send the second audio information to the voiceover server.
  • the first video and the information indicating the allowed dubbing are sent to the first client.
  • the first client plays the first video, and according to the indication, the information of the dubbing is displayed to display the dubbing trigger information during the playing of the first video.
  • the second video is obtained by removing first audio information of a character in the first video.
  • the dubbing trigger information may be triggered on the interface of playing the video during the process of playing the video, and the dubbing operation may be performed.
  • FIG. 1 is a schematic structural diagram of an implementation environment involved in each example of the present application.
  • FIG. 2 is a schematic structural diagram of an implementation environment involved in each example of the present application.
  • FIG. 3 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application
  • FIG. 4 is a schematic flowchart of a method for processing media information provided by an example of the present application.
  • FIG. 5 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • FIG. 6 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • 6A is a schematic diagram of a dubbing request triggering interface in media information processing provided by an example of the present application
  • 6B is a schematic diagram of a dubbing request triggering interface in media information processing provided by an example of the present application.
  • 6C is a schematic diagram of a role selection interface in media information processing according to an example of the present application.
  • 6D is an interface for collecting second audio information on a mobile phone in media information processing provided by an example of the present application.
  • 6E is a schematic diagram of an end collection interface in a media information processing process provided by an example of the present application.
  • 6F is a schematic diagram of an interface for playing audio information of a character in a media information processing process according to an example of the present application
  • FIG. 7 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • FIG. 8 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • FIG. 8A is a schematic diagram of a video playing interface provided by an implementation of the present application.
  • FIG. 8B is a schematic diagram of a video wall provided by an example of the present application.
  • 8C is a schematic diagram of a video switching interface in media information processing provided by an example of the present application.
  • FIG. 9 is a schematic structural diagram of a media information processing server according to an example of the present application.
  • FIG. 10 is a schematic structural diagram of a media information processing client according to an example of the present application.
  • FIG. 11 is a schematic structural diagram of a media information processing client according to an example of the present application.
  • FIG. 12 is a schematic structural diagram of a media information processing server in an example of the present application.
  • FIG. 13 is a schematic structural diagram of a media information processing client in an example of the present application.
  • FIG. 14 is a schematic structural diagram of a media information processing client in an example of the present application.
  • FIG. 1 is a schematic structural diagram of an implementation environment involved in each example of the present application. As shown in FIG. 1 , the implementation environment includes: a playback device 110, a smart terminal 120, and a dubbing server 130.
  • the playback device 110 can be a television or the like.
  • the smart terminal 120 can be a mobile phone, a tablet, or the like.
  • the playing device 110 can communicate with the dubbing server 130 through a wired or wireless network, acquire a video file from the dubbing server 130, and play, and display dubbing trigger information, such as a two-dimensional code, on the playing interface during the playing of the video. .
  • the smart terminal 120 can trigger the dubbing trigger information, for example, scan the two-dimensional code, and communicate with the dubbing server 130 to complete a subsequent dubbing operation.
  • the dubbing server 130 can be a server, or a server cluster consisting of several servers, or a cloud computing service center.
  • the dubbing server 130 can be used to store video files or to communicate with a database to retrieve video files from a database.
  • FIG. 2 is a schematic structural diagram of an implementation environment involved in each example of the present application. As shown in FIG. 2, the implementation environment includes: a terminal device 210 and a dubbing server 220.
  • the terminal device 210 can be a PC, a laptop, a mobile phone, or a tablet.
  • the dubbing server 220 can be a server, or a server cluster consisting of several servers, or a cloud computing service center.
  • the dubbing server 220 can be used to store video files or to communicate with a database to retrieve video files from a database.
  • the terminal device 210 can communicate with the dubbing server 220 through a wired or wireless network, acquire a video file from the dubbing server 220 and play it, and display dubbing trigger information, such as a dubbing trigger button, on the playing interface during playback. In response to the user triggering the dubbing trigger button, communication with the dubbing server completes the subsequent dubbing operation.
  • dubbing trigger information such as a dubbing trigger button
  • FIG. 3 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application. As shown in FIG. 3, the method includes the following steps.
  • Step 301 Send a first video and information indicating that voiceover is allowed to the first client, so that the first client plays the first video, and the information that allows voiceover according to the indication is playing the first video.
  • the dubbing trigger information is displayed during the process.
  • Step 302 Receive a first dubbing request generated by the second client according to the triggering operation of the dubbing trigger information.
  • Step 303 Acquire a second video according to an identifier (Identifier, ID) of the first video carried in the first voiceover request.
  • ID an identifier
  • Step 304 Send the second video to the second client, where the second video is obtained by removing first audio information of a character in the first video.
  • Step 305 Receive second audio information of the character collected according to the played second video from the second client, and perform synthesis processing on the second video and the second audio information of the character. Obtaining a third video, or receiving, by the second client, the third video obtained by the second client by using the second video and the second audio information to perform a composite process.
  • the first video and the information indicating the allowed dubbing are sent to the first client.
  • the first client plays the first video, and according to the indication, the information of the dubbing is displayed to display the dubbing trigger information during the playing of the first video.
  • the second video is obtained by removing first audio information of a character in the first video.
  • the dubbing trigger information may be triggered on the interface of playing the video during the process of playing the video, and the dubbing operation may be performed.
  • FIG. 4 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application. As shown in FIG. 4, the method includes the following steps.
  • Step 401 The first client performs a trigger operation on the dubbing trigger information displayed by the second client during the playing of the first video to generate a dubbing request carrying the ID of the first video.
  • Step 402 The first client sends the dubbing request to a dubbing server.
  • Step 403 The first client receives a second video from the voice-over server.
  • the second video is obtained by removing first audio information of a character in the first video.
  • Step 404 The first client plays the second video, and collects second audio information for the character according to the played second video.
  • Step 405 The first client sends the second audio information to the voiceover server.
  • the dubbing trigger information can be triggered on the interface of playing the video during the process of playing the video, and the dubbing operation can be performed, thereby improving the dubbing efficiency and saving system resources.
  • the video and dubbing trigger information can be played on the second client, and the dubbing trigger information is triggered by the first client, and the dubbing operation is implemented by using the interaction between the first client and the dubbing server. Therefore, two devices can be utilized.
  • the video playback and dubbing operations are implemented separately to realize distributed processing of tasks and reasonable allocation of system load.
  • the dubbing operation performed by the first client does not affect the video playback of the second client, Conducive to the parallel processing of tasks.
  • FIG. 5 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application. As shown in FIG. 5, the method includes the following steps.
  • Step 501 Receive a first video from the dubbing server and information indicating that dubbing is allowed.
  • Step 502 Play the first video, and display the dubbing trigger information on the interface for playing the first video according to the information that allows the dubbing according to the indication.
  • Step 503 Generate a dubbing request carrying an ID of the first video in response to a triggering operation on the dubbing trigger information.
  • Step 504 Send the dubbing request to the dubbing server.
  • Step 505 Receive a second video from the voice-over server, where the second video is obtained by removing first audio information of a character in the first video.
  • Step 506 playing the second video.
  • Step 507 Collect second audio information for the character according to the played second video.
  • Step 508 Send the second audio information to the voiceover server.
  • the terminal device plays the first video, and displays the dubbing trigger information on the interface that plays the first video, and interacts with the dubbing server to complete the dubbing operation.
  • the media information processing operation can be completed on a terminal device having a video playing and dubbing function.
  • FIG. 6 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • the triggering operation of the dubbing is completed by the first client and the second client.
  • the first client and the second client may be the same terminal or different terminals.
  • the first client and the second client are different terminals.
  • the description is made by taking the first client as the TV and the second client as the mobile phone.
  • the processed media information is an advertisement video, and the advertisement video includes two characters.
  • the mobile phone and the tablet cooperate to complete the dubbing operation of the two characters as an example.
  • the method includes the following steps.
  • Step 601 The television acquires the first advertisement video and the information indicating the allowed dubbing from the dubbing server.
  • the television can be a smart television or a non-smart television installed in a vehicle or elevator, such as a video playback terminal.
  • the dubbing server can save the first advertisement video, and then the television acquires the first advertisement video from the dubbing server.
  • the dubbing server may also be only the dubbing control server, and the first advertisement video is not stored, and the television obtains the first advertisement video from the database storing the first advertisement video through the dubbing server.
  • the first advertisement video includes screen information of the advertisement and pre-recorded audio information, and plays audio corresponding to the screen while playing the screen.
  • the television may synchronously acquire the first advertisement video and the information indicating the allowed dubbing from the dubbing server; or obtain the first advertisement video first, and then obtain the information indicating that the dubbing is allowed.
  • the indication allows the information of the dubbing to be used to indicate that the first advertisement video is allowed to be dubbed.
  • Step 602 The television plays the first advertisement video, and according to the instruction, the information of the voice-over is allowed to display the voice-over trigger information on the interface for playing the first advertisement video during the process of playing the first advertisement video.
  • the dubbing trigger information may be displayed on the interface that plays the first advertisement video in a floating form, or may be displayed in an embedded manner on the interface that plays the first advertisement video.
  • FIG. 6A is a schematic diagram of a dubbing request triggering interface in media information processing provided by an example of the present application.
  • the dubbing trigger information that is, the two-dimensional code information
  • FIG. 6B is a schematic diagram of a dubbing request triggering interface in media information processing provided by an example of the present application.
  • a dubbing trigger button is displayed in the lower right corner of the interface for playing the first advertisement video, when the user uses the remote
  • the controller triggers the dubbing trigger button
  • the television generates a two-dimensional code display request in response to the triggering operation of the dubbing trigger button, and searches for a two-dimensional code corresponding to the two-dimensional code display request, and triggers the interface on the dubbing request.
  • a two-dimensional code as shown in Fig. 6A is displayed.
  • Step 603 The mobile phone runs a two-dimensional code scanning function, and triggers the dubbing trigger information, that is, scans the two-dimensional code to generate a first dubbing request carrying the ID of the first advertisement video.
  • the mobile phone scans the two-dimensional code shown in FIG. 6A using its own two-dimensional code scanning function to generate a first dubbing request.
  • the first dubbing request carries an ID of the first advertisement video.
  • Step 604 The mobile phone sends the first dubbing request to the dubbing server.
  • the mobile phone scans the two-dimensional code shown in FIG. 6A by using its own two-dimensional code scanning function, obtains the IP address of the dubbing server from the two-dimensional code, and the first dubbing according to the IP address of the dubbing server. The request is sent to the dubbing server.
  • Step 605 The dubbing server finds the second advertisement video according to the ID of the first advertisement video carried in the first dubbing request.
  • the second advertisement video is obtained by removing the first audio information of the character in the first advertisement video.
  • the dubbing server may find the first advertisement video according to the ID of the first advertisement video, remove the first audio information of the character in the first advertisement video, and obtain the second advertisement video; or the dubbing server according to the first
  • the second advertisement video is found by the ID of the advertisement video and the pre-stored correspondence between the ID of the first advertisement video and the second advertisement video.
  • Step 606 The dubbing server sends the dubbing interface, the description information of the second advertisement video, and the information indicating that the second advertisement video is allowed to be played to the mobile phone.
  • the dubbing server can send a dubbing interface in the form of a web page to the mobile phone.
  • the dubbing server searches for the description information of the second advertisement video according to the ID of the first advertisement video carried in the first dubbing request, and sends the description information to the mobile phone.
  • Step 607 The mobile phone displays the voice-over interface, and displays a play trigger button of the second advertisement video on the voice-over interface according to the description information of the second advertisement video and the information indicating that the second advertisement video is allowed to be played.
  • step 608 the voice server determines whether the first advertisement video includes one character or multiple roles.
  • step 609 is performed.
  • the voiceover server sends the second advertisement video and the line information to the mobile phone, indicating that the mobile phone is
  • the second advertisement video and the line information played are the audio information collected for the character.
  • the mobile phone sends the audio information collected by the character to the voice-over server, and the voice-over server synthesizes the second advertisement video and the audio information to obtain a synthesized third advertisement video.
  • the dubbing server sends information indicating that the individual dubbing is allowed to be selected to the mobile phone and information indicating that the cooperative dubbing is allowed to be selected.
  • Step 610 The mobile phone allows to select the information of the individual dubbing according to the indication and the information indicating that the co-over dubbing is allowed to be displayed, and the individual dubbing trigger button and the cooperative dubbing trigger button are displayed on the dubbing interface.
  • the mobile phone does not send a request for cooperative dubbing to the dubbing server, and the default user selects a separate dubbing mode.
  • Step 611 The mobile phone generates a second dubbing request in response to the triggering operation of the cooperative dubbing trigger button by the user, and sends the second dubbing request to the dubbing server.
  • the dubbing server will use the second advertisement.
  • the video is sent to the phone and instructs the phone to collect audio information for all characters.
  • Step 612 the dubbing server selects the third client as the cooperative dubbing client.
  • the third client is a tablet.
  • the dubbing server selects the third client as the cooperative dubbing client.
  • Step 613 The dubbing server acquires the information of the first character and the information of the second role from the first advertisement video according to the second dubbing request, and sends the information to the mobile phone and the tablet.
  • the information of the first role and the information of the second role may be acquired first, and then the third client is selected as the cooperative voiceover client.
  • Step 614 The mobile phone and the tablet display the first character selection button and the second character selection button on the voiceover interface according to the information of the first character and the information of the second character.
  • the handset can also display a randomly assigned character trigger button on the voice over interface.
  • FIG. 6C is a schematic diagram of a role selection interface in media information processing provided by an example of the present application.
  • a first character selection button, a second character selection button, and a randomly assigned character trigger button are displayed in the interface.
  • the mobile phone When the user triggers the randomly assigned role trigger button, the mobile phone generates a voiceover request carrying the information indicating the randomly assigned role, and the voiceover server randomly assigns a role to the mobile phone according to the voiceover request for randomly assigning the role information.
  • Steps 609 to 614 complete the operation of selecting the first role by the mobile phone and selecting the second character by the tablet.
  • step 615 the dubbing server assigns the first role to the mobile phone.
  • the mobile phone generates a third dubbing request carrying the information of the first character in response to the triggering operation of the first character selection button by the user, and sends the third dubbing request to the dubbing server, where the dubbing server is the mobile phone Assign the first character.
  • step 616 the dubbing server assigns a second role to the tablet.
  • the tablet generates a fourth dubbing request carrying information of the second character in response to a triggering operation of the second character selection button by the user, and sends the fourth dubbing request to the dubbing server, the dubbing server assigning the dubbing server to the tablet Second role.
  • the third dubbing request and the fourth dubbing request are typically not sent simultaneously.
  • the mobile phone first sends a third dubbing request carrying information of the first role
  • the dubbing server assigns the first role to the mobile phone, and instructs the mobile phone and the tablet to select the first role.
  • the button is set to an unselectable state. In this case, the user of the tablet can only choose to trigger the second character selection button to generate a fourth dubbing request carrying the information of the second character.
  • Step 617 The mobile phone generates a second advertisement video play request in response to the trigger operation of the user to trigger the play trigger button of the second advertisement video, and sends the second advertisement video play request to the voice server.
  • Step 618 The dubbing server sends the content of the second advertisement video to the mobile phone in time sequence, and when the line information is read, determines whether the line information belongs to the first role or the second role, and when the line information is determined to belong to the first In the case of a role, step 619 is performed; otherwise, step 624 is performed.
  • Step 619 The voice-over server sends a first audio information collection instruction and the word information to the mobile phone.
  • Step 620 The mobile phone invokes its own audio information collecting device according to the first audio information collecting instruction, plays the received second advertising video and the line information, and collects the second audio information for the first character according to the played line information.
  • FIG. 6D is a schematic diagram of an interface for collecting second audio information on a mobile phone in media information processing provided by an example of the present application.
  • the second advertisement video is played above the dubbing interface, and the speech information of the first character is displayed on the interface for playing the second advertisement video.
  • the lower part of the dubbing interface Shows the progress of the second audio information collection.
  • the progress of the second audio information collection is consistent with the playback progress of the word information.
  • a play trigger button of the second audio information of the first character recorded before the recording and a play trigger button of the third character third audio information recorded before the recording may be displayed on the voice-over interface.
  • the mobile phone user triggers the second audio information of the previously recorded first character
  • the mobile phone plays the second audio information, and displays a dynamic play icon next to the play trigger button of the second audio information of the first character.
  • Step 621 When the dubbing server determines that the line information is sent, the first closing instruction is sent to the mobile phone.
  • Step 622 The mobile phone turns off its own audio information collecting device according to the first closing command, and sends the second audio information to the dubbing server.
  • FIG. 6E is a schematic diagram of an acquisition end interface in a media information processing process provided by an example of the present application.
  • the audio server After the playing of the first character's speech information, the audio server sends an acquisition end interface to the mobile phone, and the mobile phone displays the collection end interface, and displays the completion information on the interface to notify the user that the dubbing process ends.
  • Step 623 the dubbing server synthesizes the second video and the second audio information of the first character to obtain a third advertisement video.
  • Step 624 The dubbing server sends a second audio information collection instruction and the line information to the tablet.
  • Step 625 The tablet computer calls its own audio information collecting device according to the second audio information collecting instruction, plays the received second advertising video and the line information, and collects the third audio information for the second character according to the played line information. .
  • Step 626 When the dubbing server determines that the line information is sent, the second closing instruction is sent to the tablet.
  • Step 627 The tablet computer turns off its own audio information collecting device according to the second closing instruction, and sends the third audio information to the voice-over server.
  • steps 624-627 are repeatedly executed until all the line information of the second character is played, and the corresponding third audio information is collected and sent to the dubbing server according to the line information of the second character.
  • Step 628 the dubbing server synthesizes the second video and the third audio information of the second character to obtain a fourth advertisement video.
  • Step 629 after receiving the second video information of the first character from the mobile phone, and receiving the third video information of the second character from the tablet, performing the second advertisement video, the second audio information, and the third audio information.
  • the synthesis process results in a fifth advertisement video.
  • FIG. 6F is a schematic diagram of an interface for playing audio information of a character during media information processing according to an example of the present application. As shown in FIG.
  • a play trigger button of the second audio information of the first character and a play trigger button of the third audio information of the second character are displayed in the interface.
  • the client participating in the dubbing that is, the mobile phone or tablet can request to change the dubbing partner.
  • the phone is playing the third sound collected by the tablet.
  • the user thinks that the third audio is not the desired audio, and then initiates a partner replacement request by the mobile phone. For example, as shown in FIG. 6F, the user performs a right swipe operation under the third audio on the dubbing interface.
  • the mobile phone Responding to the user's right-wiping operation, the mobile phone generates the partner replacement request, and sends the partner replacement request to the dubbing server, and the dubbing server acquires a client that initiates a dubbing request carrying the ID of the first advertisement video.
  • Information selecting a client as the dubbing partner of the mobile phone, assigning the role to the selected client, and interacting with the selected client, obtaining the second role from the selected client Four audio messages are sent to the phone display.
  • FIG. 7 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • the first client and the second client are the same terminal.
  • the terminal is a personal computer (PC) as an example.
  • the processed media information is a video and video, and the first video and video includes a role as an example.
  • the method includes the following steps.
  • step 701 the PC runs the player.
  • the PC can run the player program or run the player on a web page.
  • Step 702 The PC acquires the first video and video from the dubbing server and the information indicating that the dubbing is allowed.
  • Step 703 The PC plays the first video and video in the player, and the information that allows the dubbing according to the indication displays the dubbing trigger information on the interface that plays the first video and video.
  • the dubbing trigger information may be a dubbing trigger button.
  • Step 704 when the user clicks the dubbing trigger button by using an input device, such as a mouse, the PC generates a dubbing request carrying the ID of the first video and video, and generates a dubbing program call instruction.
  • an input device such as a mouse
  • step 705 the PC uses the dubbing program call instruction to run the dubbing program installed by itself.
  • Step 706 the dubbing program sends the dubbing request to the dubbing server.
  • Step 707 The voice-over server acquires the second video and video according to the ID of the first video and video carried in the voice-over request.
  • the voice-over server can obtain the first video and video according to the ID of the first video and video carried in the voice-over request, and remove the first audio information of the character in the first video and video to obtain the second video and video.
  • the dubbing server searches for the pre-acquired second video and video according to the ID of the first video and video and the pre-stored correspondence between the ID of the first video and the second video.
  • Step 708 the dubbing server sends the second video and video to the dubbing program.
  • Step 709 the dubbing program acquires description information of the second video and video, and displays description information of the second video and video and a play trigger button of the second video and video in the dubbing interface.
  • Step 710 When the user clicks the play trigger button of the second video and video, the dubbing program plays the second video and video in the dubbing interface.
  • Step 711 When the voice program reads the line information of the character in the second video and video, the audio information collection instruction is sent to the audio collection device of the PC.
  • Step 712 The PC calls its own audio information collecting device according to the audio information collecting instruction, and collects second audio information of the character.
  • Step 713 When the dubbing program determines that the speech information is finished playing, send a shutdown command to the PC.
  • Step 714 the PC turns off the audio information collecting device according to the closing instruction.
  • steps 711 to 714 are performed until all the line information is played.
  • Step 715 After the second audio information of the character is collected, the dubbing program performs a synthesizing process on the second video and the second audio information to obtain a third video.
  • Step 716 the dubbing program sends the third video to the dubbing server.
  • FIG. 8 is a schematic flowchart diagram of a method for processing media information provided by an example of the present application.
  • the dubbing server performs subsequent processing on the video. The method includes the following steps.
  • Step 801 The dubbing server acquires the synthesized video obtained by using the above method.
  • Step 802 the dubbing server obtains audio information from the video.
  • step 803 the dubbing server converts the audio information into text information.
  • Step 804 the dubbing server determines whether the text information includes information matching the pre-stored keyword. When it is determined that the text information includes information matching the pre-stored keyword, step 805 is performed; otherwise, step 808 is performed.
  • the keyword can be: fun, Shanghai dialect, reverse string, and the like. For example, if the text message contains "thank you”, compare “thank you” with the pre-stored data in the database, and find that the "thank you” matches the keyword "Shanghai dialect", then the text information is included and pre-stored. Keyword matching information. The textual information can be matched to a plurality of pre-stored keywords.
  • Step 805 Set a first association relationship between the ID of the video and the pre-stored keyword.
  • Step 806 when receiving the first video presentation request sent by the first client, obtaining the ID of the video from the first video presentation request, acquiring the video and the pre-stored keyword according to the ID of the video, and the video And the pre-stored keyword is sent to the first client.
  • Step 807 The first client can play the video and display the pre-stored keyword on the play interface.
  • FIG. 8A is a schematic diagram of a video playing interface provided by the implementation of the present application.
  • the first client can play the video in the interface to display keywords received from the dubbing server.
  • the first client can share the video playing interface in the form of a webpage, for example, by clicking the "Publish Dub” button on the interface to share the video playing interface to the social network.
  • Audio processing buttons such as “Lolly”, “Electronic Sound”, “Cat” and “Smoke” can also be displayed on the interface shown in FIG. 8A. For example, when the user clicks on the "cat” button, the first client can process the audio in the video into the form of a cat sound.
  • the acquired video may be displayed in the form of a video wall.
  • FIG. 8B is a schematic diagram of a video wall provided by an example of the present application. Multiple videos can be displayed on the video wall, and users can like the videos they are interested in.
  • Step 808 The dubbing server may create a custom keyword corresponding to the text information, and set a second association relationship between the ID of the video and the custom keyword.
  • Step 809 when receiving a second video presentation request sent by the second client, obtaining an ID of the video from the second video presentation request, acquiring the video and the customized keyword according to the ID of the video, and the video and the video The custom keyword is sent to the second client.
  • the second client can play the video and display the customized keyword on the play interface.
  • the video and the custom keyword may also be published in the manner described in 807.
  • the dubbing server stores the original video, for example, the first video in FIG. 3 to FIG. 7 above, and may also store the synthesized video synthesized by using the audio information collected by the above example, which is called a dubbed video.
  • a dubbed video For example, the third video in FIGS. 3 to 7 above.
  • the client plays the original video and displays the video switching trigger information on the interface that plays the original video.
  • the client When the user triggers the video switching trigger information, the client generates a video switching request and sends the video switching request to the voiceover server.
  • FIG. 8C is a schematic diagram of a video switching interface in media information processing provided by an example of the present application.
  • a switching button of "original video/dubbing version video" is displayed, by which a switching request can be sent to the dubbing server to acquire and play another version of the video.
  • FIG. 9 is a schematic structural diagram of a media information processing server according to an example of the present application.
  • the server includes: a first video sending module 901, a first dubbing request receiving module 902, a first video obtaining module 903, a second video sending module 904, and a second video acquiring module 905.
  • the first video sending module 901 is configured to send, to the first client, a first video and information indicating that the voice-over is allowed, so that the first client plays the first video, and the information of the voice-over is allowed according to the indication.
  • the dubbing trigger information is displayed on the interface that plays the first video.
  • the first dubbing request receiving module 902 is configured to receive a first dubbing request generated by the second client according to the triggering operation of the dubbing trigger information.
  • the first video acquisition module 903 is configured to acquire a second video according to the ID of the first video that is carried in the first voice-over request, where the second video is removed from the first video.
  • the first audio information of the character is obtained.
  • the second video sending module 904 is configured to send the second video to the second client.
  • the second video obtaining module 905 is configured to receive second audio information of the character collected according to the played second video from the second client, for the second video and the Performing a synthesis process on the second audio information of the character to obtain a third video; or receiving, from the second client, the first process obtained by the second client using the second video and the second audio information to perform a synthesis process Three videos.
  • the server further includes: a role judging module 906, a dubbing interface sending module 907, a second dubbing request receiving module 908, a cooperative dubbing client selecting module 909, a role information sending module 910, a role assigning module 911, and The third video acquisition module 912.
  • the role determining module 906 is configured to determine whether the role includes multiple roles.
  • the voice-over interface sending module 907 is configured to, when determining that the character includes multiple roles, send a voice-over interface to the second client, and information indicating that the individual voice-over is allowed to be selected and information indicating that the cooperative voice-over is allowed to be selected, so that the The second client allows the selection of the individual dubbing information and the information indicating the selection of the cooperative dubbing according to the indication to display the individual dubbing trigger button and the cooperative dubbing trigger button on the dubbing interface.
  • the second dubbing request receiving module 908 is configured to receive a second dubbing request from the second client, where the second dubbing request is a triggering operation by the second client in response to the user to the cooperative dubbing trigger button And generated.
  • the cooperative voice-over client selection module 909 is configured to acquire the information of the first role and the information of the second role according to the second voice-over request, and select the third client as the cooperative voice-over client.
  • the role information sending module 910 is configured to send information about the first role and information about the second role to the second client and the third client, respectively, so that the second client And the third client displays the first character selection button and the second role selection button according to the information of the first character and the information of the second character.
  • the role assignment module 911 is configured to respond to a third voiceover request that is sent by the second client to carry information about the first role, and a third message that is sent by the third client to carry the second role.
  • a fourth dubbing request, the first character and the second role are respectively assigned to the second client and the third client; wherein the third dubbing request and the fourth dubbing request are respectively The second client and the third client are generated in response to a triggering operation of the first character selection button and the second character selection button by the user.
  • the second video obtaining module 905 is further configured to receive second audio information of the first character that is collected according to the played second video from the second client.
  • the third video obtaining module 912 is configured to receive third audio information of the second character collected according to the played second video from the third client, to the second video and the The third audio information of the second character performs a synthesizing process to obtain a fourth video.
  • the server further includes: a conversion module 913, a keyword determination module 914, and a processing module 915.
  • the conversion module 913 is configured to convert the second audio information of the character into text information.
  • the keyword determining module 914 is configured to determine whether the text information includes information that matches a pre-stored keyword.
  • the processing module 915 is configured to: when determining that the text information includes information that matches the pre-stored keyword, set a first association relationship between an ID of the third video and the pre-stored keyword, Receiving, by the fourth client, a first video presentation request that carries an ID of the third video, and searching for the pre-stored keyword and the third video according to the ID of the third video and the first association relationship, The third video and the pre-stored keyword are sent to the fourth client for display, and when it is determined that the text information does not include information matching the pre-stored keyword, setting corresponding to the text information Customizing a keyword, setting a second association relationship between the ID of the third video and the custom keyword, and receiving, by the fifth client, a second video presentation request that carries an ID of the third video, according to The ID of the third video and the second association relationship find the customized keyword and the third video, and send the third video and the customized keyword to the fifth client Show it.
  • the server further includes: a handover indication information sending module 916, a video switching request receiving module 917, a fourth video acquiring module 918, and a third video sending module 919.
  • the change indication information sending module 916 is configured to send the first video to the first client, and send information indicating that video switching is allowed to the first client, so that the first client plays the Decoding the first video, and displaying video switching trigger information on the interface for playing the first video according to the information indicating that the video switching is allowed.
  • the video switching request receiving module 917 is configured to receive, from the first client, a video switching request that carries an ID of the third video, where the video switching request is that the first client responds to the user Generated by the triggering operation of the video switching trigger information.
  • the fourth video obtaining module 918 is configured to acquire the third video according to the ID of the third video.
  • the third video sending module 919 is configured to send the third video to the first client, so that the first client plays the third video.
  • the first client and the second client may be the same client or different clients.
  • FIG. 10 is a schematic structural diagram of a media information processing client according to an example of the present application.
  • the client includes: a dubbing request generating module 1001, a dubbing request sending module 1002, a video receiving module 1003, a video playing module 1004, an audio collecting module 1005, and an audio sending module 1006.
  • a dubbing request generating module 1001 configured to perform a triggering operation on the dubbing trigger information displayed by another client on the interface for playing the first video, to generate a dubbing request carrying the identifier ID of the first video;
  • a dubbing request sending module 1002 configured to send the dubbing request to the dubbing server, so that the dubbing server acquires the second video according to the ID of the first video carried in the dubbing request, where the second The video is obtained by removing the first audio information of the character in the first video;
  • a video receiving module 1003, configured to receive a second video from the voice-over server
  • the audio collection module 1005 is configured to collect second audio information for the role according to the played second video.
  • the audio sending module 1006 is configured to send the second audio information to the voiceover server.
  • the client further includes: a first information receiving module 1007, a first display module 1008, a video play request generating module 1009, and a video play request sending module 1010.
  • the first information receiving module 1007 is configured to receive, from the voice-over server, a voice-over interface in the form of a webpage sent by the voice-over server to the client according to the voice-over request, where the description information of the second video and the indication allow playback Information of the second video.
  • the first display module 1008 is configured to display the voice-over interface, and display a play trigger button of the second video according to the description information of the first video and the information indicating that the second video is allowed to be played.
  • the video play request generating module 1009 is configured to generate a second video play request in response to a user triggering a play trigger button of the second video.
  • the video play request sending module 1010 is configured to send the second video play request to the voice-over server, so that the voice-over server sends the second video to the client according to the second video play request. .
  • the client further includes: a dubbing program running module 1011.
  • the dubbing program running module 1011 is configured to run a self-installed dubbing program according to the dubbing request.
  • the video receiving module 1003 is further configured to receive, in a voice-over interface of the voice-over program, the second video that the voice-over server pushes to the client according to the voice-over request.
  • the client further includes: a first audio information collection instruction receiving module 1012, a first shutdown instruction receiving module 1013, and a first shutdown processing module 1014.
  • the first audio information collection instruction receiving module 1012 is configured to receive, when the voice distribution server sends the second video and the line information of the character to the client in time sequence, receive the first sent by the voiceover server. Audio information collection instructions.
  • the video playing module 1004 is further configured to play the second video and the line information of the character.
  • the audio collection module 1005 is further configured to acquire the second audio information for the role according to the audio information collection device that invokes itself according to the first audio information collection instruction.
  • the first closing instruction receiving module 1013 is configured to receive a first closing instruction, where the first closing instruction is sent by the voice-over server to the client when the line information of the character is sent.
  • the first shutdown processing module 1014 is configured to close the audio information collection device according to the first shutdown instruction.
  • the client further includes: a second audio information collection instruction receiving module 1015, a second shutdown instruction receiving module 1016, and a second shutdown processing module 1017.
  • the second audio information collection instruction receiving module 1015 is configured to receive the second audio information from the voiceover program when the second video is played on the voice-over interface in time sequence and played to the character information of the character. Acquisition instructions.
  • the audio collection module 1005 is further configured to: call the audio information collection device of the second audio information collection instruction to acquire the second audio information for the role.
  • the second closing instruction receiving module 1016 is configured to receive a second closing instruction from the dubbing program when the character information of the character is played.
  • the second shutdown processing module 1017 is configured to close the audio information collection according to the second shutdown instruction. Device.
  • FIG. 11 is a schematic structural diagram of a media information processing client according to an example of the present application.
  • the client includes: a video receiving module 1101, a video playing module 1102, a dubbing request generating module 1103, a dubbing request sending module 1104, an audio collecting module 1105, and an audio information transmitting module 1106.
  • the video receiving module 1101 is configured to receive a first video from the voice-over server and information indicating that voice-over is allowed.
  • the video playing module 1102 is configured to play the first video, and display the dubbing trigger information on the interface that plays the first video according to the information that allows the dubbing according to the indication.
  • the voice-over request generating module 1103 is configured to generate a voice-over request carrying an ID of the first video in response to a triggering operation on the voice-over trigger information.
  • the dubbing request sending module 1104 is configured to send the dubbing request to the dubbing server, so that the dubbing server obtains the second video according to the ID of the first video carried in the dubbing request, where the The two videos are obtained by removing the first audio information of the character in the first video.
  • the video receiving module 1101 is further configured to receive a second video from the voice-over server.
  • the video playing module 1102 is further configured to play the second video.
  • the audio collection module 1105 is configured to collect second audio information for the character according to the played second video.
  • the audio information sending module 1106 is configured to send the second audio information to the voiceover server.
  • FIG. 12 is a schematic structural diagram of a media information processing server in an example of the present application.
  • the device includes a processor 1201, a non-volatile computer readable memory 1202, a display unit 1203, and a network communication interface 1204. These components communicate over bus 1205.
  • a plurality of program modules are stored in the memory 1202, including an operating system 1206, a network communication module 1207, and an application 1208.
  • the processor 1201 can read various modules (not shown) included in the application in the memory 1202 to perform media information processing various functional applications and data processing.
  • the processor 1201 in this example may be one or multiple, and may be a CPU, a processing unit/module, an ASIC, a logic module, or a programmable gate array.
  • the operating system 1206 can be: a Windows operating system, an Android operating system, or an Apple iPhone OS operating system.
  • the application 1208 can include a media information processing module 1209.
  • the media information processing module 1209 can include the computer executable instruction set 1209-1 and the corresponding metadata and heuristic algorithm 1209-2 formed by the modules in FIG. 9 described above. These sets of computer executable instructions may be executed by the processor 1201 and perform the functions of the method illustrated in Figures 3, 6-8, or the media information processing server illustrated in Figure 9.
  • the network communication interface 1204 cooperates with the network communication module 1207 to complete transmission and reception of various network signals of the media information processing server.
  • the display unit 1203 has a display panel for completing input and display of related information.
  • FIG. 13 is a schematic structural diagram of a media information processing client in an example of the present application.
  • the device includes a processor 1301, a non-volatile computer readable memory 1302, a display unit 1303, and a network communication interface 1304. These components communicate over bus 1305.
  • a plurality of program modules are stored in the memory 1302, including an operating system 1306, a network communication module 1307, and an application 1308.
  • the processor 1301 can read various modules (not shown) included in the application in the memory 1302 to perform media information processing various function applications and data processing.
  • the processor 1301 in this example may be one or multiple, and may be a CPU, a processing unit/module, an ASIC, a logic module, or a programmable Chengmen array and so on.
  • the operating system 1306 can be: a Windows operating system, an Android operating system, or an Apple iPhone OS operating system.
  • the application 1308 can include a media information processing module 1309.
  • the media information processing module 1309 can include the computer executable instruction set 1309-1 and the corresponding metadata and heuristic algorithm 1309-2 formed by the modules in FIG. 10 described above. These sets of computer executable instructions may be executed by the processor 1301 and perform the functions of the method illustrated in Figures 4, 6-8, or the media information processing client illustrated in Figure 10.
  • the network communication interface 1304 cooperates with the network communication module 1307 to complete the transmission and reception of various network signals of the media information processing client.
  • the display unit 1303 has a display panel for completing input and display of related information.
  • FIG. 14 is a schematic structural diagram of a media information processing client in an example of the present application.
  • the device includes a processor 1401, a non-volatile computer readable memory 1402, a display unit 1403, and a network communication interface 1404. These components communicate over bus 1405.
  • a plurality of program modules are stored in the memory 1402, including an operating system 1406, a network communication module 1407, and an application 1408.
  • the processor 1401 can read various modules (not shown) included in the application in the memory 1402 to perform media information processing various functional applications and data processing.
  • the processor 1401 in this example may be one or multiple, and may be a CPU, a processing unit/module, an ASIC, a logic module, or a programmable gate array.
  • the operating system 1406 can be: a Windows operating system, an Android operating system, or an Apple iPhone OS operating system.
  • the application 1408 can include a media information processing module 1409.
  • the media information processing module 1409 can include the computer executable instruction set 1409-1 formed by each module in FIG. 11 and the corresponding metadata and heuristic algorithm 1409-2. These sets of computer executable instructions may be executed by the processor 1401 and perform the functions of the method illustrated in Figures 5-8 or the media information processing client illustrated in Figure 11.
  • the network communication interface 1404 cooperates with the network communication module 1407 to complete the transmission and reception of various network signals of the media information processing client.
  • the display unit 1403 has a display panel for completing input and display of related information.
  • each functional module in each example of the present application may be integrated into one processing unit, or each module may exist physically separately, or two or more modules may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the functional modules of the examples may be located at one terminal or network node, or may be distributed to multiple terminals or network nodes.
  • each of the examples of the present application can be implemented by a data processing program such as a computer.
  • the data processing program constitutes the present application.
  • a data processing program usually stored in a storage medium is executed by directly reading a program out of a storage medium or by installing or copying the program to a storage device (such as a hard disk and or a memory) of the data processing device. Therefore, such a storage medium also constitutes the present application.
  • the storage medium can use any type of recording method, such as paper storage medium (such as paper tape, etc.), magnetic storage medium (such as floppy disk, hard disk, flash memory, etc.), optical storage medium (such as CD-ROM, etc.), magneto-optical storage medium ( Such as MO, etc.).
  • the present application therefore also provides a storage medium in which is stored a data processing program, such as machine readable instructions, for performing any of the above examples of the methods of the present application.
  • the steps to achieve the above examples may be by hardware.
  • the completion may also be performed by a program to instruct related hardware.
  • the program may be stored in a computer readable storage medium.
  • the storage medium mentioned above may be a read only memory, a magnetic disk or an optical disk.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

本申请公开了一种媒体信息处理方法,包括:向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求;根据所述第一配音请求中携带的所述第一视频的ID,获取第二视频;将所述第二视频发送给所述第二客户端;从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收合成处理得到的所述第三视频。

Description

媒体信息处理方法和装置
本申请要求于2016年11月24日提交中国专利局、申请号为2016110508334、发明名称为“媒体信息处理方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及多媒体技术领域,尤其涉及一种媒体信息处理方法和装置。
背景技术
随着互联网技术的发展,越来越多的数据(包括:文本、图片、音频、视频等)会通过互联网推送给各种用户。比如:用户在使用手机、PC、TV等终端设备时,会接收到网络侧推送的各种数据,比如:图片或视频格式的广告、公益宣传信息、新闻等。这样,用户可以及时获知时讯、感兴趣的内容等。这类数据可被称为推送信息或推送媒体内容等。
发明内容
本申请实例提供一种媒体信息处理方法,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理方法,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理方法,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理服务器,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理客户端,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理客户端,可以提高配音效率,节省系统资源。
本申请实例提供一种媒体信息处理方法,应用于配音服务器,包括:
向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求;
根据所述第一配音请求中携带的所述第一视频的标识符ID,获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
将所述第二视频发送给所述第二客户端;
从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。
本申请实例提供一种媒体信息处理方法,包括:
第一客户端对第二客户端在播放第一视频的界面上显示的配音触发信息进行触发操作生成携带所述第一视频的标识符ID的配音请求;
所述第一客户端将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
所述第一客户端从所述配音服务器接收第二视频;
所述第一客户端播放所述第二视频,并根据播放的所述第二视频为所述角色采集第二音频信息;
所述第一客户端将所述第二音频信息发送给所述配音服务器。
本申请实例提供一种媒体信息处理方法,应用于媒体信息处理客户端,包括:
从配音服务器接收第一视频和指示允许配音的信息;
播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
响应于对所述配音触发信息的触发操作生成携带所述第一视频的标识符ID的配音请求;
将所述配音请求发送给所述配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
从所述配音服务器接收所述第二视频;
播放所述第二视频;
根据播放的所述第二视频为所述角色采集第二音频信息;
将所述第二音频信息发送给所述配音服务器。
本申请实例提供一种媒体信息处理服务器,包括:处理器和存储器,所述存储器存储:
第一视频发送模块,用于向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
第一配音请求接收模块,用于接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求;
第一视频获取模块,用于根据所述第一配音请求中携带的所述第一视频的标识符ID,获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
第二视频发送模块,用于将所述第二视频发送给所述第二客户端;
第二视频获取模块,用于从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,用于对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频;或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。
本申请实例提供一种媒体信息处理客户端,包括:处理器和存储器,所述存储器存储:
配音请求生成模块,用于对另一客户端在播放第一视频的界面上显示的配音触发信息进行触发操作生成携带所述第一视频的标识符ID的配音请求;
配音请求发送模块,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
视频接收模块,用于从所述配音服务器接收第二视频;
视频播放模块,用于播放所述第二视频;
音频采集模块,用于并根据播放的所述第二视频为所述角色采集第二音频信息;
音频发送模块,用于将所述第二音频信息发送给所述配音服务器。
本申请实例提供一种媒体信息处理客户端,包括:处理器和存储器,所述存储器存储:
视频接收模块,用于从配音服务器接收第一视频和指示允许配音的信息;
视频播放模块,用于播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
配音请求生成模块,用于响应于对所述配音触发信息的触发操作生成携带所述第一视频的标识符ID的配音请求;
配音请求发送模块,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
所述视频接收模块,进一步用于从所述配音服务器接收第二视频;
所述视频播放模块,进一步用于播放所述第二视频;
音频采集模块,用于根据播放的所述第二视频为所述角色采集第二音频信息;
音频信息发送模块,用于将所述第二音频信息发送给所述配音服务器。
在本申请实例中,向第一客户端发送第一视频和指示允许配音的信息。该第一客户端播放该第一视频,并根据该指示允许配音的信息在播放该第一视频的过程中显示配音触发信息。接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求。根据该第一配音请求中携带的该第一视频的ID,获取第二视频。将所述第二视频发送给所述第二客户端。其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。在本申请实例中,可以在播放视频的过程中,在播放视频的界面上触发配音触发信息,执行配音操作。利用本方案提供的媒体信息处理方法,可以减少客户端和配音服务器之间的交互操作,提高配音效率,节省系统资源。
附图说明
为了更清楚地说明本申请实例或现有技术中的技术方案,下面将对实例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本申请各实例涉及的一种实施环境的结构示意图;
图2为本申请各实例涉及的一种实施环境的结构示意图;
图3为本申请实例提供的一种媒体信息处理方法的流程示意图;
图4为本申请实例提供的一种媒体信息处理方法的流程示意图;
图5为本申请实例提供的一种媒体信息处理方法的流程示意图;
图6为本申请实例提供的一种媒体信息处理方法的流程示意图;
图6A为本申请实例提供的媒体信息处理中配音请求触发界面的示意图;
图6B为本申请实例提供的媒体信息处理中配音请求触发界面的示意图;
图6C为本申请实例提供的在媒体信息处理中角色选择界面的示意图;
图6D为本申请实例提供的媒体信息处理中在手机上采集第二音频信息的界面 的示意图;
图6E为本申请实例提供的在媒体信息处理过程中采集结束界面的示意图;
图6F为本申请实例提供的在媒体信息处理过程中播放角色的音频信息的界面的示意图;
图7为本申请实例提供的一种媒体信息处理方法的流程示意图;
图8为本申请实例提供的一种媒体信息处理方法的流程示意图;
图8A为本申请实施提供的视频播放界面的示意图;
图8B为本申请实例提供的视频墙的示意图;
图8C为本申请实例提供的媒体信息处理中的视频切换界面示意图;
图9为本申请实例提供的一种媒体信息处理服务器的结构示意图;
图10为本申请实例提供的一种媒体信息处理客户端的结构示意图;
图11为本申请实例提供的一种媒体信息处理客户端的结构示意图;
图12为本申请实例中媒体信息处理服务器的结构示意图;
图13为本申请实例中媒体信息处理客户端的结构示意图;
图14为本申请实例中媒体信息处理客户端的结构示意图。
具体实施方式
下面将结合本申请实例中的附图,对本申请实例中的技术方案进行清楚、完整地描述,显然,所描述的实例仅是本申请一部分实例,而不是全部的实例。基于本申请中的实例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实例,都属于本申请保护的范围。
图1为本申请各实例涉及的一种实施环境的结构示意图。如图1所示,该实施环境包括:播放设备110、智能终端120和配音服务器130。
该播放设备110可以为电视等。该智能终端120可以为手机、平板电脑等。
该播放设备110可以通过有线或无线网络与该配音服务器130进行通信,从该配音服务器130获取视频文件,并进行播放,并在播放视频的过程中在播放界面显示配音触发信息,例如二维码。
该智能终端120可以触发该配音触发信息,例如扫描该二维码,并与该配音服务器130进行通信完成后续的配音操作。
该配音服务器130可以是一台服务器,或者由若干台服务器组成的服务器集群,或者是一个云计算服务中心。该配音服务器130可以用于存储视频文件,或与数据库进行通信,从数据库获取视频文件。
图2为本申请各实例涉及的一种实施环境的结构示意图。如图2所示,该实施环境包括:终端设备210和配音服务器220。
该终端设备210可以为PC、笔记本电脑、手机或平板电脑等。
该配音服务器220可以是一台服务器,或者由若干台服务器组成的服务器集群,或者是一个云计算服务中心。该配音服务器220可以用于存储视频文件,或与数据库进行通信,从数据库获取视频文件。
该终端设备210可以通过有线或无线网络与该配音服务器220进行通信,从该配音服务器220获取视频文件并进行播放,并在播放的过程中在播放界面上显示配音触发信息,例如配音触发按钮,响应于用户对该配音触发按钮的触发操作,与配音服务器进行通信完成后续的配音操作。
图3为本申请实例提供的一种媒体信息处理方法的流程示意图。如图3所示,该方法包括如下步骤。
步骤301,向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的过程中显示配音触发信息。
步骤302,接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求。
步骤303,根据所述第一配音请求中携带的所述第一视频的标识符(Identifier,ID),获取第二视频。
步骤304,将所述第二视频发送给所述第二客户端,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。
步骤305,从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。
在本申请实例中,向第一客户端发送第一视频和指示允许配音的信息。该第一客户端播放该第一视频,并根据该指示允许配音的信息在播放该第一视频的过程中显示配音触发信息。接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求。根据该第一配音请求中携带的该第一视频的ID,获取第二视频。将所述第二视频发送给所述第二客户端。其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。在本申请实例中,可以在播放视频的过程中,在播放视频的界面上触发配音触发信息,执行配音操作。利用本方案提供的媒体信息处理方法,可以减少客户端和配音服务器之间的交互操作,提高配音效率,节省系统资源。
图4为本申请实例提供的一种媒体信息处理方法的流程示意图。如图4所示,该方法包括如下步骤。
步骤401,第一客户端对第二客户端在播放第一视频的过程中显示的配音触发信息进行触发操作生成携带所述第一视频的ID的配音请求。
步骤402,所述第一客户端将所述配音请求发送给配音服务器。
步骤403,所述第一客户端从所述配音服务器接收第二视频。其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。
步骤404,所述第一客户端播放所述第二视频,并根据播放的所述第二视频为所述角色采集第二音频信息。
步骤405,所述第一客户端将所述第二音频信息发送给所述配音服务器。
利用本实例提供的媒体信息处理方法,可以在播放视频的过程中,在播放视频的界面上触发配音触发信息,执行配音操作,因而可以提高配音效率,节省系统资源。此外,还可以实现在第二客户端上播放视频和配音触发信息,利用第一客户端触发该配音触发信息,利用第一客户端和配音服务器的交互实现配音操作,因此,可以利用两个设备分别来实现视频播放和配音操作,实现任务分布式处理,合理分配系统负载。此外,第一客户端执行的配音操作不影响第二客户端的视频播放,有 利于任务的并行处理。
图5为本申请实例提供的一种媒体信息处理方法的流程示意图。如图5所示,该方法包括如下步骤。
步骤501,从配音服务器接收第一视频和指示允许配音的信息。
步骤502,播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息。
步骤503,响应于对所述配音触发信息的触发操作生成携带所述第一视频的ID的配音请求。
步骤504,将所述配音请求发送给配音服务器。
步骤505,从所述配音服务器接收第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。
步骤506,播放所述第二视频。
步骤507,根据播放的所述第二视频为所述角色采集第二音频信息。
步骤508,将所述第二音频信息发送给所述配音服务器。
在本实例中,该终端设备播放该第一视频,并在播放该第一视频的界面上显示配音触发信息,并与配音服务器进行交互完成配音操作。利用本实例,可以在具备视频播放和配音功能的终端设备上完成媒体信息处理操作。
图6为本申请实例提供的一种媒体信息处理方法的流程示意图。在本实例中,由第一客户端和第二客户端合作完成配音的触发操作。第一客户端和第二客户端可以为同一个终端,也可以是不同的终端,在本实例中,该第一客户端和第二客户端为不同的终端。其中,以第一客户端为电视,第二客户端为手机为例进行描述。在本实例中,以处理的媒体信息为广告视频,广告视频中包含两个角色,由该手机和平板电脑合作完成这两个角色的配音操作为例进行介绍,该方法包括如下步骤。
步骤601,电视从配音服务器获取第一广告视频和指示允许配音的信息。
在本实例中,电视可以为智能电视,也可以为安装在交通工具或电梯间的非智能电视,例如视频播放终端。在本实例中,配音服务器可以保存该第一广告视频,那么,该电视从该配音服务器中获取该第一广告视频。该配音服务器也可以仅为配音控制服务器,不存储该第一广告视频,则该电视通过该配音服务器从存储该第一广告视频的数据库获取该第一广告视频。该第一广告视频包含有广告的画面信息和预先录制好的音频信息,在播放画面的同时播放与该画面对应的音频。
在本实施中,电视可以同步从该配音服务器获取该第一广告视频和指示允许配音的信息;也可以先获取该第一广告视频,再获取该指示允许配音的信息。该指示允许配音的信息用于指示允许为该第一广告视频进行配音。
步骤602,电视播放该第一广告视频,并根据该指示允许配音的信息在播放该第一广告视频的过程中,在播放该第一广告视频的界面上显示配音触发信息。
在本实例中,可以以浮动的形式在播放该第一广告视频的界面上显示该配音触发信息,也可以以嵌入式的方式在播放该第一广告视频的界面上显示该配音触发信息。
图6A为本申请实例提供的媒体信息处理中配音请求触发界面的示意图。在该图6A中,在播放第一广告视频的过程中,在播放第一广告视频的界面的右下角显示配音触发信息,即二维码信息。也可以如图6B所示,图6B为本申请实例提供的媒体信息处理中配音请求触发界面的示意图。在该图6B中,在播放该第一广告视频的过程中,在播放第一广告视频的界面的右下角显示配音触发按钮,当用户用遥 控器触发该配音触发按钮时,该电视响应于对该配音触发按钮的触发操作,生成二维码显示请求,查找与该二维码显示请求对应的二维码,并在该配音请求触发界面上显示如图6A所示的二维码。
步骤603,手机运行二维码扫描功能,对该配音触发信息进行触发操作,即对该二维码进行扫描,生成携带该第一广告视频的ID的第一配音请求。
在本实例中,手机利用自身的二维码扫描功能对图6A所示的二维码进行扫描,生成第一配音请求。该第一配音请求携带该第一广告视频的ID。
步骤604,该手机将该第一配音请求发送给该配音服务器。
在本实例中,手机利用自身的二维码扫描功能扫描图6A所示的二维码,从该二维码中获取该配音服务器的IP地址,根据该配音服务器的IP地址将该第一配音请求发送给该配音服务器。
步骤605,该配音服务器根据该第一配音请求中携带的该第一广告视频的ID查找到第二广告视频。
在本实例中,该第二广告视频是去除了该第一广告视频中的角色的第一音频信息而得到的。该配音服务器可以根据该第一广告视频的ID查找到该第一广告视频,去除该第一广告视频中角色的第一音频信息,得到该第二广告视频;也可以是该配音服务器根据该第一广告视频的ID和预存的该第一广告视频的ID和第二广告视频之间的对应关系查找到该第二广告视频。
步骤606,该配音服务器将配音界面、该第二广告视频的描述信息和指示允许播放该第二广告视频的信息发送给该手机。
在本实例中,该配音服务器可以向该手机发送网页形式的配音界面。该配音服务器根据该第一配音请求中携带的该第一广告视频的ID查找到该第二广告视频的描述信息,并发送给该手机。
步骤607,该手机显示该配音界面,并根据该第二广告视频的描述信息和指示允许播放该第二广告视频的信息在该配音界面上显示该第二广告视频的播放触发按钮。
步骤608,该配音服务器判断该第一广告视频中包含一个角色还是多个角色,当该第一广告视频中包含多个角色,例如第一角色和第二角色时,执行步骤609。
在本实例中,当该第一广告视频中仅包含一个角色,且用户点击第二广告视频的播放触发按钮时,该配音服务器向该手机发送该第二广告视频和台词信息,指示该手机根据播放的该第二广告视频和台词信息为该角色采集音频信息。该手机将为该角色采集的音频信息发送给该配音服务器,该配音服务器对该第二广告视频和该音频信息进行合成处理,得到合成后的第三广告视频
步骤609,该配音服务器向该手机发送指示允许选择单独配音的信息和指示允许选择合作配音的信息。
步骤610,该手机根据该指示允许选择单独配音的信息和指示允许选择合作配音的信息在该配音界面上显示单独配音触发按钮和合作配音触发按钮。
在本步骤中,也可以只显示合作配音触发按钮。当用户不触发该合作配音触发按钮时,该手机不向该配音服务器发送请求合作配音的请求,则默认用户选择了单独配音方式
步骤611,该手机响应于用户对于该合作配音触发按钮的触发操作,生成第二配音请求,并将该第二配音请求发送给该配音服务器。
在本实例中,若用户触发了单独配音触发按钮,则该配音服务器将该第二广告 视频发送给该手机,并指示该手机为所有角色采集音频信息。
步骤612,该配音服务器选择第三客户端作为合作配音客户端。
在本实施中,该第三客户端为平板电脑。在本实例中,也可以在该手机完成角色选择和上传该角色的音频信息之后,该配音服务器选择该第三客户端作为合作配音客户端。
步骤613,该配音服务器根据该第二配音请求从该第一广告视频中获取第一角色的信息和第二角色的信息,并发送给该手机和平板电脑。
在本申请一实例中,也可以先获取第一角色的信息和第二角色的信息,再选择第三客户端作为合作配音客户端。
步骤614,该手机和平板电脑根据该第一角色的信息和该第二角色的信息在该配音界面上显示第一角色选择按钮和第二角色选择按钮。
在本实例中,该手机还可以在该配音界面上显示随机分配角色触发按钮。如图6C所示,图6C为本申请实例提供的在媒体信息处理中角色选择界面的示意图。在该界面中显示了第一角色选择按钮、第二角色选择按钮和随机分配角色触发按钮。当用户触发该随机分配角色触发按钮时,该手机生成携带指示随机分配角色信息的配音请求,该配音服务器根据该指示随机分配角色信息的配音请求随机为手机分配角色。
步骤609~步骤614完成了手机选择第一角色,平板电脑选择第二角色的操作。
步骤615,该配音服务器为该手机分配第一角色。
在本实例中,该手机响应于用户对第一角色选择按钮的触发操作生成携带该第一角色的信息的第三配音请求,将该第三配音请求发送给配音服务器,该配音服务器为该手机分配该第一角色。
步骤616,该配音服务器为该平板电脑分配第二角色。
该平板电脑响应于用户对第二角色选择按钮的触发操作生成携带该第二角色的信息的第四配音请求,将该第四配音请求发送给该配音服务器,该配音服务器为该平板电脑分配该第二角色。
在本实例中,该第三配音请求和第四配音请求通常不是同时发送的。在本申请一实例中,该手机先发送携带该第一角色的信息的第三配音请求,该配音服务器为该手机分配该第一角色,并指示该手机和该平板电脑将该第一角色选择按钮置为不可选状态。在这种情况下,该平板电脑的用户只能选择触发该第二角色选择按钮,生成携带该第二角色的信息的第四配音请求。
步骤617,该手机响应于用户对该第二广告视频的播放触发按钮的触发操作,生成第二广告视频播放请求,并发送给该配音服务器。
步骤618,该配音服务器按照时间顺序向所述手机发送第二广告视频的内容,在读取到台词信息时,判断该台词信息属于第一角色还是第二角色,当判定该台词信息属于第一角色时,执行步骤619;否则,执行步骤624。
步骤619,该配音服务器向该手机发送第一音频信息采集指令和该台词信息。
步骤620,该手机根据该第一音频信息采集指令调用自身的音频信息采集装置,播放接收到的第二广告视频和台词信息,根据播放的台词信息为所述第一角色采集第二音频信息。
如图6D所示,图6D为本申请实例提供的媒体信息处理中在手机上采集第二音频信息的界面的示意图。如图6D所示,在该配音界面的上方播放该第二广告视频,在播放该第二广告视频的界面上显示该第一角色的台词信息。该配音界面的下方显 示第二音频信息采集的进度。同时,该第二音频信息采集的进度和该台词信息的播放进度一致。此外,还可以在该配音界面上显示在此次录音之前录制的第一角色的第二音频信息的播放触发按钮和第二角色第三音频信息的播放触发按钮。例如,当手机用户触发之前录制的第一角色的第二音频信息时,该手机播放该第二音频信息,并在该第一角色的第二音频信息的播放触发按钮旁显示动态的播放图标。
步骤621,当该配音服务器判定该台词信息发送完毕时,向该手机发送第一关闭指令。
步骤622,该手机根据该第一关闭指令关闭自身的音频信息采集装置,并将该第二音频信息发送给该配音服务器。
在采集第二音频的过程中,重复执行步骤618~622,直到播放完该第一角色的所有台词信息,并根据该第一角色的台词信息采集相应的第二音频信息并发送给该配音服务器。如图6E所示,图6E为本申请实例提供的在媒体信息处理过程中采集结束界面的示意图。当该第一角色的台词信息播放完毕后,该音频服务器向该手机发送采集结束界面,该手机显示该采集结束界面,在该界面上显示完成信息,以告知用户配音过程结束。
步骤623,该配音服务器对该第二视频和该第一角色的第二音频信息进行合成处理,得到第三广告视频。
步骤624,该配音服务器向该平板电脑发送第二音频信息采集指令和该台词信息。
步骤625,该平板电脑根据该第二音频信息采集指令调用自身的音频信息采集装置,播放接收到的第二广告视频和台词信息,根据播放的台词信息为所述第二角色采集第三音频信息。
步骤626,当该配音服务器判定该台词信息发送完毕时,向该平板电脑发送第二关闭指令。
步骤627,该平板电脑根据该第二关闭指令关闭自身的音频信息采集装置,并将该第三音频信息发送给该配音服务器。
在采集第三音频的过程中,重复执行步骤624~627,直到播放完该第二角色的所有台词信息,并根据该第二角色的台词信息采集相应的第三音频信息发送给配音服务器。
步骤628,该配音服务器对该第二视频和该第二角色的第三音频信息进行合成处理,得到第四广告视频。
步骤629,在配音服务器从手机接收到第一角色的第二视频信息,从平板电脑接收到第二角色的第三视频信息后,对第二广告视频、第二音频信息和第三音频信息进行合成处理得到第五广告视频。
在本实例中,在该手机将该第二音频信息发送给该配音服务器后,该配音服务器可以将该第二音频信息发送给该手机和平板电脑,以使该手机和平板电脑在自身显示的配音界面上播放该第二音频。同样,该平板电脑也可以将该第三音频信息发送给该配音服务器,该配音服务器可以将该第三音频信息发送给该手机和该平板电脑。该手机和平板电脑在自身显示的配音界面上播放该第三音频。图6F为本申请实例提供的在媒体信息处理过程中播放角色的音频信息的界面的示意图。如图6F所示,在该界面中显示了该第一角色的第二音频信息的播放触发按钮和该第二角色的第三音频信息的播放触发按钮。在配音过程中,参与配音的客户端,即该手机或平板电脑可以请求更换配音合作伙伴。例如,该手机在播放该平板电脑采集的第三音 频后,用户认为该第三音频不是想要的音频,则通过手机发起合作伙伴更换请求,例如,如该图6F所示,该用户在配音界面上该第三音频的下方执行右划操作。响应于用户的右划操作,该手机生成该合作伙伴更换请求,并将该合作伙伴更换请求发送给该配音服务器,该配音服务器获取发起了携带该第一广告视频的ID的配音请求的客户端的信息,从中选择一个客户端作为该手机的配音合作伙伴,并将该角色分配给该选择的客户端,并与该选择的客户端进行交互,从该选择的客户端获取该第二角色的第四音频信息发送给该手机显示。
图7为本申请实例提供的一种媒体信息处理方法的流程示意图。在本实例中,第一客户端和第二客户端为同一个终端。在本实例中,以该终端为个人计算机(PC)为例进行介绍。以处理的媒体信息为影视视频,且该第一影视视频中包含一个角色为例进行介绍。该方法包括如下步骤。
步骤701,该PC运行播放器。
在本步骤中,该PC可以运行播放器程序或在网页上运行播放器。
步骤702,该PC从配音服务器获取第一影视视频和指示允许配音的信息。
步骤703,该PC在该播放器中播放该第一影视视频,根据该指示允许配音的信息在播放该第一影视视频的界面上显示配音触发信息。
在步骤中,该配音触发信息可以为配音触发按钮。
步骤704,当用户利用输入设备,例如鼠标点击该配音触发按钮时,该PC生成携带该第一影视视频的ID的配音请求,以及生成配音程序调用指令。
步骤705,该PC利用该配音程序调用指令运行自身安装的配音程序。
步骤706,该配音程序将该配音请求发送给该配音服务器。
步骤707,该配音服务器根据该配音请求中携带的该第一影视视频的ID获取第二影视视频。
在本步骤中,该配音服务器器可以根据该配音请求中携带的第一影视视频的ID获取该第一影视视频,去除该第一影视视频中角色的第一音频信息,得到该第二影视视频。也可以是,该配音服务器根据该第一影视视频的ID和预存的该第一影视视频的ID和第二影视视频之间的对应关系查找到该预先获取到的第二影视视频。
步骤708,该配音服务器将该第二影视视频发送给该配音程序。
步骤709,该配音程序获取该第二影视视频的描述信息,在配音界面中显示该第二影视视频的描述信息以及该第二影视视频的播放触发按钮。
步骤710,当用户点击该第二影视视频的播放触发按钮时,该配音程序在该配音界面中播放该第二影视视频。
步骤711,当该配音程序读取到该第二影视视频中的角色的台词信息时,向PC的音频采集装置发送音频信息采集指令。
步骤712,该PC根据该音频信息采集指令调用自身的音频信息采集装置,采集该角色的第二音频信息。
步骤713,当该配音程序判断该台词信息播放完毕时,向该PC发送关闭指令。
步骤714,该PC根据该关闭指令关闭该音频信息采集装置。
在按照时间顺序播放所述第二影视视频的过程中,每到读到新的一条台词信息,则执行步骤711~步骤714,直到播放完所有的台词信息。
步骤715,当该角色的第二音频信息采集完毕后,该配音程序对该第二影视视频和该第二音频信息执行合成处理,得到第三影视视频。
步骤716,该配音程序将该第三影视视频发送给该配音服务器。
图8为本申请实例提供的一种媒体信息处理方法的流程示意图。在本实例中,在客户端上传合成得到的视频后,该配音服务器对该视频进行后续处理。该方法包括如下步骤。
步骤801,配音服务器获取利用上述方法获取到的合成得到的视频。
步骤802,该配音服务器从该视频中获取音频信息。
步骤803,该配音服务器将该音频信息转换成文本信息。
步骤804,该配音服务器判断该文本信息是否包含和预存的关键词匹配的信息,当判定该文本信息包含和预存的关键词匹配的信息时,执行步骤805;否则,执行步骤808。
在本实施中,该关键词可以为:趣味、上海话,反串等。例如,如果该文本信息包含“谢谢侬”,将“谢谢侬”与数据库中预存的数据进行比较,发现该“谢谢侬”和关键词“上海话”匹配,则判定该文本信息包含和预存的关键词匹配的信息。该文本信息可以和多个预存的关键词相匹配。
步骤805,设置该视频的ID和该预存的关键词之间的第一关联关系。
步骤806,当接收到第一客户端发送的第一视频展示请求,从该第一视频展示请求中获取该视频的ID,根据该视频的ID获取该视频和该预存的关键词,将该视频和该预存的关键词发送给该第一客户端。
步骤807,该第一客户端可以播放该视频,并在播放界面上显示该预存的关键词。
如图8A所示,图8A为本申请实施提供的视频播放界面的示意图。在该图示中,该第一客户端可以在该界面中播放该视频,显示从配音服务器接收的关键词。该第一客户端可以将该视频播放界面通过网页的形式进行分享,例如,通过点击界面上的“发布配音”按钮,将该视频播放界面分享到社交网络。在该图8A所示的界面上还可以显示音频处理按钮,例如“萝莉音”、“电子音”、“猫咪”和“烟嗓音”等。例如,当用户点击了“猫咪”按钮,该第一客户端可以将视频中的音频处理成猫咪音的形式。
当该第一客户端从该配音服务器获取多个视频时,可以以视频墙的形式显示获取到的视频。如图8B所示,图8B为本申请实例提供的视频墙的示意图。在该视频墙上可以显示多个视频,用户可以为其感兴趣的视频点赞。
步骤808,该配音服务器可以创建和该文本信息对应的自定义关键词,设置该视频的ID和该自定义关键词之间的第二关联关系。
步骤809,当接收到第二客户端发送的第二视频展示请求,从该第二视频展示请求中获取该视频的ID,根据该视频的ID获取该视频和自定义关键词,将该视频和该自定义关键词发送给该第二客户端。
步骤810,该第二客户端可以播放该视频,并在播放界面上显示该自定义关键词。
在本步骤中,也可以按照807所述的方法发布该视频和该自定义关键词。
利用本申请实例,该配音服务器保存有原视频,例如上述图3至图7中的第一视频,也可以保存有利用上述实例采集到的音频信息合成得到的合成视频,称为配音版视频。例如,上述图3至图7中的第三视频。在再次向客户端推送视频时,可以向该客户端发送原视频和指示允许视频切换的信息。该客户端播放该原视频,并在播放该原视频的界面上显示视频切换触发信息。当用户触发该视频切换触发信息时,该客户端生成视频切换请求,并发送给该配音服务器。该配音服务器根据该视 频切换请求获取该配音版视频,并发送给该客户端进行播放。图8C为本申请实例提供的媒体信息处理中的视频切换界面示意图。在该图8C中,显示“原视频/配音版视频”的切换按钮,利用该切换按钮可以向配音服务器发送切换请求,获取并播放另外一种版本的视频。
图9为本申请实例提供的一种媒体信息处理服务器的结构示意图。在图9所示的实例中,该服务器包括:第一视频发送模块901,第一配音请求接收模块902,第一视频获取模块903,第二视频发送模块904和第二视频获取模块905。
该第一视频发送模块901,用于向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息。
该第一配音请求接收模块902,用于接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求。
该第一视频获取模块903,用于根据所述第一配音请求中携带的所述第一视频的ID,获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。
该第二视频发送模块904,用于将所述第二视频发送给所述第二客户端。
该第二视频获取模块905,用于从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,用于对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频;或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。
在本申请一实例中,该服务器进一步包括:角色判断模块906,配音界面发送模块907,第二配音请求接收模块908,合作配音客户端选择模块909,角色信息发送模块910,角色分配模块911和第三视频获取模块912。
该角色判断模块906,用于判断所述角色是否包括多个角色。
该配音界面发送模块907,用于当判定所述角色包含多个角色时,向所述第二客户端发送配音界面和指示允许选择单独配音的信息和指示允许选择合作配音的信息,以使所述第二客户端根据所述指示允许选择单独配音的信息和指示允许选择合作配音的信息在所述配音界面上显示单独配音触发按钮和合作配音触发按钮。
该第二配音请求接收模块908,并用于从所述第二客户端接收第二配音请求,所述第二配音请求是所述第二客户端响应于用户对所述合作配音触发按钮的触发操作而生成的。
该合作配音客户端选择模块909,用于根据所述第二配音请求获取第一角色的信息和第二角色的信息,选择第三客户端作为合作配音客户端。
该角色信息发送模块910,用于分别将所述第一角色的信息和所述第二角色的信息发送给所述第二客户端和所述第三客户端,以使所述第二客户端和第三客户端根据所述第一角色的信息和第二角色的信息展示第一角色选择按钮和第二角色选择按钮。
该角色分配模块911,用于响应于所述第二客户端发送的携带所述第一角色的信息的第三配音请求和所述第三客户端发送的携带所述第二角色的信息的第四配音请求,将所述第一角色和第二角色分别分配给所述第二客户端和所述第三客户端;其中,所述第三配音请求和所述第四配音请求分别是所述第二客户端和所述第三客户端响应于用户对所述第一角色选择按钮和第二角色选择按钮的触发操作而生成的。
该第二视频获取模块905,进一步用于从所述第二客户端接收根据播放的所述第二视频而采集到的所述第一角色的第二音频信息。
该第三视频获取模块912,用于从所述第三客户端接收根据播放的所述第二视频而采集到的所述第二角色的第三音频信息,对所述第二视频和所述第二角色的第三音频信息执行合成处理得到第四视频。
在本申请一实例中,该服务器进一步包括:转换模块913,关键词判断模块914和处理模块915。
该转换模块913,用于将所述角色的第二音频信息转换成文本信息。
该关键词判断模块914,用于判断所述文本信息是否包含和预存的关键词匹配的信息。
该处理模块915,用于当判定所述文本信息包含和所述预存的关键词匹配的信息,设置所述第三视频的ID和所述预存的关键词之间的第一关联关系,从第四客户端接收携带所述第三视频的ID的第一视频展示请求,根据所述第三视频的ID和所述第一关联关系查找到所述预存的关键词和所述第三视频,将所述第三视频和所述预存的关键词发送给所述第四客户端进行展示,当判定所述文本信息不包含和所述预存的关键词匹配的信息,设置和所述文本信息对应的自定义关键词,设置所述第三视频的ID和所述自定义关键词之间的第二关联关系,从第五客户端接收携带所述第三视频的ID的第二视频展示请求,根据所述第三视频的ID和所述第二关联关系查找到所述自定义关键词和所述第三视频,将所述第三视频和所述自定义关键词发送给所述第五客户端进行展示。
在本申请一实例中,该服务器进一步包括:切换指示信息发送模块916、视频切换请求接收模块917、第四视频获取模块918、第三视频发送模块919。
该换指示信息发送模块916,用于向所述第一客户端发送所述第一视频,并向所述第一客户端发送指示允许视频切换的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许视频切换的信息在播放所述第一视频的所述界面上显示视频切换触发信息。
该视频切换请求接收模块917,用于从所述第一客户端接收携带所述第三视频的ID的视频切换请求,其中,所述视频切换请求是所述第一客户端响应于用户对于所述视频切换触发信息的触发操作而生成的。
该第四视频获取模块918,用于根据所述第三视频的ID获取所述第三视频。
该第三视频发送模块919,用于向所述第一客户端发送所述第三视频,以使所述第一客户端播放所述第三视频。
在本申请一实例中,该第一客户端和第二客户端可以为同一客户端或不同客户端。
图10为本申请实例提供的一种媒体信息处理客户端的结构示意图。在图10所示的实例中,该客户端包括:配音请求生成模块1001,配音请求发送模块1002,视频接收模块1003,视频播放模块1004,音频采集模块1005和音频发送模块1006。
配音请求生成模块1001,用于对另一客户端在播放第一视频的界面上显示的配音触发信息进行触发操作生成携带所述第一视频的标识符ID的配音请求;
配音请求发送模块1002,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
视频接收模块1003,用于从所述配音服务器接收第二视频;
视频播放模块1004,用于播放所述第二视频;
音频采集模块1005,用于并根据播放的所述第二视频为所述角色采集第二音频信息;
音频发送模块1006,用于将所述第二音频信息发送给所述配音服务器。
在本申请一实例中,该客户端进一步包括:第一信息接收模块1007,第一显示模块1008,视频播放请求生成模块1009,视频播放请求发送模块1010。
该第一信息接收模块1007,用于从所述配音服务器接收所述配音服务器根据所述配音请求向所述客户端发送的网页形式的配音界面,所述第二视频的描述信息以及指示允许播放所述第二视频的信息。
该第一显示模块1008,用于显示所述配音界面,根据所述第一视频的描述信息以及所述指示允许播放所述第二视频的信息显示所述第二视频的播放触发按钮。
该视频播放请求生成模块1009,用于响应于用户对所述第二视频的播放触发按钮的触发操作生成第二视频播放请求。
该视频播放请求发送模块1010,用于将所述第二视频播放请求发送给所述配音服务器,以使所述配音服务器根据所述第二视频播放请求向所述客户端发送所述第二视频。
在本申请一实例中,该客户端进一步包括:配音程序运行模块1011。
该配音程序运行模块1011,用于根据所述配音请求运行自身安装的配音程序。
该视频接收模块1003,进一步用于在所述配音程序的配音界面中接收所述配音服务器根据所述配音请求向所述客户端推送的所述第二视频。
在本申请一实例中,该客户端进一步包括:第一音频信息采集指令接收模块1012,第一关闭指令接收模块1013和第一关闭处理模块1014。
该第一音频信息采集指令接收模块1012,用于当所述配音服务器按照时间顺序向所述客户端发送所述第二视频和所述角色的台词信息时,接收所述配音服务器发送的第一音频信息采集指令。
该视频播放模块1004,进一步用于播放所述第二视频和所述角色的所述台词信息。
该音频采集模块1005,进一步用于根据所述第一音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息。
该第一关闭指令接收模块1013,用于接收第一关闭指令,所述第一关闭指令是当所述角色的所述台词信息发送完毕时,所述配音服务器向所述客户端发送的。
该第一关闭处理模块1014,用于根据所述第一关闭指令关闭所述音频信息采集装置。
在本申请一实例中,该客户端进一步包括:第二音频信息采集指令接收模块1015,第二关闭指令接收模块1016和第二关闭处理模块1017。
该第二音频信息采集指令接收模块1015,用于当按照时间顺序在所述配音界面上播放所述第二视频并播放到所述角色的台词信息时,从所述配音程序接收第二音频信息采集指令。
该音频采集模块1005,进一步用于根据所述第二音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息。
该第二关闭指令接收模块1016,用于当所述角色的台词信息播放完毕时,从所述配音程序接收第二关闭指令。
该第二关闭处理模块1017,用于根据所述第二关闭指令关闭所述音频信息采集 装置。
图11为本申请实例提供的一种媒体信息处理客户端的结构示意图。在图11所示的实例中,该客户端包括:视频接收模块1101,视频播放模块1102,配音请求生成模块1103,配音请求发送模块1104,音频采集模块1105和音频信息发送模块1106。
该视频接收模块1101,用于从配音服务器接收第一视频和指示允许配音的信息。
该视频播放模块1102,用于播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息。
该配音请求生成模块1103,用于响应于对所述配音触发信息的触发操作生成携带所述第一视频的ID的配音请求。
该配音请求发送模块1104,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的。
该视频接收模块1101,进一步用于从所述配音服务器接收第二视频。
该视频播放模块1102,进一步用于播放所述第二视频。
该音频采集模块1105,用于根据播放的所述第二视频为所述角色采集第二音频信息。
该音频信息发送模块1106,用于将所述第二音频信息发送给所述配音服务器。
图12为本申请实例中媒体信息处理服务器的结构示意图。如图12所示,该设备包括:处理器1201、非易失性计算机可读存储器1202、显示单元1203、网络通信接口1204。这些组件通过总线1205进行通信。
本实例中,存储器1202中存储有多个程序模块,包括操作系统1206、网络通信模块1207和应用程序1208。
处理器1201可以读取存储器1202中的应用程序所包括的各种模块(图中未示出)来执行媒体信息处理各种功能应用以及数据处理。本实例中的处理器1201可以为一个,也可以为多个,其可以为CPU,处理单元/模块,ASIC,逻辑模块或可编程门阵列等。
其中,操作系统1206可以为:Windows操作系统、Android操作系统或苹果iPhone OS操作系统。
应用程序1208可包括:媒体信息处理模块1209。该媒体信息处理模块1209可包括上述图9中各模块所形成的计算机可执行指令集1209-1及对应的元数据及启发式算法1209-2。这些计算机可执行指令集可以由所述处理器1201执行并完成图3、图6至图8所示方法或图9所示媒体信息处理服务器的功能。
在本实例中,网络通信接口1204与网络通信模块1207相配合完成媒体信息处理服务器的各种网络信号的收发。
显示单元1203具有一显示面板,用于完成相关信息的输入及显示。
图13为本申请实例中媒体信息处理客户端的结构示意图。如图13所示,该设备包括:处理器1301、非易失性计算机可读存储器1302、显示单元1303、网络通信接口1304。这些组件通过总线1305进行通信。
本实例中,存储器1302中存储有多个程序模块,包括操作系统1306、网络通信模块1307和应用程序1308。
处理器1301可以读取存储器1302中的应用程序所包括的各种模块(图中未示出)来执行媒体信息处理各种功能应用以及数据处理。本实例中的处理器1301可以为一个,也可以为多个,其可以为CPU,处理单元/模块,ASIC,逻辑模块或可编 程门阵列等。
其中,操作系统1306可以为:Windows操作系统、Android操作系统或苹果iPhone OS操作系统。
应用程序1308可包括:媒体信息处理模块1309。该媒体信息处理模块1309可包括上述图10中各模块所形成的计算机可执行指令集1309-1及对应的元数据及启发式算法1309-2。这些计算机可执行指令集可以由所述处理器1301执行并完成图4、图6至图8所示方法或图10所示媒体信息处理客户端的功能。
在本实例中,网络通信接口1304与网络通信模块1307相配合完成媒体信息处理客户端的各种网络信号的收发。
显示单元1303具有一显示面板,用于完成相关信息的输入及显示。
图14为本申请实例中媒体信息处理客户端的结构示意图。如图14所示,该设备包括:处理器1401、非易失性计算机可读存储器1402、显示单元1403、网络通信接口1404。这些组件通过总线1405进行通信。
本实例中,存储器1402中存储有多个程序模块,包括操作系统1406、网络通信模块1407和应用程序1408。
处理器1401可以读取存储器1402中的应用程序所包括的各种模块(图中未示出)来执行媒体信息处理各种功能应用以及数据处理。本实例中的处理器1401可以为一个,也可以为多个,其可以为CPU,处理单元/模块,ASIC,逻辑模块或可编程门阵列等。
其中,操作系统1406可以为:Windows操作系统、Android操作系统或苹果iPhone OS操作系统。
应用程序1408可包括:媒体信息处理模块1409。该媒体信息处理模块1409可包括上述图11中各模块所形成的计算机可执行指令集1409-1及对应的元数据及启发式算法1409-2。这些计算机可执行指令集可以由所述处理器1401执行并完成图5至图8所示方法或图11所示媒体信息处理客户端的功能。
在本实例中,网络通信接口1404与网络通信模块1407相配合完成媒体信息处理客户端的各种网络信号的收发。
显示单元1403具有一显示面板,用于完成相关信息的输入及显示。
另外,在本申请各个实例中的各功能模块可以集成在一个处理单元中,也可以是各个模块单独物理存在,也可以两个或两个以上模块集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。所述各实例的功能模块可以位于一个终端或网络节点,或者也可以分布到多个终端或网络节点上。
另外,本申请的每一个实例可以通过如计算机执行的数据处理程序来实现。显然,数据处理程序构成了本申请。此外,通常存储在一个存储介质中的数据处理程序通过直接将程序读取出存储介质或者通过将程序安装或复制到数据处理设备的存储设备(如硬盘和或内存)中执行。因此,这样的存储介质也构成了本申请。存储介质可以使用任何类型的记录方式,例如纸张存储介质(如纸带等)、磁存储介质(如软盘、硬盘、闪存等)、光存储介质(如CD-ROM等)、磁光存储介质(如MO等)等。
因此本申请还提供了一种存储介质,其中存储有数据处理程序,例如机器可读指令,该数据处理程序用于执行本申请上述方法的任何一种实例。
本领域普通技术人员可以理解实现上述实例的全部或部分步骤可以通过硬件来 完成,也可以通过程序来指令相关的硬件完成,所述的程序可以存储于一种计算机可读存储介质中,上述提到的存储介质可以是只读存储器,磁盘或光盘等。
以上所述仅为本申请的较佳实例而已,并不用以限制本申请,凡在本申请的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本申请保护的范围之内。

Claims (28)

  1. 一种媒体信息处理方法,其特征在于,应用于配音服务器,包括:
    向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
    接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求;
    根据所述第一配音请求中携带的所述第一视频的标识符ID,获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    将所述第二视频发送给所述第二客户端;
    从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频,或从所述第二客户端接收所述第二客户端利用所述第二视频和所述第二音频信息进行合成处理得到的所述第三视频。
  2. 根据权利要求1所述的方法,其特征在于,进一步包括:
    根据所述第一配音请求中携带的所述第一视频的ID查找所述第二视频的描述信息;
    将所述第二视频的描述信息以及指示允许播放所述第二视频的信息发送给所述第二客户端,以使所述第二客户端根据所述第二视频的描述信息以及所述指示允许播放所述第二视频的信息显示所述第二视频的播放触发按钮;
    将所述第二视频发送给所述第二客户端包括:
    从所述第二客户端接收第二视频播放请求,所述第二视频播放请求是所述第二客户端响应于用户对所述第二视频的播放触发按钮的触发操作而生成的;
    将根据所述第一视频的ID获取到的所述第二视频发送给所述第二客户端。
  3. 根据权利要求1所述的方法,其特征在于,所述获取第二视频包括:
    根据所述第一视频的ID查找到所述第一视频,去除所述第一视频中所述角色的所述第一音频信息,得到所述第二视频;或
    根据所述第一视频的ID和预存的所述第一视频的ID和所述第二视频之间的对应关系查找到所述第二视频。
  4. 根据权利要求1所述的方法,其特征在于,该方法进一步包括:
    判断所述角色是否包括多个角色;
    当判定所述角色包括多个角色时,执行如下处理:
    向所述第二客户端发送配音界面和指示允许选择单独配音的信息和指示允许选择合作配音的信息,以使所述第二客户端根据所述指示允许选择单独配音的信息和指示允许选择合作配音的信息在所述配音界面上显示单独配音触发按钮和合作配音触发按钮;
    从所述第二客户端接收第二配音请求,所述第二配音请求是所述第二客户端响应于用户对所述合作配音触发按钮的触发操作而生成的;
    根据所述第二配音请求获取第一角色的信息和第二角色的信息,选择第三客户端作为合作配音客户端;
    将所述第一角色的信息和所述第二角色的信息分别发送给所述第二客户端和所述第三客户端,以使所述第二客户端和第三客户端根据所述第一角色的信息和第二 角色的信息展示第一角色选择按钮和第二角色选择按钮;
    响应于所述第二客户端发送的携带所述第一角色的信息的第三配音请求和所述第三客户端发送的携带所述第二角色的信息的第四配音请求,将所述第一角色和第二角色分别分配给所述第二客户端和所述第三客户端;其中,所述第三配音请求和所述第四配音请求分别是所述第二客户端和所述第三客户端响应于用户对所述第一角色选择按钮和第二角色选择按钮的触发操作而生成的;
    所述从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息包括:从所述第二客户端接收根据播放的所述第二视频而采集到的所述第一角色的第二音频信息;
    该方法进一步包括:从所述第三客户端接收根据播放的所述第二视频而采集到的所述第二角色的第三音频信息,对所述第二视频和所述第二角色的第三音频信息执行合成处理得到第四视频。
  5. 根据权利要求1所述的方法,其特征在于,进一步包括:
    将所述角色的第二音频信息转换成文本信息;
    判断所述文本信息是否包含和预存的关键词匹配的信息;
    如果所述文本信息包含和所述预存的关键词匹配的信息,设置所述第三视频的ID和所述预存的关键词之间的第一关联关系,从第四客户端接收携带所述第三视频的ID的第一视频展示请求,根据所述第三视频的ID和所述第一关联关系查找到所述预存的关键词和所述第三视频,将所述第三视频和所述预存的关键词发送给所述第四客户端进行展示;
    如果所述文本信息不包含和所述预存的关键词匹配的信息,设置和所述文本信息对应的自定义关键词,设置所述第三视频的ID和所述自定义关键词之间的第二关联关系,从第五客户端接收携带所述第三视频的ID的第二视频展示请求,根据所述第三视频的ID和所述第二关联关系查找到所述自定义关键词和所述第三视频,将所述第三视频和所述自定义关键词发送给所述第五客户端进行展示。
  6. 根据权利要求1所述的方法,其特征在于,进一步包括:
    向所述第一客户端发送所述第一视频,并向所述第一客户端发送指示允许视频切换的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许视频切换的信息在播放所述第一视频的所述界面上显示视频切换触发信息;
    从所述第一客户端接收携带所述第三视频的ID的视频切换请求,其中,所述视频切换请求是所述第一客户端响应于用户对于所述视频切换触发信息的触发操作而生成的;
    根据所述第三视频的ID获取所述第三视频,向所述第一客户端发送所述第三视频,以使所述第一客户端播放所述第三视频。
  7. 根据权利要求1至6任一项所述的方法,其特征在于,所述第一客户端和所述第二客户端为同一客户端或不同客户端。
  8. 一种媒体信息处理方法,其特征在于,包括:
    第一客户端对第二客户端在播放第一视频的界面上显示的配音触发信息进行触发操作生成携带所述第一视频的标识符ID的配音请求;
    所述第一客户端将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    所述第一客户端从所述配音服务器接收第二视频;
    所述第一客户端播放所述第二视频,并根据播放的所述第二视频为所述角色采集第二音频信息;
    所述第一客户端将所述第二音频信息发送给所述配音服务器。
  9. 根据权利要求8所述的方法,其特征在于,第一客户端对第二客户端在播放所述第一视频的界面上显示的所述配音触发信息进行触发操作生成携带所述第一视频的ID的所述配音请求包括:
    所述第一客户端利用图像采集装置对所述第二客户端在播放所述第一视频的界面上显示的二维码进行扫描,生成携带所述第一视频的ID的所述配音请求。
  10. 根据权利要求8所述的方法,其特征在于,进一步包括:
    所述第一客户端从所述配音服务器接收所述配音服务器根据所述配音请求向所述第一客户端发送的网页形式的配音界面,所述第二视频的描述信息以及指示允许播放所述第二视频的信息;
    所述第一客户端显示所述配音界面,根据所述第二视频的描述信息以及所述指示允许播放所述第二视频的信息显示所述第二视频的播放触发按钮;
    所述第一客户端响应于用户对所述第二视频的播放触发按钮的触发操作生成第二视频播放请求,并将所述第二视频播放请求发送给所述配音服务器,以使所述配音服务器根据所述第二视频播放请求向所述第一客户端发送所述第二视频。
  11. 根据权利要求8所述的方法,其特征在于,进一步包括:
    所述第一客户端根据所述配音请求运行自身安装的配音程序;
    所述第一客户端从所述配音服务器接收所述第二视频包括:所述第一客户端在所述配音程序的配音界面中接收所述配音服务器根据所述配音请求向所述第一客户端推送的所述第二视频。
  12. 根据权利要求10或11所述的方法,其特征在于,当所述第一视频包含多个角色时,该方法进一步包括:
    所述第一客户端选择第一角色;
    所述第一客户端向所述配音服务器发送携带所述第一角色的信息的角色选择请求,以使所述配音服务器为所述第一客户端分配所述第一角色;
    所述第一客户端根据播放的所述第二视频为所述角色采集第二音频信息包括:所述第一客户端根据播放的所述第二视频为所述第一角色采集所述第二音频信息。
  13. 根据权利要求12所述的方法,其特征在于,所述角色进一步包括:第二角色,该方法进一步包括:
    所述第一客户端从所述配音服务器接收第三客户端发送的所述第二角色的第三音频信息,所述第三客户端是所述配音服务器为所述第二角色选定的;
    所述第一客户端播放所述第二角色的所述第三音频信息;
    所述第一客户端向所述配音服务器发送合作伙伴更换请求,以使所述配音服务器根据所述合作伙伴更换请求为所述第二角色选择第四客户端;
    所述第一客户端从所述配音服务器接收所述第四客户端发送的所述第二角色的第四音频信息;
    所述第一客户端播放所述第二角色的第四音频信息。
  14. 根据权利要求10所述的方法,其特征在于,所述第一客户端根据播放的所述第二视频为所述角色采集第二音频信息包括:
    当所述配音服务器按照时间顺序向所述第一客户端发送所述第二视频和所述角色的台词信息时,所述第一客户端接收所述配音服务器发送的第一音频信息采集指 令;
    所述第一客户端播放所述第二视频和所述角色的所述台词信息;
    所述第一客户端根据所述第一音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息;
    接收第一关闭指令,所述第一关闭指令是当所述角色的所述台词信息发送完毕时,所述配音服务器向所述第一客户端发送的;
    所述第一客户端根据所述第一关闭指令关闭所述音频信息采集装置。
  15. 根据权利要求11所述的方法,其特征在于,所述第一客户端根据播放的所述第二视频为所述角色采集第二音频信息包括:
    当按照时间顺序在所述配音界面上播放所述第二视频并播放到所述角色的台词信息时,所述第一客户端从所述配音程序接收第二音频信息采集指令;
    所述第一客户端根据所述第二音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息;
    当所述角色的台词信息播放完毕时,所述第一客户端从所述配音程序接收第二关闭指令;
    所述第一客户端根据所述第二关闭指令关闭所述音频信息采集装置。
  16. 一种媒体信息处理方法,其特征在于,应用于媒体信息处理客户端,包括:
    从配音服务器接收第一视频和指示允许配音的信息;
    播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
    响应于对所述配音触发信息的触发操作生成携带所述第一视频的标识符ID的配音请求;
    将所述配音请求发送给所述配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    从所述配音服务器接收所述第二视频;
    播放所述第二视频;
    根据播放的所述第二视频为所述角色采集第二音频信息;
    将所述第二音频信息发送给所述配音服务器。
  17. 一种媒体信息处理服务器,其特征在于,包括:处理器和存储器,所述存储器存储:
    第一视频发送模块,用于向第一客户端发送第一视频和指示允许配音的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
    第一配音请求接收模块,用于接收第二客户端根据对所述配音触发信息的触发操作而生成的第一配音请求;
    第一视频获取模块,用于根据所述第一配音请求中携带的所述第一视频的标识符ID,获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    第二视频发送模块,用于将所述第二视频发送给所述第二客户端;
    第二视频获取模块,用于从所述第二客户端接收根据播放的所述第二视频而采集到的所述角色的第二音频信息,用于对所述第二视频和所述角色的第二音频信息执行合成处理得到第三视频;或从所述第二客户端接收所述第二客户端利用所述第 二视频和所述第二音频信息进行合成处理得到的所述第三视频。
  18. 根据权利要求17所述的服务器,其特征在于,所述存储器进一步包括:角色判断模块,配音界面发送模块,第二配音请求接收模块,合作配音客户端选择模块,角色信息发送模块,角色分配模块和第三视频获取模块;
    所述角色判断模块,用于判断所述角色是否包括多个角色;
    所述配音界面发送模块,用于当判定所述角色包含多个角色时,向所述第二客户端发送配音界面和指示允许选择单独配音的信息和指示允许选择合作配音的信息,以使所述第二客户端根据所述指示允许选择单独配音的信息和指示允许选择合作配音的信息在所述配音界面上显示单独配音触发按钮和合作配音触发按钮;
    所述第二配音请求接收模块,并用于从所述第二客户端接收第二配音请求,所述第二配音请求是所述第二客户端响应于用户对所述合作配音触发按钮的触发操作而生成的;
    所述合作配音客户端选择模块,用于根据所述第二配音请求获取第一角色的信息和第二角色的信息,选择第三客户端作为合作配音客户端;
    所述角色信息发送模块,用于分别将所述第一角色的信息和所述第二角色的信息发送给所述第二客户端和所述第三客户端,以使所述第二客户端和第三客户端根据所述第一角色的信息和第二角色的信息展示第一角色选择按钮和第二角色选择按钮;
    所述角色分配模块,用于响应于所述第二客户端发送的携带所述第一角色的信息的第三配音请求和所述第三客户端发送的携带所述第二角色的信息的第四配音请求,将所述第一角色和第二角色分别分配给所述第二客户端和所述第三客户端;其中,所述第三配音请求和所述第四配音请求分别是所述第二客户端和所述第三客户端响应于用户对所述第一角色选择按钮和第二角色选择按钮的触发操作而生成的;
    所述第二视频获取模块,进一步用于从所述第二客户端接收根据播放的所述第二视频而采集到的所述第一角色的第二音频信息;
    所述第三视频获取模块,用于从所述第三客户端接收根据播放的所述第二视频而采集到的所述第二角色的第三音频信息,对所述第二视频和所述第二角色的第三音频信息执行合成处理得到第四视频。
  19. 根据权利要求17所述的服务器,其特征在于,所述存储器进一步包括:转换模块,关键词判断模块,处理模块;
    所述转换模块,用于将所述角色的第二音频信息转换成文本信息;
    所述关键词判断模块,用于判断所述文本信息是否包含和预存的关键词匹配的信息;
    所述处理模块,用于当判定所述文本信息包含和所述预存的关键词匹配的信息,设置所述第三视频的ID和所述预存的关键词之间的第一关联关系,从第四客户端接收携带所述第三视频的ID的第一视频展示请求,根据所述第三视频的ID和所述第一关联关系查找到所述预存的关键词和所述第三视频,将所述第三视频和所述预存的关键词发送给所述第四客户端进行展示,当判定所述文本信息不包含和所述预存的关键词匹配的信息时,设置和所述文本信息对应的自定义关键词,设置所述第三视频的ID和所述自定义关键词之间的第二关联关系,从第五客户端接收携带所述第三视频的ID的第二视频展示请求,根据所述第三视频的ID和所述第二关联关系查找到所述自定义关键词和所述第三视频,将所述第三视频和所述自定义关键词发送给所述第五客户端进行展示。
  20. 根据权利要求17所述的服务器,其特征在于,所述存储器进一步包括:切换指示信息发送模块、视频切换请求接收模块、第四视频获取模块、第三视频发送模块;
    所述换指示信息发送模块,用于向所述第一客户端发送所述第一视频,并向所述第一客户端发送指示允许视频切换的信息,以使所述第一客户端播放所述第一视频,并根据所述指示允许视频切换的信息在播放所述第一视频的所述界面上显示视频切换触发信息;
    所述视频切换请求接收模块,用于从所述第一客户端接收携带所述第三视频的ID的视频切换请求,其中,所述视频切换请求是所述第一客户端响应于用户对于所述视频切换触发信息的触发操作而生成的;
    所述第四视频获取模块,用于根据所述第三视频的ID获取所述第三视频;
    所述第三视频发送模块,用于向所述第一客户端发送所述第三视频,以使所述第一客户端播放所述第三视频。
  21. 根据权利要求17至20任一项所述的服务器,其特征在于,所述第一客户端和所述第二客户端为同一客户端或不同客户端。
  22. 一种媒体信息处理客户端,其特征在于,包括:处理器和存储器,所述存储器存储:
    配音请求生成模块,用于对另一客户端在播放第一视频的界面上显示的配音触发信息进行触发操作生成携带所述第一视频的标识符ID的配音请求;
    配音请求发送模块,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    视频接收模块,用于从所述配音服务器接收第二视频;
    视频播放模块,用于播放所述第二视频;
    音频采集模块,用于并根据播放的所述第二视频为所述角色采集第二音频信息;
    音频发送模块,用于将所述第二音频信息发送给所述配音服务器。
  23. 根据权利要求22所述的客户端,其特征在于,所述存储器进一步包括:
    第一信息接收模块,用于从所述配音服务器接收所述配音服务器根据所述配音请求向所述客户端发送的网页形式的配音界面,所述第二视频的描述信息以及指示允许播放所述第二视频的信息;
    第一显示模块,用于显示所述配音界面,根据所述第一视频的描述信息以及所述指示允许播放所述第二视频的信息显示所述第二视频的播放触发按钮;
    视频播放请求生成模块,用于响应于用户对所述第二视频的播放触发按钮的触发操作生成第二视频播放请求,
    视频播放请求发送模块,用于将所述第二视频播放请求发送给所述配音服务器,以使所述配音服务器根据所述第二视频播放请求向所述客户端发送所述第二视频。
  24. 根据权利要求22所述的客户端,其特征在于,所述存储器进一步包括:
    配音程序运行模块,用于根据所述配音请求运行自身安装的配音程序;
    所述视频接收模块,进一步用于在所述配音程序的配音界面中接收所述配音服务器根据所述配音请求向所述客户端推送的所述第二视频。
  25. 根据权利要求23所述的客户端,其特征在于,所述存储器进一步包括:第一音频信息采集指令接收模块,第一关闭指令接收模块,第一关闭处理模块;
    所述第一音频信息采集指令接收模块,用于当所述配音服务器按照时间顺序向 所述客户端发送所述第二视频和所述角色的台词信息时,接收所述配音服务器发送的第一音频信息采集指令;
    所述视频播放模块,进一步用于播放所述第二视频和所述角色的所述台词信息;
    所述音频采集模块,进一步用于根据所述第一音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息;
    所述第一关闭指令接收模块,用于接收第一关闭指令,所述第一关闭指令是当所述角色的所述台词信息发送完毕时,所述配音服务器向所述客户端发送的;
    所述第一关闭处理模块,用于根据所述第一关闭指令关闭所述音频信息采集装置。
  26. 根据权利要求24所述的客户端,其特征在于,所述存储器进一步包括:第二音频信息采集指令接收模块,第二关闭指令接收模块,第二关闭处理模块;
    所述第二音频信息采集指令接收模块,用于当按照时间顺序在所述配音界面上播放所述第二视频并播放到所述角色的台词信息时,从所述配音程序接收第二音频信息采集指令;
    所述音频采集模块,进一步用于根据所述第二音频信息采集指令调用自身的音频信息采集装置为所述角色采集所述第二音频信息;
    所述第二关闭指令接收模块,用于当所述角色的台词信息播放完毕时,接收从所述配音程序发送的第二关闭指令;
    所述第二关闭处理模块,用于根据所述第二关闭指令关闭所述音频信息采集装置。
  27. 一种媒体信息处理客户端,其特征在于,包括:处理器和存储器,所述存储器存储:
    视频接收模块,用于从配音服务器接收第一视频和指示允许配音的信息;
    视频播放模块,用于播放所述第一视频,并根据所述指示允许配音的信息在播放所述第一视频的界面上显示配音触发信息;
    配音请求生成模块,用于响应于对所述配音触发信息的触发操作生成携带所述第一视频的标识符ID的配音请求;
    配音请求发送模块,用于将所述配音请求发送给配音服务器,以使所述配音服务器根据所述配音请求中携带的所述第一视频的ID获取第二视频,其中,所述第二视频是去除了所述第一视频中的角色的第一音频信息而得到的;
    所述视频接收模块,进一步用于从所述配音服务器接收第二视频;
    所述视频播放模块,进一步用于播放所述第二视频;
    音频采集模块,用于根据播放的所述第二视频为所述角色采集第二音频信息;
    音频信息发送模块,用于将所述第二音频信息发送给所述配音服务器。
  28. 一种计算机可读存储介质,存储有计算机可读指令,被至少一个处理器执行以
    执行权利要求1至16任一项所述的媒体信息处理方法。
PCT/CN2017/109927 2016-11-24 2017-11-08 媒体信息处理方法和装置 WO2018095219A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611050833.4A CN107659850B (zh) 2016-11-24 2016-11-24 媒体信息处理方法和装置
CN201611050833.4 2016-11-24

Publications (1)

Publication Number Publication Date
WO2018095219A1 true WO2018095219A1 (zh) 2018-05-31

Family

ID=61127618

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/109927 WO2018095219A1 (zh) 2016-11-24 2017-11-08 媒体信息处理方法和装置

Country Status (2)

Country Link
CN (1) CN107659850B (zh)
WO (1) WO2018095219A1 (zh)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110392273A (zh) * 2019-07-16 2019-10-29 北京达佳互联信息技术有限公司 音视频处理的方法、装置、电子设备及存储介质
CN111223502A (zh) * 2020-01-11 2020-06-02 青岛希望鸟科技有限公司 一种双录交互控制系统和控制方法
CN112954390A (zh) * 2021-01-26 2021-06-11 北京有竹居网络技术有限公司 视频处理方法、装置、存储介质及设备
CN113873034A (zh) * 2021-09-27 2021-12-31 深圳Tcl新技术有限公司 数据传输方法、装置、电子设备及存储介质
CN115037975A (zh) * 2021-02-24 2022-09-09 花瓣云科技有限公司 一种视频配音的方法、相关设备以及计算机可读存储介质

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109361954B (zh) * 2018-11-02 2021-03-26 腾讯科技(深圳)有限公司 视频资源的录制方法、装置、存储介质及电子装置
CN109391842B (zh) * 2018-11-16 2021-01-26 维沃移动通信有限公司 一种配音方法、移动终端
CN110650366B (zh) * 2019-10-29 2021-09-24 成都超有爱科技有限公司 互动配音方法、装置、电子设备及可读存储介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080092047A1 (en) * 2006-10-12 2008-04-17 Rideo, Inc. Interactive multimedia system and method for audio dubbing of video
CN103763480A (zh) * 2014-01-24 2014-04-30 三星电子(中国)研发中心 获取视频配音的方法及设备
CN105047187A (zh) * 2015-08-03 2015-11-11 广州艾美网络科技有限公司 一种点歌装置的点歌控制方法
CN105709416A (zh) * 2016-03-14 2016-06-29 上海科睿展览展示工程科技有限公司 对多用户操作游戏的个性化配音方法及系统

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101860537B (zh) * 2010-05-25 2015-08-12 中兴通讯股份有限公司 一种媒体播放业务的实现方法及媒体服务器
CN102868862A (zh) * 2012-09-13 2013-01-09 北京富年科技有限公司 应用于移动终端的视频配音方法及该设备
KR101374661B1 (ko) * 2012-09-14 2014-03-17 류승범 컨텐츠 객체 재사용이 가능한 컨텐츠 융합 프레임워크 시스템
CN105227966A (zh) * 2015-09-29 2016-01-06 深圳Tcl新技术有限公司 电视播放控制方法、服务器及电视播放控制系统
CN105611337B (zh) * 2015-12-25 2018-12-04 深圳市九洲电器有限公司 一种广告休闲和推广方法及系统

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080092047A1 (en) * 2006-10-12 2008-04-17 Rideo, Inc. Interactive multimedia system and method for audio dubbing of video
CN103763480A (zh) * 2014-01-24 2014-04-30 三星电子(中国)研发中心 获取视频配音的方法及设备
CN105047187A (zh) * 2015-08-03 2015-11-11 广州艾美网络科技有限公司 一种点歌装置的点歌控制方法
CN105709416A (zh) * 2016-03-14 2016-06-29 上海科睿展览展示工程科技有限公司 对多用户操作游戏的个性化配音方法及系统

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110392273A (zh) * 2019-07-16 2019-10-29 北京达佳互联信息技术有限公司 音视频处理的方法、装置、电子设备及存储介质
CN110392273B (zh) * 2019-07-16 2023-08-08 北京达佳互联信息技术有限公司 音视频处理的方法、装置、电子设备及存储介质
CN111223502A (zh) * 2020-01-11 2020-06-02 青岛希望鸟科技有限公司 一种双录交互控制系统和控制方法
CN112954390A (zh) * 2021-01-26 2021-06-11 北京有竹居网络技术有限公司 视频处理方法、装置、存储介质及设备
CN112954390B (zh) * 2021-01-26 2023-05-09 北京有竹居网络技术有限公司 视频处理方法、装置、存储介质及设备
CN115037975A (zh) * 2021-02-24 2022-09-09 花瓣云科技有限公司 一种视频配音的方法、相关设备以及计算机可读存储介质
CN115037975B (zh) * 2021-02-24 2024-03-01 花瓣云科技有限公司 一种视频配音的方法、相关设备以及计算机可读存储介质
CN113873034A (zh) * 2021-09-27 2021-12-31 深圳Tcl新技术有限公司 数据传输方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN107659850B (zh) 2019-09-17
CN107659850A (zh) 2018-02-02

Similar Documents

Publication Publication Date Title
WO2018095219A1 (zh) 媒体信息处理方法和装置
WO2020187086A1 (zh) 一种视频剪辑方法、装置、设备和存储介质
CN109474843B (zh) 语音操控终端的方法、客户端、服务器
JP6450029B2 (ja) 広告プッシュシステム、装置及び方法
US9621950B2 (en) TV program identification method, apparatus, terminal, server and system
WO2019196628A1 (zh) 推广内容推送方法、装置及存储介质
US8788544B2 (en) Method and apparatus for managing data
US8588824B2 (en) Transferring media context information based on proximity to a mobile device
US20090063645A1 (en) System and method for supporting messaging using a set top box
JP7551773B2 (ja) インタラクション記録生成方法、装置、デバイス及び媒体
EP3023930A1 (en) Screenshot processing device and method for same
CN109271130B (zh) 音频播放方法、介质、装置和计算设备
KR101991188B1 (ko) 홍보 정보의 처리방법, 장치, 디바이스 및 비휘발성 컴퓨터 저장매체
CN111629253A (zh) 视频处理方法及装置、计算机可读存储介质、电子设备
CN105704504A (zh) 视频直播中插入推送信息的方法和装置
WO2015062224A1 (en) Tv program identification method, apparatus, terminal, server and system
KR101123370B1 (ko) 휴대단말용 객체기반 콘텐츠 제공방법 및 장치
CN109597929A (zh) 搜索结果的展示方法、装置、终端及可读介质
WO2024008184A1 (zh) 一种信息展示方法、装置、电子设备、计算机可读介质
CN103763312A (zh) 功能启动方法、装置及客户端
CN102707914B (zh) 信息处理设备及方法
CN117786159A (zh) 文本素材获取方法、装置、设备、介质和程序产品
US9084011B2 (en) Method for advertising based on audio/video content and method for creating an audio/video playback application
US20240126500A1 (en) Device and method for creating a sharable clip of a podcast
CN114900386A (zh) 一种终端设备及数据接力方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17872953

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17872953

Country of ref document: EP

Kind code of ref document: A1