CN110650366B - Interactive dubbing method and device, electronic equipment and readable storage medium - Google Patents

Interactive dubbing method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN110650366B
CN110650366B CN201911041873.6A CN201911041873A CN110650366B CN 110650366 B CN110650366 B CN 110650366B CN 201911041873 A CN201911041873 A CN 201911041873A CN 110650366 B CN110650366 B CN 110650366B
Authority
CN
China
Prior art keywords
dubbing
dubbed
video
terminal
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911041873.6A
Other languages
Chinese (zh)
Other versions
CN110650366A (en
Inventor
尤国懿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Chaoyouai Technology Co ltd
Original Assignee
Chengdu Chaoyouai Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Chaoyouai Technology Co ltd filed Critical Chengdu Chaoyouai Technology Co ltd
Priority to CN201911041873.6A priority Critical patent/CN110650366B/en
Publication of CN110650366A publication Critical patent/CN110650366A/en
Application granted granted Critical
Publication of CN110650366B publication Critical patent/CN110650366B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/27Server based end-user applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Abstract

The application provides an interactive dubbing method, an interactive dubbing device, electronic equipment and a readable storage medium, wherein when a dubbing terminal receives a dubbing request, the dubbing terminal acquires an audio/video to be dubbed, which does not contain audio data; and displaying dubbing trigger information when receiving dubbing indication information sent by the opposite terminal equipment. And the dubbing indication information represents that the second user finishes dubbing the current to-be-dubbed segment in the to-be-dubbed video. When dubbing triggering information is triggered, playing a next to-be-dubbed fragment of a current to-be-dubbed fragment in the to-be-dubbed audio and video, and acquiring dubbing data of a first user; and sending dubbing data to the opposite terminal equipment. Therefore, for the same video, dubbing can be performed through the mutual matching of different dubbing terminals, dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.

Description

Interactive dubbing method and device, electronic equipment and readable storage medium
Technical Field
The present application relates to the field of data processing technologies, and in particular, to an interactive dubbing method and apparatus, an electronic device, and a readable storage medium.
Background
With the development of internet technology, more and more data can be accessed by users. Video is a common data form in life, and is widely applied to daily life by people. At present, a lot of software is available in the market, so that people can directly modify the dubbing of videos, and the videos are more suitable for the needs of users. However, in the existing dubbing mode, a single user dubs, and dubbing interaction among multiple terminal devices cannot be realized, which causes the dubbing process to be rigid and boring.
Disclosure of Invention
An objective of embodiments of the present invention is to provide an interactive dubbing method, an apparatus, an electronic device and a readable storage medium, so as to solve the problem that there is no related art for dubbing interaction among multiple users.
The embodiment of the application provides an interactive dubbing method, which is applied to a first dubbing terminal of a first user and comprises the following steps: when a dubbing request is received, acquiring an audio and video to be dubbed without audio data; when receiving dubbing indication information sent by opposite terminal equipment, displaying dubbing trigger information; the dubbing indication information represents that the dubbing of the current to-be-dubbed fragment in the to-be-dubbed audio and video is finished by a second user; when the dubbing triggering information is triggered, playing a next to-be-dubbed segment of the current to-be-dubbed segment in the to-be-dubbed audio and video, and collecting dubbing data of the first user; and sending the dubbing data to the opposite terminal equipment.
In the implementation process, after receiving dubbing indication information indicating that the dubbing of the current to-be-dubbed segment in the to-be-dubbed video is finished by the second user, the first dubbing terminal can display dubbing trigger information to inform that the first user can dub the next to-be-dubbed segment. And after the first user triggers the dubbing triggering information, playing the next dubbing fragment to be dubbed of the current dubbing fragment in the audio and video to be dubbed, and collecting dubbing data. Therefore, for the same video, dubbing can be performed through the mutual matching of different dubbing terminals, dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.
In addition, the scheme of the embodiment of the application can be applied to the spoken language learning process, namely, foreign language videos are provided for a user to dub. At the moment, the learning interaction among multiple users can be realized, and the interest of spoken language learning is improved, so that the learning effect to the users is improved to a certain extent.
Further, before the obtaining of the audio and video to be dubbed without dubbing data, the method further comprises: receiving a role selection instruction; determining a to-be-dubbed fragment corresponding to the first dubbing terminal according to the role selection instruction; and the next segment to be dubbed is one of the segments to be dubbed corresponding to the first dubbing terminal.
In the implementation process, the user can select the role in the video by issuing the role selection instruction to the first dubbing terminal. And the first dubbing terminal can determine which of the fragments which need dubbing and correspond to the first dubbing terminal from the audio and video to be dubbed based on the role selection instruction, so that the management of the fragments to be dubbed is facilitated.
Further, the dubbing indication information sent by the peer device includes: in the audio and video to be dubbed, the audio data of the current clip to be dubbed; when receiving dubbing indication information sent by the opposite terminal device, the displaying dubbing trigger information comprises: and playing the current clip to be dubbed and the audio data of the current clip to be dubbed, and displaying the dubbing triggering information after the playing is finished.
In the implementation process, the dubbing indication information may include audio data of a current dubbing clip to be collected by the second dubbing terminal, and before dubbing trigger information is displayed, the audio data of the current dubbing clip to be collected and the audio data of the current dubbing clip to be collected are played first, so that the first user can know the dubbing effect of the second user, and meanwhile, when the first user dubs the next dubbing clip to be collected, the first user can be consistent with the previous video content, and therefore the dubbing experience of the user is improved.
Further, before receiving the dubbing indication information sent by the peer device, the method further includes: acquiring original audio data corresponding to the audio and video to be matched; after the collecting the dubbing data of the first user, before sending the dubbing data to the peer device, the method further includes: comparing the collected dubbing data with the original audio data, determining the score of the dubbing data according to the comparison result, and displaying the score.
In practical application, the scheme of the embodiment of the application can be used for spoken language learning. In the implementation process, dubbing data of the user can be compared with original audio data, and scoring is given, so that the user can know the spoken language level of the dubbing of the user, corresponding adjustment is carried out, and the effect of improving the spoken language level of the user is achieved.
The embodiment of the present application further provides an interactive dubbing method, which is applied to a dubbing server, and includes: when dubbing data of a current dubbing fragment in an audio and video to be dubbed sent by a first dubbing terminal is received and a next dubbing fragment exists in the audio and video to be dubbed, determining a second dubbing terminal corresponding to the next dubbing fragment; the first dubbing terminal and the second dubbing terminal are dubbing terminals in a dubbing task; the dubbing task is a task for dubbing the audio and video to be dubbed; and sending dubbing indication information to the second dubbing terminal so that the second dubbing terminal collects dubbing data of the next to-be-dubbed clip.
In the implementation process, when the dubbing server receives dubbing data of a current to-be-dubbed segment in the to-be-dubbed audio and video sent by the first dubbing terminal and a next to-be-dubbed segment exists in the to-be-dubbed audio and video, the second dubbing terminal corresponding to the next to-be-dubbed segment can be determined, and dubbing indication information is sent to the second dubbing terminal, so that the second dubbing terminal can acquire the dubbing data of the next to-be-dubbed segment. Therefore, through continuous transfer coordination of the dubbing server, the audio and video to be dubbed can be successfully dubbed cooperatively completed by the plurality of dubbing terminals, dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.
In addition, the scheme of the embodiment of the application can be applied to the spoken language learning process, namely, foreign language videos are provided for a user to dub. At the moment, the learning interaction among multiple users can be realized, and the interest of spoken language learning is improved, so that the learning effect to the users is improved to a certain extent.
Further, when dubbing data of a current to-be-dubbed clip in the to-be-dubbed audio and video sent by the first dubbing terminal is received, the method further includes: and sending the dubbing data of the current to-be-dubbed segment to all the dubbing terminals except the first dubbing terminal in the dubbing task.
In the implementation process, the dubbing data of the current dubbing fragment to be dubbed is sent to all dubbing terminals except the first dubbing terminal in the dubbing task, so that all the dubbing terminals in the dubbing task can acquire the dubbing data of the first dubbing terminal to the current dubbing fragment to be dubbed, users of all the dubbing terminals participating in the dubbing task can know the dubbing effect of the first user, the interactivity and the interestingness of the scheme are improved, and the dubbing effect is improved to a certain extent.
Further, when dubbing data of a current to-be-dubbed segment in the audio/video to be dubbed, which is sent by the first dubbing terminal, is received and a next to-be-dubbed segment does not exist in the audio/video to be dubbed, the interactive dubbing method further includes: informing all dubbing terminals in the dubbing task of completing dubbing so that all dubbing terminals in the dubbing task respectively synthesize the audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video; or, a third dubbing terminal in the dubbing task is informed of completing dubbing, so that the third dubbing terminal synthesizes the audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video; and receiving the target dubbing video reported by the third dubbing terminal, and sending the target dubbing video to all dubbing terminals except the third dubbing terminal in the dubbing task.
In the implementation process, the audio and video to be dubbed can be synthesized with dubbing data acquired by each dubbing terminal, so that the target dubbing video is obtained. Therefore, the dubbing level of each dubbing participant can be visually reflected through the target dubbing video, and meanwhile, the user can review and share the dubbing level, so that the interestingness of the scheme is further improved.
Further, when dubbing data of a current to-be-dubbed segment in the audio/video to be dubbed, which is sent by the first dubbing terminal, is received and a next to-be-dubbed segment does not exist in the audio/video to be dubbed, the interactive dubbing method further includes: synthesizing the audio and video to be dubbed and the dubbing data to obtain a target dubbing video; and sending the target dubbing video to all dubbing terminals in the dubbing task.
In the implementation process, the dubbing server can synthesize the audio and video to be dubbed with the dubbing data collected by each dubbing terminal, so that a target dubbing video is obtained and sent to all dubbing terminals in the dubbing task. Therefore, the dubbing level of each dubbing participant can be visually reflected through the target dubbing video, and meanwhile, the user can review and share the dubbing level, so that the interestingness of the scheme is further improved.
Further, before the audio and video to be dubbed and the dubbing data are synthesized, the method further includes: and informing all dubbing terminals in the dubbing task of completing dubbing, and determining to receive a video synthesis request sent by any dubbing terminal in the dubbing task.
In the implementation process, after dubbing is performed on all the to-be-dubbed segments, all the dubbing terminals can be informed of completing dubbing, and after a video synthesis request sent by any dubbing terminal in the dubbing task is received, a video synthesis request is performed to obtain a target dubbing video, so that the manageability of the scheme of the application is improved.
Further, when dubbing data of a current to-be-dubbed clip in the to-be-dubbed audio and video sent by the first dubbing terminal is received, the method further includes: and comparing dubbing data of the current to-be-dubbed clip with pre-stored original audio data of the to-be-dubbed audio and video, determining the score of the dubbing data according to the comparison result, and sending the score to the first dubbing terminal.
In practical application, the scheme of the embodiment of the application can be used for spoken language learning. In the implementation process, dubbing data of the first dubbing terminal can be compared with original audio data, and scoring is returned, so that a user can know the spoken language level of the dubbing, corresponding adjustment is performed, and the effect of improving the spoken language level is achieved.
Further, before determining a second dubbing terminal corresponding to a next dubbing fragment of the current dubbing fragment, the method further includes: and determining that a score acceptance instruction sent by the first dubbing terminal is received.
In the implementation process, after the user approves the dubbing effect, dubbing of the subsequent segments can be performed, so that the dubbing effect is improved.
The embodiment of the present application further provides an interactive dubbing device, which is applied to a first dubbing terminal of a first user, and includes: the device comprises a first receiving module, an acquisition module, a display module, a playing module, an acquisition processing module and a first sending module; the first receiving module is used for receiving a dubbing request and receiving dubbing indication information sent by opposite-end equipment; the acquisition module is used for acquiring the audio and video to be dubbed without audio data when receiving the dubbing request; the display module is used for displaying dubbing triggering information when dubbing indication information sent by the opposite terminal equipment is received; the dubbing indication information represents that the dubbing of the current to-be-dubbed fragment in the to-be-dubbed audio and video is finished by a second user; the playing module is used for playing a next dubbing fragment to be dubbed in the audio and video to be dubbed when the dubbing triggering information is triggered; the acquisition processing module is used for acquiring dubbing data of the first user; the first sending module is configured to send the dubbing data to the peer device.
In the implementation process, after receiving dubbing indication information representing that the second user has finished dubbing the current dubbing segment in the video to be dubbed, the dubbing trigger information can be displayed to inform the first user that the next dubbing segment can be dubbed. And after the first user triggers the dubbing triggering information, playing the next dubbing fragment to be dubbed of the current dubbing fragment in the audio and video to be dubbed, and collecting dubbing data of the first user. Therefore, for the same video, dubbing can be performed through the mutual matching of different dubbing terminals, dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.
In addition, the scheme of the embodiment of the application can be applied to the spoken language learning process, namely, foreign language videos are provided for a user to dub. At the moment, the learning interaction among multiple users can be realized, and the interest of spoken language learning is improved, so that the learning effect to the users is improved to a certain extent.
The embodiment of the present application further provides an interactive dubbing device, which is applied to a dubbing server, and includes: the device comprises a second receiving module, a second sending module and a processing module; the second receiving module is used for receiving dubbing data of a current to-be-dubbed fragment in the audio and video to be dubbed sent by the first dubbing terminal; the processing module is used for determining a second dubbing terminal corresponding to a next dubbing fragment when dubbing data of a current dubbing fragment in an audio and video to be dubbed sent by a first dubbing terminal are received and the next dubbing fragment exists in the audio and video to be dubbed; the first dubbing terminal and the second dubbing terminal are dubbing terminals in a dubbing task; the dubbing task is a task for dubbing the audio and video to be dubbed; and the second sending module is used for sending dubbing indication information to the second dubbing terminal so that the second dubbing terminal can collect dubbing data of the next to-be-dubbed clip.
In the implementation process, when dubbing data of a current to-be-dubbed segment in the to-be-dubbed audio and video sent by the first dubbing terminal is received and a next to-be-dubbed segment exists in the to-be-dubbed audio and video, a second dubbing terminal corresponding to the next to-be-dubbed segment can be determined, and dubbing indication information is sent to the second dubbing terminal, so that the second dubbing terminal can collect the dubbing data of the next to-be-dubbed segment. Therefore, through continuous transfer coordination of the dubbing server, the audio and video to be dubbed can be successfully dubbed cooperatively completed by the plurality of dubbing terminals, dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.
In addition, the scheme of the embodiment of the application can be applied to the spoken language learning process, namely, foreign language videos are provided for a user to dub. At the moment, the learning interaction among multiple users can be realized, and the interest of spoken language learning is improved, so that the learning effect to the users is improved to a certain extent.
The embodiment of the application also provides electronic equipment, which comprises a processor, a memory and a communication bus; the communication bus is used for realizing connection communication between the processor and the memory; the processor is used for executing one or more first programs stored in the memory so as to realize any one of the interactive dubbing methods applied to the dubbing terminal; or the processor is used for executing one or more second programs stored in the memory to realize any one of the above interactive dubbing methods applied to the dubbing server.
The embodiment of the application also provides a readable storage medium, wherein the readable storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to realize the interactive dubbing method applied to any one of the dubbing terminals; or implementing any of the above-described interactive dubbing methods applied to a dubbing server.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are required to be used in the embodiments of the present application will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present application and therefore should not be considered as limiting the scope, and that those skilled in the art can also obtain other related drawings based on the drawings without inventive efforts.
Fig. 1 is a schematic view of a process interaction of an interactive dubbing with an audio server according to an embodiment of the present application;
fig. 2 is a schematic view of a process interaction of an interactive dubbing without an audio server according to an embodiment of the present application;
fig. 3 is a schematic view of a role selection interface provided in an embodiment of the present application;
fig. 4 is a schematic view of a dubbing trigger information display interface according to an embodiment of the present application;
fig. 5 is a schematic view of a score display interface provided in an embodiment of the present application;
fig. 6 is a schematic structural diagram of an interactive dubbing apparatus applied to a dubbing terminal according to an embodiment of the present application;
fig. 7 is a schematic structural diagram of an interactive dubbing apparatus applied to a dubbing server according to an embodiment of the present application;
fig. 8 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application.
The first embodiment is as follows:
in order to solve the problem that no related technology of dubbing interaction among multiple terminal devices exists at present, an interactive dubbing method is provided in the embodiment of the application. As shown in fig. 1 and fig. 2, fig. 1 is a schematic flow chart of the cooperative implementation of interactive dubbing among the first dubbing terminal, the second dubbing terminal, and the dubbing server; fig. 2 is a schematic flow chart of the cooperative interactive dubbing between the first dubbing terminal and the second dubbing terminal.
S101: and when the first dubbing terminal receives the dubbing request, acquiring the audio and video to be dubbed without audio data.
In this embodiment of the application, when receiving a dubbing request, the first dubbing terminal may invite the relevant dubbing terminal to dub the video to be dubbed together, so as to generate a corresponding dubbing task and notify the dubbing server, so that the dubbing server knows which video the audio/video to be dubbed of the current dubbing task is specifically and which dubbing terminals correspondingly participate in the current dubbing task are. In the embodiment of the application, the related users can be invited to participate in the dubbing task in modes of two-dimensional codes and the like. In the example process of fig. 1 in the embodiment of the present application, the dubbing terminal of the invited relevant user is the second dubbing terminal.
In the embodiment of the present application, the dubbing request may be issued by a user, and the dubbing request includes information such as a selected video identifier used for dubbing this time. In the embodiment of the present application, the video to be dubbed may be a video already existing locally in the first dubbing terminal, but may also be a video existing in the dubbing server. When the audio/video to be dubbed is the video existing locally at the first dubbing terminal, the audio/video to be dubbed can be sent to the second dubbing terminal after being invited to the second dubbing terminal. It is to be noted that the solution of the embodiment of the present application may be implemented by a corresponding application program, and the application program may download a certain video in advance when downloading. For such a situation, in the embodiment of the application, the first dubbing terminal only needs to send the video identifier of the selected audio/video to be dubbed to the second dubbing terminal. When the audio/video to be dubbed is a video existing in the dubbing server, in the embodiment of the application, the first dubbing terminal may send a video acquisition request including the dubbing task to the dubbing server after generating the corresponding dubbing task, where the dubbing task includes information of all dubbing terminals participating in the dubbing task (such as an IP address of the dubbing terminal, a user name corresponding to the dubbing terminal, and the like) and identification information of the requested audio/video to be dubbed. After receiving the video acquisition request, the dubbing server can find out the required audio and video to be dubbed according to the identification information of the audio and video to be dubbed and return the audio and video to all dubbing terminals participating in the dubbing task. It should be noted that the audio and video to be dubbed returned by the dubbing server may be a video from which the audio data has been removed, but may also be a video containing the audio data, and at this time, the dubbing terminal performs audio data removal processing on the video.
In the embodiment of the application, before the audio and video to be dubbed without dubbing data is obtained, a corresponding option interface can be provided first, so that a user can select a dubbing role. As shown in fig. 3, the user may issue a role selection instruction by clicking a corresponding option, and then determine the to-be-dubbed segment corresponding to each dubbing terminal according to the role selection instruction.
In the embodiment of the present application, after determining the to-be-dubbed segments corresponding to the dubbing terminals, the dubbing server needs to be notified, so that the dubbing server can know the to-be-dubbed segments corresponding to the dubbing terminals.
S102: and displaying dubbing triggering information.
S103: and when the dubbing triggering information is triggered, playing the current dubbing clip in the audio and video to be dubbed and collecting dubbing data.
In the embodiment of the application, dubbing trigger information can be displayed, and the dubbing trigger information can be a trigger interface, and after being triggered, the dubbing terminal can dub. For example, referring to fig. 4, the microphone in fig. 4 is dubbing trigger information, and after the user touches the microphone, the dubbing terminal can play a current clip to be dubbed and collect dubbing data.
It is to be noted that, in the embodiment of the present application, except that when the current segment to be dubbed is the first segment to be dubbed in the audio/video to be dubbed, the dubbing terminal directly displays the dubbing trigger information, in other cases, the dubbing trigger information needs to be displayed only when receiving the dubbing indication information sent by the dubbing server.
In the example shown in fig. 1 in the embodiment of the present application, the dubbing indication information is indication information that is issued by the dubbing server after receiving dubbing data reported by a certain dubbing terminal, and may represent that a last dubbing segment of a current dubbing segment is dubbed completely.
In this embodiment of the present application, the dubbing indication information may include audio data of a previous to-be-dubbed segment of a current to-be-dubbed segment in the to-be-dubbed audio/video. Therefore, the audio data of the previous dubbing clip and the audio data of the previous dubbing clip can be played before the dubbing triggering information is displayed, so that a user has better plot substituting feeling when dubbing the current dubbing clip.
S104: and sending the collected dubbing data to a dubbing server.
S105: the dubbing server judges whether the audio and video to be dubbed has a next clip to be dubbed; if yes, go to step S106; otherwise, go to step S111.
S106: and determining a dubbing terminal corresponding to the next dubbing fragment to be dubbed.
In the embodiment of the application, the dubbing server receives the dubbing task sent by the dubbing terminal in advance, and the dubbing task comprises information of all dubbing terminals participating in the dubbing task and identification information of the requested audio and video to be dubbed. Meanwhile, the dubbing server can also obtain the to-be-dubbed segments corresponding to the dubbing terminals. Based on the dubbing server, the dubbing terminal corresponding to the next to-be-dubbed segment can be found out. In the example in fig. 1 in the embodiment of the present application, it is assumed that the dubbing terminal determined in step S106 is the second dubbing terminal. If the dubbing terminal corresponding to the next dubbing segment is still the first dubbing terminal, the dubbing indication information is sent to the first dubbing terminal, and the process goes to step S102.
S107: and sending the dubbing indication information to the second dubbing terminal.
S108: and the second dubbing terminal displays the dubbing triggering information.
S109: and when the dubbing triggering information is triggered, playing the current dubbing clip in the audio and video to be dubbed and collecting dubbing data.
S110: and sending the collected dubbing data to a dubbing server.
In the embodiment of the application, the dubbing terminal and the dubbing server can continuously execute the steps until all the to-be-dubbed segments in the to-be-dubbed audio and video are dubbed completely.
S111: the dubbing server generates a target dubbing video.
In the embodiment of the application, after all the segments to be dubbed in the audio and video to be dubbed are dubbed, the dubbing server can synthesize the audio and video to be dubbed and the dubbing data to obtain the target dubbing video and send the target dubbing video to all the dubbing terminals in the dubbing task.
In a possible implementation manner of the embodiment of the present application, before generating a target dubbing video, all dubbing terminals in the dubbing task may be notified of completion of dubbing, and meanwhile, a corresponding interface for determining whether video synthesis is required or not may be provided in the dubbing terminal, so that when a user determines that video synthesis is required through the interface, the audio terminal sends a video synthesis request to the dubbing server. And the dubbing server regenerates the target dubbing video after receiving the video synthesis request.
It is noted that, in the embodiment of the present application, the synthesis of the target dubbing video may also be implemented by the dubbing terminal. Specifically, in the embodiment of the present application, when receiving dubbing data transmitted from a certain dubbing terminal, the dubbing server may transmit the dubbing data to all the dubbing terminals except the dubbing terminal from which the dubbing data is transmitted in the dubbing task.
At this time, all dubbing terminals in the dubbing task can be notified of completion of dubbing, and after receiving the notification, each dubbing terminal respectively synthesizes the audio and video to be dubbed and the obtained dubbing data to obtain the target dubbing video.
Or, any one of the dubbing terminals (for example, the first dubbing terminal to be notified) in the dubbing task may be notified of completion of dubbing, and after receiving the notification, the dubbing terminal may synthesize the audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video and send the target dubbing video to the dubbing server. After receiving the target dubbing video reported by the dubbing terminal, the dubbing server can send the target dubbing video to all dubbing terminals except the dubbing terminal in the dubbing task.
It is to be noted that, in the embodiment of the present application, in addition to the interactive dubbing through the audio server, the interactive dubbing of the multi-audio terminal without the participation of the audio server may also be implemented through a local area network such as WiFi. Taking the process of implementing interactive dubbing by the first dubbing terminal and the second dubbing terminal shown in fig. 2 as an example:
s201: and when the first dubbing terminal receives the dubbing request, acquiring the audio and video to be dubbed without audio data.
Similar to the description for fig. 1, the first dubbing terminal, upon receiving the dubbing request, may invite the associated dubbing terminals to dub the dubbed video together, thereby generating a corresponding dubbing task. The dubbing task can be notified to all dubbing terminals participating in the dubbing task, so that each dubbing terminal can know which video the audio and video to be dubbed of the dubbing task is and which dubbing terminals correspondingly participating in the dubbing task are.
In the process of interactive dubbing of the multi-audio-frequency terminal without the participation of the audio-frequency server, the audio and video to be dubbed is the local video of at least one dubbed terminal which participates in the dubbing task. The video can be sent to the dubbing terminal participating in the dubbing task through the established local area network. It should be noted that the transmitted video may be a video from which audio data has been removed, but may also be a video containing audio data, and the audio data removal processing is performed on the video by the dubbing terminal on the receiving end.
Similar to the description in fig. 1, in the embodiment of the present application, before the audio/video to be dubbed without dubbing data is obtained, the dubbing terminal may further provide a corresponding option interface to allow the user to select a dubbing role, and then determine a to-be-dubbed segment corresponding to each dubbing terminal according to the role selection instruction. In the embodiment of the application, the segment to be dubbed corresponding to each dubbing terminal is notified to all dubbing terminals in the dubbing task.
In the embodiment of the application, before the corresponding option interface is provided for the user to select the role, the complete audio and video to be matched with the original audio data can be played first for the user to know the video condition, so that the role selection is made better.
S202: and displaying dubbing triggering information.
S203: and when the dubbing triggering information is triggered, playing the current dubbing clip in the audio and video to be dubbed and collecting dubbing data.
The dubbing trigger information may be a trigger interface which is triggered to allow the dubbing terminal to dub. In the embodiment of the application, except that when the current segment to be dubbed is the first segment to be dubbed in the audio/video to be dubbed, the dubbing terminal can directly display the dubbing trigger information, under other conditions, the dubbing trigger information can be displayed only when dubbing indication information sent by other dubbing terminals in the dubbing task is received.
S204: judging whether a next dubbing fragment exists; if yes, go to step S205; otherwise, go to step S214.
S205: and determining a dubbing terminal corresponding to the next dubbing fragment to be dubbed.
S206: when the dubbing terminal corresponding to the next dubbing fragment is the first dubbing terminal, go to step S202.
When the process goes to step S202 and the steps after S202 are executed again, the next segment to be dubbed in step S206 becomes the current segment to be dubbed in step S203 as time goes by.
S207: and when the dubbing terminal corresponding to the next dubbing fragment to be dubbed is the second dubbing terminal, transmitting dubbing indication information to the second dubbing terminal.
In the embodiment of the application, the first dubbing terminal can send the collected dubbing data to all dubbing terminals in the dubbing task, so that each dubbing terminal can play the current to-be-dubbed clip and the audio data of the current to-be-dubbed clip, the audio and video to be dubbed are continuously played along with the dubbing, and users participating in the dubbing have better plot substituting feeling. Meanwhile, users participating in dubbing can better know the dubbing conditions of other participants, so that the interaction sense is enhanced.
S208: and the second dubbing terminal displays the dubbing triggering information.
S209: and when the dubbing triggering information is triggered, playing the current dubbing clip in the audio and video to be dubbed and collecting dubbing data.
S210: and judging whether the next dubbing fragment exists. If yes, go to step S211; otherwise, go to step S215.
S211: and determining a dubbing terminal corresponding to the next dubbing fragment to be dubbed.
S212: if the dubbing terminal corresponding to the next dubbing segment is the second dubbing terminal, go to step S208.
S213: when the dubbing terminal corresponding to the next dubbing segment is the first dubbing terminal, the dubbing indication information is sent to the second dubbing terminal, and the process goes to step S202.
S214: the first dubbing terminal generates a target dubbing video.
S215: and the second dubbing terminal generates a target dubbing video.
It should be noted that, in the embodiment of the present application, when it is determined that there is no next dubbing segment, the target dubbing video may be generated not by the dubbing terminal that performs the determination operation but by notifying the remaining dubbing terminals.
It should be noted that, in the embodiment of the present application, in the process of interactive dubbing for a multi-audio terminal without participation of an audio server, the function of the dubbing server in the above-mentioned scheme of fig. 1 may also be simultaneously implemented by selecting one dubbing terminal, so as to implement interactive dubbing for a multi-audio terminal without participation of an audio server, and the process may refer to the description of fig. 1 above, and the operation performed by the management server is performed by replacing the selected dubbing terminal.
It should be noted that, for convenience of description in the embodiment of the present application, fig. 1 and fig. 2 are both described by taking a process of implementing interactive dubbing through two dubbing terminals as an example, but it should be understood that, in the embodiment of the present application, a greater number of dubbing terminals may also be used to implement interactive dubbing, and the implementation principle of the implementation principle is consistent with the principle shown in fig. 1 and fig. 2, and therefore, the detailed description is not repeated.
In the embodiment of the application, after the dubbing terminal collects the dubbing data, the dubbing data can be scored and displayed to the user before being sent, so that the user can visually confirm the dubbing effect. In the embodiment of the present application, after the score is displayed, a corresponding confirmation interface may be provided in the interface, for example, as shown in fig. 5, the user may select to perform dubbing on the current dubbing segment again through the interface, or may select to receive the current dubbing, so as to send the dubbing data.
It should be noted that, in the scheme of implementing interactive dubbing through the audio server, the dubbing terminal can implement scoring on the dubbing data, and the dubbing server can also implement scoring. Specifically, the dubbing server may score the dubbing data reported by the dubbing terminal and return the scoring result to the dubbing terminal after receiving the dubbing data. The dubbing terminal displays the score, and simultaneously can provide a corresponding confirmation interface in the interface, and a user can select to dub the segment to be dubbed again through the interface and can also select to accept the dubbing. If the user selects to receive the dubbing, the dubbing terminal sends a grading receiving instruction, and the dubbing server can enter the step of judging whether the next dubbing fragment exists.
It should be noted that in the embodiment of the present application, scoring of dubbing data may be performed by comparing the dubbing data of the current to-be-dubbed segment with the original audio data of the current to-be-dubbed segment in the to-be-dubbed audio/video, and determining a score according to a comparison result (e.g., similarity between the speech rate, pronunciation, and the like) and the original audio.
In the embodiment of the application, the mode of synthesizing the audio and video to be dubbed and the dubbing data to obtain the target dubbing video can be realized by the following modes:
the dubbing server or the dubbing terminal can extract the voice channel of the audio and video to be dubbed with the original audio data and decode the voice channel into an original PCM voice file; then calculating the position of each pcm dubbing data acquired by the dubbing terminal in the whole video pcm voice file and replacing the corresponding pcm fragment with the voice pcm data acquired by the dubbing terminal; and (3) the video part to be provided with the audio and video uses the h264 video frame coded by the source file, the voice part uses the replaced pcm file to be re-coded into an aac audio frame, and finally the video frame and the voice frame are synthesized into a new mp4 file.
In summary, the interactive dubbing method provided by the embodiment of the application can be used for dubbing the same video through the mutual cooperation of different dubbing terminals, so that dubbing interaction among a plurality of users is realized, and the interest of the dubbing process is improved.
In addition, the scheme of the embodiment of the application can be applied to the spoken language learning process, namely, foreign language videos are provided for a user to dub. At the moment, the learning interaction among multiple users can be realized, the dubbing of the same video is realized by the mutual matching of the multiple users, the role playing of each user can be realized, the interest of spoken language learning is improved, and the learning effect to the users is improved to a certain extent.
Example two:
based on the first embodiment, the present embodiment takes a two-player role-playing dubbing process applied in an english oral learning scenario as an example, and further illustrates the present application.
In the embodiment of the application, a user can initiate a dubbing task through the dubbing terminal. When the dubbing task is initiated, the initiator can invite another user to participate in the dubbing task in the forms of two-dimensional codes and the like, and the users can select different roles for dubbing.
For example, suppose that user 1 initiates a dubbing task for a piece of video and invites user 2 to participate, and notifies the dubbing server of the dubbing task, in which the dubbing terminals that are participated in and by the video are indicated. After the dubbing task is started, the dubbing terminals of the user 1 and the user 2 play the video, and in the embodiment of the application, the user can be allowed to select a play mode, wherein the play mode comprises English subtitle play, English-Chinese bilingual subtitle play, subtitle-free play and the like. After the video is played, a character option box as shown in fig. 3 is provided for users 1 and 2 to select the character to be dubbed respectively. After the role is selected, the selection result is submitted to the dubbing server, so that the server can determine which of the two dubbing terminals correspond to the to-be-dubbed segments respectively.
Suppose the video comprises two roles AB, each role having 3 sentences, each sentence having corresponding numbers a1, a2, A3; b1, B2, B3; the order in the video is a1, B1, a2, B2, A3, B3. Assuming that the role selected by the user 1 is a and the role selected by the user B is B, when dubbing, the interface shown in fig. 4 is displayed at the dubbing terminal of the user 1, and after the user 1 touches the microphone, the dubbing terminal starts playing the picture corresponding to a1 and collects dubbing data of the user 1. After the collection is completed, the dubbing terminal of the user 1 scores the dubbing data in sentence a1 according to the accuracy, the fluency and the integrity, and displays the scoring result, for example, as shown in fig. 5. If the user 1 is not satisfied with the result, the user 1 may choose to dub again, so as to replay the picture corresponding to a1, and collect dubbing data of the user 1, and enter the interface shown in fig. 5 again. If the user 1 is satisfied with the result, the user can select to receive the dubbing, the dubbing terminal uploads the dubbing data of the A1 sentence to the dubbing server, after the dubbing server receives the A1 sentence of the A character, the state of the A1 sentence of the video A character in the database is updated from unlocked to recorded, the state of the A1 sentence of the video B character is updated from unlocked to unlocked, and the dubbing terminal of the user 2 is informed to dub. And the dubbing terminal of the user 2 collects dubbing data of B1 sentences according to the execution operation of the dubbing terminal of the user 1 on the sentence A1, and selects to receive the dubbing of this time and uploads the dubbing data to the dubbing server in the user 2. The above steps are sequentially performed until the dubbing data corresponding to a1, B1, a2, B2, A3 and B3 are completely acquired.
After the dubbing data corresponding to a1, B1, a2, B2, A3 and B3 are all collected, the dubbing server notifies the dubbing terminal of the user 1, and sends the dubbing data corresponding to a1, B1, a2, B2, A3 and B3 to the dubbing terminal of the user 1, and the user 1 can select to perform video synthesis to obtain target dubbing videos dubbing of which the dubbing is the dubbing of the user 1 and the user 2, and uploads the target dubbing videos to the dubbing server. And the dubbing server sends the target dubbing video to all users participating in the dubbing task for review, sharing and the like.
In addition, in the embodiment of the application, a scheme of non-real-time dubbing is further provided, and a dubbing initiator can select a character and complete dubbing of all sentences of the character. After completing dubbing of each sentence, similar to the scheme, the dubbing terminal can score and display, and the user can choose to dub again or accept the dubbing of this time, so as to dub the next sentence. After all sentences of the selected character are dubbed, the selected character can be shared with another user to dub other characters in the video. After the dubbing data of all the roles are acquired, the dubbing data can be executed according to the operation in the above mode, and the details are not repeated here.
Through the scheme, the user can listen to the mouth shape in time and exercise the sense of speech by synthesizing the complete work by dubbing and the video prototype, and the user can have great sense of achievement and encourages the user to learn continuously.
Example three:
based on the same inventive concept, the embodiment of the present application further provides an interactive dubbing apparatus 100 and an interactive dubbing apparatus 200. Referring to fig. 6 and 7, fig. 6 shows an interactive dubbing apparatus corresponding to the steps executed by the dubbing terminal in the method shown in fig. 1 or 2, and fig. 7 shows an interactive dubbing apparatus corresponding to the steps executed by the dubbing server in the method shown in fig. 1. It should be understood that the specific functions of the apparatus 100, the apparatus 200 and the apparatus 300 can be referred to the above description, and the detailed description is appropriately omitted here to avoid redundancy. The apparatus 100, 200 comprises at least one software functional module that can be stored in a memory in the form of software or firmware or solidified in an operating system of the apparatus 100, 200. Specifically, the method comprises the following steps:
referring to fig. 6, the apparatus 100 is applied to a first dubbing terminal of a first user, and includes: the system comprises a first receiving module 101, an obtaining module 102, a display module 103, a playing module 104, an acquisition processing module 105 and a first sending module 106;
the first receiving module 101 is configured to receive a dubbing request and to receive dubbing indication information sent from an opposite-end device;
the obtaining module 102 is configured to obtain an audio/video to be dubbed without audio data when receiving a dubbing request;
the display module 103 is configured to display dubbing trigger information when receiving dubbing indication information sent by an opposite-end device; the dubbing indication information represents that the second user dubs the current dubbing fragment in the video to be dubbed completely;
the playing module 104 is configured to play a next to-be-dubbed segment of the current to-be-dubbed segment in the to-be-dubbed audio/video when the dubbing triggering information is triggered; the acquisition processing module 105 is configured to acquire dubbing data of a first user; the first sending module 106 is configured to send the dubbing data to the peer device.
In this embodiment of the present application, the first receiving module 101 is further configured to receive a role selection instruction; the acquisition processing module 105 is configured to determine a to-be-dubbed segment corresponding to the first dubbing terminal according to the role selection instruction; the next segment to be dubbed is one of the segments to be dubbed corresponding to the first dubbing terminal.
In this embodiment of the present application, the dubbing indication information sent by the peer device includes: in the audio and video to be dubbed, the audio data of the current clip to be dubbed; the playing module 104 is further configured to play the current clip to be dubbed and the audio data of the current clip to be dubbed, and the display module 103 is configured to display the dubbing trigger information after the playing is completed.
In this embodiment of the present application, the obtaining module 102 is further configured to obtain original audio data corresponding to an audio and video to be dubbed before receiving dubbing indication information sent by an opposite end device;
the collecting and processing module 105 is further configured to, after collecting the dubbing data and before the first sending module 106 sends the dubbing data to the peer device, compare the collected dubbing data with the original audio data, and determine a score of the dubbing data according to a comparison result. The display module 103 is also used for displaying the scores.
Referring to fig. 7, the apparatus 200 is applied to a dubbing server, and includes: a second receiving module 201, a second sending module 202 and a processing module 203;
the second receiving module 201 is configured to receive dubbing data of a current to-be-dubbed clip in the to-be-dubbed audio and video sent by the first dubbing terminal;
the processing module 203 is configured to determine a second dubbing terminal corresponding to a next to-be-dubbed fragment when dubbing data of a current to-be-dubbed fragment in the to-be-dubbed audio and video sent by the first dubbing terminal is received and the next to-be-dubbed fragment exists in the to-be-dubbed audio and video; the first dubbing terminal and the second dubbing terminal are dubbing terminals in the dubbing task; the dubbing task is a task of dubbing the audio and video to be dubbed;
the second sending module 202 is configured to send dubbing indication information to the second dubbing terminal, so that the second dubbing terminal collects dubbing data of a next to-be-dubbed segment.
In this embodiment of the application, the second sending module 202 is further configured to send dubbing data of a current to-be-dubbed segment to all dubbing terminals except the first dubbing terminal in the dubbing task when the second receiving module 201 receives the dubbing data of the current to-be-dubbed segment in the to-be-dubbed audio and video sent by the first dubbing terminal.
In this embodiment of the application, when the second receiving module 201 receives dubbing data of a current to-be-dubbed segment in the to-be-dubbed audio and video sent by the first dubbing terminal, and a next to-be-dubbed segment does not exist in the to-be-dubbed audio and video, the processing module 203 is further configured to:
informing all dubbing terminals in the dubbing task of completing dubbing so that all dubbing terminals in the dubbing task respectively synthesize audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video;
or informing a third dubbing terminal in the dubbing task of completing dubbing so that the third dubbing terminal synthesizes the audio and video to be dubbed and the obtained dubbing data to obtain the target dubbing video. At this time, the second receiving module 201 is further configured to receive the target dubbing video reported by the third dubbing terminal, and the second sending module 202 is further configured to send the target dubbing video to all the dubbing terminals except the third dubbing terminal in the dubbing task.
In this embodiment of the application, when the second receiving module 201 receives dubbing data of a current to-be-dubbed segment in an audio/video to be dubbed sent by the first dubbing terminal, and a next to-be-dubbed segment does not exist in the audio/video to be dubbed, the processing module 203 is further configured to synthesize the audio/video to be dubbed and the dubbing data to obtain a target dubbing video; the second sending module 202 is further configured to send the target dubbing video to all dubbing terminals in the dubbing task.
In this embodiment of the application, the processing module 203 is further configured to notify all dubbing terminals in the dubbing task of completing dubbing before synthesizing the audio/video to be dubbed and the dubbing data, and determine that a video synthesis request sent by any dubbing terminal in the dubbing task is received.
In this embodiment of the application, the processing module 203 is further configured to, when the second receiving module 201 receives dubbing data of a current to-be-dubbed segment in the to-be-dubbed audio/video sent by the first dubbing terminal, compare the dubbing data of the current to-be-dubbed segment with pre-stored original audio data of the to-be-dubbed audio/video, and determine a score of the dubbing data according to a comparison result. The second sending module 202 is further configured to send the score to the first dubbing terminal.
In this embodiment of the present application, the processing module 203 is further configured to determine that a score acceptance instruction sent by the first dubbing terminal is received before determining the second dubbing terminal corresponding to the next dubbing segment of the current dubbing segment.
It should be understood that, for the sake of brevity, the contents described in some embodiments are not repeated in this embodiment.
Example four:
the present embodiment provides an electronic device, which is shown in fig. 8 and includes a processor 801, a memory 802, and a communication bus 803. Wherein:
the communication bus 803 is used to enable communications between the processor 801 and the memory 802.
The processor 801 is configured to execute one or more first programs stored in the memory 802 to implement the steps performed by the dubbing terminal according to the first embodiment and/or the second embodiment;
alternatively, the processor 801 is configured to execute one or more second programs stored in the memory 802 to implement the steps performed by the dubbing server in the first embodiment and/or the second embodiment.
It will be appreciated that the configuration shown in fig. 8 is merely illustrative and that the electronic device may include more or fewer components than shown in fig. 8 or have a different configuration than shown in fig. 8. For example, when the electronic device is a dubbing terminal, it also has an audio data collection device such as a microphone to collect audio data. In addition, the system is also provided with a display device such as a display screen and the like so as to realize the playing display of the video and the display of the scores. And when the electronic equipment is a dubbing server, the electronic equipment is provided with a communication module so as to realize communication with the dubbing terminal.
The present embodiment further provides a readable storage medium, such as a floppy disk, an optical disk, a hard disk, a flash Memory, a usb (Secure Digital Card), an MMC (Multimedia Card), etc., in which one or more programs for implementing the above steps are stored, and the one or more programs can be executed by one or more processors to implement the steps executed by the dubbing terminal according to the first embodiment and/or the second embodiment, or the steps executed by the dubbing server according to the first embodiment and/or the second embodiment. And will not be described in detail herein.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described embodiments of the apparatus are merely illustrative, and for example, the division of the units is only one logical division, and there may be other divisions when actually implemented, and for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some communication interfaces, and may be in an electrical, mechanical or other form.
In addition, units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
Furthermore, the functional modules in the embodiments of the present application may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
In this document, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions.
In this context, a plurality means two or more.
The above description is only an example of the present application and is not intended to limit the scope of the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (15)

1. An interactive dubbing method applied to a first dubbing terminal of a first user, comprising:
when a dubbing request is received, acquiring an audio and video to be dubbed without audio data;
when receiving dubbing indication information sent by opposite terminal equipment, displaying dubbing trigger information; the dubbing indication information represents that the dubbing of the current to-be-dubbed fragment in the to-be-dubbed audio and video is finished by a second user;
when the dubbing triggering information is triggered, playing a next to-be-dubbed segment of the current to-be-dubbed segment in the to-be-dubbed audio and video, and collecting dubbing data of the first user;
and sending the dubbing data to the opposite terminal equipment.
2. The interactive dubbing method of claim 1, wherein before the obtaining of the audio-video to be dubbed without audio data, further comprising:
receiving a role selection instruction;
determining a to-be-dubbed fragment corresponding to the first dubbing terminal according to the role selection instruction; and the next segment to be dubbed is one of the segments to be dubbed corresponding to the first dubbing terminal.
3. The interactive dubbing method of claim 1, wherein the dubbing indication information sent from the peer device comprises: in the audio and video to be dubbed, the second dubbing terminal of the second user acquires the audio data of the current clip to be dubbed;
when receiving dubbing indication information sent by the opposite terminal device, the displaying dubbing trigger information comprises:
and playing the current clip to be dubbed and the audio data of the current clip to be dubbed, and displaying the dubbing triggering information after the playing is finished.
4. The interactive dubbing method of any one of claims 1 to 3, wherein before receiving dubbing indication information from the peer device, the method further comprises: acquiring original audio data corresponding to the audio and video to be matched;
after the collecting the dubbing data of the first user, before sending the dubbing data to the peer device, the method further includes:
comparing the collected dubbing data with the original audio data, determining the score of the dubbing data according to the comparison result, and displaying the score.
5. An interactive dubbing method, applied to a dubbing server, comprising:
when dubbing data of a current dubbing fragment in an audio and video to be dubbed sent by a first dubbing terminal is received and a next dubbing fragment exists in the audio and video to be dubbed, determining a second dubbing terminal corresponding to the next dubbing fragment; the first dubbing terminal and the second dubbing terminal are dubbing terminals in a dubbing task; the dubbing task is a task for dubbing the audio and video to be dubbed;
and sending dubbing indication information to the second dubbing terminal so that the second dubbing terminal collects dubbing data of the next to-be-dubbed clip.
6. The interactive dubbing method of claim 5, wherein when receiving dubbing data of a current dubbing clip in the audio-video to be dubbed sent by the first dubbing terminal, the method further comprises:
and sending the dubbing data of the current to-be-dubbed segment to all the dubbing terminals except the first dubbing terminal in the dubbing task.
7. The interactive dubbing method of claim 6, wherein when dubbing data of a current dubbing fragment in the audio-video to be dubbed sent by the first dubbing terminal is received and a next dubbing fragment does not exist in the audio-video to be dubbed, the interactive dubbing method further comprises:
informing all dubbing terminals in the dubbing task of completing dubbing so that all dubbing terminals in the dubbing task respectively synthesize the audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video;
or, a third dubbing terminal in the dubbing task is informed of completing dubbing, so that the third dubbing terminal synthesizes the audio and video to be dubbed and the obtained dubbing data to obtain a target dubbing video; receiving a target dubbing video reported by the third dubbing terminal, and sending the target dubbing video to all dubbing terminals except the third dubbing terminal in the dubbing task; and the third dubbing terminal is any dubbing terminal in the dubbing task.
8. The interactive dubbing method of claim 5, wherein when dubbing data of a current dubbing fragment in the audio-video to be dubbed sent by the first dubbing terminal is received and a next dubbing fragment does not exist in the audio-video to be dubbed, the interactive dubbing method further comprises:
synthesizing the audio and video to be dubbed and the dubbing data to obtain a target dubbing video;
and sending the target dubbing video to all dubbing terminals in the dubbing task.
9. The interactive dubbing method of claim 8, wherein before the synthesizing the audio-video to be dubbed and the dubbing data, further comprising:
and informing all dubbing terminals in the dubbing task of completing dubbing, and determining to receive a video synthesis request sent by any dubbing terminal in the dubbing task.
10. The interactive dubbing method of any one of claims 5 to 9, wherein upon receiving dubbing data of a current dubbing clip in the video to be dubbed sent from the first dubbing terminal, the method further comprises:
and comparing dubbing data of the current to-be-dubbed clip with pre-stored original audio data of the to-be-dubbed audio and video, determining the score of the dubbing data according to the comparison result, and sending the score to the first dubbing terminal.
11. The interactive dubbing method of claim 10, wherein before determining the second dubbing terminal corresponding to the next dubbing segment of the current dubbing segment, further comprising:
and determining that a score acceptance instruction sent by the first dubbing terminal is received.
12. An interactive dubbing apparatus, for use in a first dubbing terminal of a first user, comprising: the device comprises a first receiving module, an acquisition module, a display module, a playing module, an acquisition processing module and a first sending module;
the first receiving module is used for receiving a dubbing request and receiving dubbing indication information sent by opposite-end equipment;
the acquisition module is used for acquiring the audio and video to be dubbed without audio data when receiving the dubbing request;
the display module is used for displaying dubbing triggering information when dubbing indication information sent by the opposite terminal equipment is received; the dubbing indication information represents that the dubbing of the current to-be-dubbed fragment in the to-be-dubbed audio and video is finished by a second user;
the playing module is used for playing a next dubbing fragment to be dubbed in the audio and video to be dubbed when the dubbing triggering information is triggered;
the acquisition processing module is used for acquiring dubbing data of the first user;
the first sending module is configured to send the dubbing data to the peer device.
13. An interactive dubbing device, applied to a dubbing server, comprising: the device comprises a second receiving module, a second sending module and a processing module;
the second receiving module is used for receiving dubbing data of a current to-be-dubbed fragment in the audio and video to be dubbed sent by the first dubbing terminal;
the processing module is used for determining a second dubbing terminal corresponding to a next dubbing fragment when dubbing data of a current dubbing fragment in an audio and video to be dubbed sent by a first dubbing terminal are received and the next dubbing fragment exists in the audio and video to be dubbed; the first dubbing terminal and the second dubbing terminal are dubbing terminals in a dubbing task; the dubbing task is a task for dubbing the audio and video to be dubbed;
and the second sending module is used for sending dubbing indication information to the second dubbing terminal so that the second dubbing terminal can collect dubbing data of the next to-be-dubbed clip.
14. An electronic device comprising a processor, a memory, and a communication bus;
the communication bus is used for realizing connection communication between the processor and the memory;
the processor is configured to execute one or more first programs stored in the memory to implement the interactive dubbing method of any of claims 1 to 4;
or the processor is configured to execute one or more second programs stored in the memory to implement the interactive dubbing method according to any one of claims 5 to 11.
15. A readable storage medium storing one or more programs, the one or more programs being executable by one or more processors to implement the interactive dubbing method of any one of claims 1 to 4 or the interactive dubbing method of any one of claims 5 to 11.
CN201911041873.6A 2019-10-29 2019-10-29 Interactive dubbing method and device, electronic equipment and readable storage medium Active CN110650366B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911041873.6A CN110650366B (en) 2019-10-29 2019-10-29 Interactive dubbing method and device, electronic equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911041873.6A CN110650366B (en) 2019-10-29 2019-10-29 Interactive dubbing method and device, electronic equipment and readable storage medium

Publications (2)

Publication Number Publication Date
CN110650366A CN110650366A (en) 2020-01-03
CN110650366B true CN110650366B (en) 2021-09-24

Family

ID=68995049

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911041873.6A Active CN110650366B (en) 2019-10-29 2019-10-29 Interactive dubbing method and device, electronic equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN110650366B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111359209B (en) * 2020-02-28 2022-03-29 腾讯科技(深圳)有限公司 Video playing method and device and terminal
CN111462553B (en) * 2020-04-17 2021-03-30 杭州菲助科技有限公司 Language learning method and system based on video dubbing and sound correction training
CN112752142B (en) * 2020-08-26 2022-07-29 腾讯科技(深圳)有限公司 Dubbing data processing method and device and electronic equipment
CN112261435B (en) * 2020-11-06 2022-04-08 腾讯科技(深圳)有限公司 Social interaction method, device, system, equipment and storage medium
CN112954390B (en) * 2021-01-26 2023-05-09 北京有竹居网络技术有限公司 Video processing method, device, storage medium and equipment
CN115037975B (en) * 2021-02-24 2024-03-01 花瓣云科技有限公司 Method for dubbing video, related equipment and computer readable storage medium
CN113630630B (en) * 2021-08-09 2023-08-15 咪咕数字传媒有限公司 Method, device and equipment for processing video comment dubbing information

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101374306A (en) * 2008-09-18 2009-02-25 中兴通讯股份有限公司 Method for prompting key-press sound of mobile phone keys
WO2010051059A1 (en) * 2008-11-03 2010-05-06 At&T Intellectual Property I, L.P. System and method for distributing media content
CN105959773A (en) * 2016-04-29 2016-09-21 魔方天空科技(北京)有限公司 Multimedia file processing method and device
CN106792013A (en) * 2016-11-29 2017-05-31 青岛海尔多媒体有限公司 A kind of method, the TV interactive for television broadcast sounds
EP3285455A1 (en) * 2015-06-16 2018-02-21 Huawei Technologies Co., Ltd. Content item aggregation method and related device and communication system
CN107809666A (en) * 2017-10-26 2018-03-16 费非 Voice data merging method, device storage medium and processor
CN110366032A (en) * 2019-08-09 2019-10-22 腾讯科技(深圳)有限公司 Video data handling procedure, device and video broadcasting method, device

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7447126B2 (en) * 2003-05-12 2008-11-04 Matsushita Electrical Industrial Co., Ltd. Recording device, and related control method, computer program and system LSI
JP2007336116A (en) * 2006-06-14 2007-12-27 Hitachi Ltd Video recording and reproducing apparatus
CN101937570A (en) * 2009-10-11 2011-01-05 上海本略信息科技有限公司 Animation mouth shape automatic matching implementation method based on voice and text recognition
CN105975568B (en) * 2016-04-29 2020-04-03 腾讯科技(深圳)有限公司 Audio processing method and device
CN106060424A (en) * 2016-06-14 2016-10-26 徐文波 Video dubbing method and device
CN107659850B (en) * 2016-11-24 2019-09-17 腾讯科技(北京)有限公司 Media information processing method and device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101374306A (en) * 2008-09-18 2009-02-25 中兴通讯股份有限公司 Method for prompting key-press sound of mobile phone keys
WO2010051059A1 (en) * 2008-11-03 2010-05-06 At&T Intellectual Property I, L.P. System and method for distributing media content
EP3285455A1 (en) * 2015-06-16 2018-02-21 Huawei Technologies Co., Ltd. Content item aggregation method and related device and communication system
CN105959773A (en) * 2016-04-29 2016-09-21 魔方天空科技(北京)有限公司 Multimedia file processing method and device
CN106792013A (en) * 2016-11-29 2017-05-31 青岛海尔多媒体有限公司 A kind of method, the TV interactive for television broadcast sounds
CN107809666A (en) * 2017-10-26 2018-03-16 费非 Voice data merging method, device storage medium and processor
CN110366032A (en) * 2019-08-09 2019-10-22 腾讯科技(深圳)有限公司 Video data handling procedure, device and video broadcasting method, device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"基于web的远程教学系统交互设计研究";付增福;《中国优秀硕士学位论文全文数据库》;20110315;全文 *

Also Published As

Publication number Publication date
CN110650366A (en) 2020-01-03

Similar Documents

Publication Publication Date Title
CN110650366B (en) Interactive dubbing method and device, electronic equipment and readable storage medium
WO2022121601A1 (en) Live streaming interaction method and apparatus, and device and medium
CN110505491B (en) Live broadcast processing method and device, electronic equipment and storage medium
CN107316520B (en) Video teaching interaction method, device, equipment and storage medium
CN104333783B (en) The order method and system, relevant device of a kind of formulation task
CN112653902B (en) Speaker recognition method and device and electronic equipment
CN106971009B (en) Voice database generation method and device, storage medium and electronic equipment
CN108847214A (en) Method of speech processing, client, device, terminal, server and storage medium
CN108960158A (en) A kind of system and method for intelligent sign language translation
CN105704504A (en) A method and apparatus for inserting push information in video direct broadcast
CN111294606B (en) Live broadcast processing method and device, live broadcast client and medium
CN111599359A (en) Man-machine interaction method, server, client and storage medium
CN108573393A (en) Comment information processing method, device, server and storage medium
CN108989864A (en) A kind of even wheat method, apparatus, equipment and storage medium
TW201225669A (en) System and method for synchronizing with multimedia broadcast program and computer program product thereof
KR102131106B1 (en) Karaoke player that enables personal broadcasting by connecting to a broadcasting service server based on a two-dimensional code and operating method thereof
CN114363650B (en) Live broadcast room public screen text display method, electronic equipment and storage medium
CN109413455A (en) A kind of user information display methods and device connecting wheat interaction for voice
CN111741333B (en) Live broadcast data acquisition method and device, computer equipment and storage medium
JP7376235B2 (en) Video distribution system
CN110275860B (en) System and method for recording teaching process
JP6344731B1 (en) Content evaluation system
CN111160051A (en) Data processing method and device, electronic equipment and storage medium
CN114341866A (en) Simultaneous interpretation method, device, server and storage medium
JP2020188346A (en) Video distribution system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant