WO2021068558A1 - 一种同声字幕翻译方法、智能电视及存储介质 - Google Patents
一种同声字幕翻译方法、智能电视及存储介质 Download PDFInfo
- Publication number
- WO2021068558A1 WO2021068558A1 PCT/CN2020/098074 CN2020098074W WO2021068558A1 WO 2021068558 A1 WO2021068558 A1 WO 2021068558A1 CN 2020098074 W CN2020098074 W CN 2020098074W WO 2021068558 A1 WO2021068558 A1 WO 2021068558A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- subtitle
- subtitles
- cloud server
- simultaneous
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/233—Processing of audio elementary streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
- H04N21/42212—Specific keyboard arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/42204—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
- H04N21/42206—User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
- H04N21/42224—Touch pad or touch panel provided on the remote control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/441—Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
- H04N21/4415—Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
- H04N21/4856—End-user interface for client configuration for language selection, e.g. for the menu or subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
- H04N21/4858—End-user interface for client configuration for modifying screen layout parameters, e.g. fonts, size of the windows
Definitions
- the present disclosure relates to the technical field of smart TVs, such as a simultaneous subtitle translation method, smart TVs, and storage media.
- TVs large-screen smart display devices
- TV sets have unique advantages in watching movies and distance education in the current and future of intelligent development.
- society the process of people's integration into globalization is accelerating.
- language is the biggest obstacle to integrating into the international society and understanding international culture.
- it will bring great obstacles to hope to learn and cultural exchanges.
- the main purpose of the present disclosure is to provide a simultaneous subtitle translation method, a smart TV, and a storage medium, and aims to solve the problem of language barriers in the process of using high-quality audiovisual resources for users in the prior art.
- the simultaneous subtitle translation method includes the following steps:
- the cloud server is configured to translate the voice information into the target subtitles corresponding to the target language
- the target subtitles are matched with the time axis of the playback resource in real time, and the target subtitles are displayed synchronously during the playback.
- the simultaneous subtitle translation method wherein the receiving operation instruction of the user to select a target language, collecting voice information in the playback resource, and sending the target language and the voice information to a cloud server, specifically include:
- the collection of voice information in the playback resource specifically includes:
- the voice information in the video resource is identified and separated.
- the simultaneous subtitle translation method wherein the receiving an operation instruction of a user to select a target language, afterwards, further includes:
- the simultaneous subtitle translation method wherein the receiving an operation instruction of a user to select a target language, afterwards, further includes:
- the subtitle style of the target subtitle selected by the user to be displayed on the display interface is received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size.
- the display position of the target subtitle is above or below the original subtitle.
- the subtitle color of the target subtitle is different from the subtitle color of the original subtitle.
- the simultaneous subtitle translation method wherein the real-time matching of the target subtitles with the time axis of the playback resource, and synchronously displaying the target subtitles during the playback process specifically includes:
- the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on the display interface during the playback of the playback resource.
- the simultaneous subtitle translation method wherein the synchronously displaying the target subtitles during the playback process further includes:
- the original subtitles and the target subtitles are displayed on the display interface to form a comparison effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language.
- the cloud server is configured to translate the voice information into the target subtitle corresponding to the target language, which specifically includes:
- the cloud server receives the voice information and the target language
- the cloud server recognizes and translates the voice information according to the target language, and generates the target subtitles corresponding to the target language;
- the cloud server returns the target subtitles to the smart TV.
- the cloud server uses artificial intelligence to perform voice recognition and translation to generate the target subtitles desired by the user.
- the translation by the cloud server is online real-time translation.
- the present disclosure also provides a smart TV, wherein the smart TV includes: a memory, a processor, and a simultaneous subtitle translation program stored on the memory and running on the processor When the simultaneous subtitle translation program is executed by the processor, the steps of the simultaneous subtitle translation method described above are implemented.
- the present disclosure also provides a simultaneous subtitle translation system, wherein the simultaneous subtitle translation system includes the smart TV as described above, and also includes a cloud that has established a communication connection with the smart TV.
- the smart TV is used to receive the user's operation instructions to select the target language, collect the voice information in the playback resource and send it to the cloud server; also used to receive the target subtitles returned by the cloud server, and the target subtitles Perform real-time matching with the playback time axis of the playback resource, and display the target subtitles synchronously during playback;
- the cloud server is used to translate the voice information into the target subtitles corresponding to the target language, and The target subtitles are sent to the smart TV.
- the present disclosure also provides a storage medium, wherein the storage medium stores a simultaneous subtitle translation program, and the simultaneous subtitle translation program is executed by a processor to realize the simultaneous subtitle as described above.
- the steps of the translation method are described above.
- the present disclosure collects the voice information in the playback resource by receiving the operation instruction of the user to select the target language, and sends the target language and the voice information to the cloud server; receiving the target subtitles returned by the cloud server, the cloud server It is used to translate the voice information into the target subtitles corresponding to the target language; match the target subtitles with the time axis of the playback resource in real time, and display the target subtitles synchronously during the playback process.
- the present disclosure collects the voice information in the playback resources, according to the language subtitles required by the user, sends the voice information to the cloud server for synchronous translation into the target subtitles corresponding to the target language selected by the user, and performs synchronous display during playback to satisfy the user
- the demand for subtitles in multiple languages enables users to easily understand the content and information expressed by audio and video in different languages, and realize convenient viewing or learning.
- Fig. 1 is a flowchart of a preferred embodiment of the simultaneous subtitle translation method of the present disclosure
- step S10 is a flowchart of step S10 in the preferred embodiment of the simultaneous subtitle translation method of the present disclosure
- step S30 is a flowchart of step S30 in the preferred embodiment of the simultaneous subtitle translation method of the present disclosure
- FIG. 5 is a schematic diagram of the operating environment of the preferred embodiment of the disclosed smart TV.
- the simultaneous subtitle translation method includes the following steps:
- Step S10 Receive an operation instruction for the user to select a target language, collect voice information in the playback resource, and send the target language and the voice information to a cloud server.
- FIG. 2 is a flowchart of step S10 in the simultaneous subtitle translation method provided by the present disclosure.
- the step S10 includes:
- S12. Collect the voice information in the playback resource that currently needs to be played, and send the compressed voice information and the target language to the cloud server.
- the remote control The user can send an operation instruction to select the target language to the TV through the remote controller (such as a smart TV).
- the remote controller such as a smart TV
- the user can directly issue an operation instruction to select the target language through the touch screen touch menu
- the TV or touch device system has pre-stored multiple languages, and you can add other languages that are not stored.
- the subtitle style of the target subtitle selected by the user to be displayed on the display interface can also be received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size
- the user selects the color of the final output target subtitle to be red (for example, the original subtitle is black, the two colors are different for comparison and viewing), the subtitle transparency can be selected according to the user’s needs, and the display position can be above or below the original subtitle.
- the font size can also be selected according to user needs, and other subtitle parameters or related settings can also be selected.
- the TV or touch device After the TV or touch device receives the operation instruction for the user to select the target language, it recognizes the operation instruction and executes the corresponding function, that is, collects the playback resource that needs to be played (or is currently playing) according to the target language
- the voice information in the collected playback resource specifically includes: when the playback resource is an audio resource (for example, audio is played through a TV, although there is no changed picture, there will be a display interface, showing When the interface may be a certain playback interface that remains unchanged during the audio playback process, then the voice information in the audio resource can be directly obtained; when the playback resource is a video resource (including audio information and video information), Then it is necessary to identify and separate the voice information in the video resource. Because the target caption generation only needs to obtain the voice information.
- an audio resource for example, audio is played through a TV, although there is no changed picture, there will be a display interface, showing When the interface may be a certain playback interface that remains unchanged during the audio playback process, then the voice information
- Compress the acquired voice information (to facilitate data transmission), and finally send the compressed voice information and the target language to the cloud server, so that the cloud server can respond to the needs of the user
- the voice data is processed.
- Step S20 Receive the target subtitles returned by the cloud server, where the cloud server is used to translate the voice information into the target subtitles corresponding to the target language.
- the cloud server mainly performs online real-time translation operations on the voice information according to the target language, for example, the The cloud server can perform voice recognition and translation through artificial intelligence (AI), and generate the target subtitles that the user wants (target subtitles mainly refer to the user's native language).
- AI artificial intelligence
- target subtitles mainly refer to the user's native language.
- the cloud server After the cloud server receives the voice information and the target language, the cloud server according to the target language (for example, the original audio and video resources are French, and the user needs Chinese, then the target language is Chinese, the original The subtitles are French) recognize and translate the voice information, and generate the target subtitles corresponding to the target language (for example, generate Chinese subtitles), and the cloud server transmits the target subtitles back to the TV (for example, a smart TV). ) Or touch the device.
- the target language for example, the original audio and video resources are French, and the user needs Chinese
- the target language is Chinese
- the original The subtitles are French
- the cloud server transmits the target subtitles back to the TV (for example, a smart TV). ) Or touch the device.
- the translation of the cloud server is online real-time translation, which does not affect the user's viewing.
- the TV or touch device can also choose to broadcast and translate according to the user's network environment, or translate first and then broadcast. Provide users with a consistent and transparent viewing experience.
- Step S30 Match the target subtitles with the time axis of the playback resource in real time, and display the target subtitles synchronously during the playback process.
- FIG. 3 is a flowchart of step S30 in the simultaneous subtitle translation method provided by the present disclosure.
- the step S30 includes:
- the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on a display interface during the playback of the playback resource.
- the cloud server when the cloud server recognizes and translates the voice information according to the target language, and generates the target subtitles corresponding to the target language, the cloud server transmits the target subtitles back to the TV or After touching the device, the TV or touch device matches the target subtitles and pictures according to the target subtitles, and outputs synchronously according to the time axis, that is, real-time matching of the target subtitles and the time axis of the playback resource, so that the picture, voice and target subtitles Synchronous matching; when the time axis matching is completed, the target subtitles are displayed synchronously on the display interface (i.e. display screen) during audio or video playback. For example, during audio or video playback, the original subtitles are compared and displayed on the display screen.
- the display interface i.e. display screen
- the subtitles and the target subtitles are used to form the contrast effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language, which not only preserves the original sound but also allows viewing Understand the meaning to be expressed. For example, if the user wants to learn the language represented by the original subtitles, then the comparison display can assist the user in learning other languages.
- the simultaneous subtitle translation method of the present disclosure decodes the decodable audio and video resources and samples the key information.
- the key sampling information is transferred to the cloud server, which is translated by the cloud server and then returned to the smart
- subtitles are used to cooperate with the original language output and presented to the user, so that the user can easily understand the content information expressed by audio and video in different languages, realize convenient communication or learning, can give the user a real context, and the user can immediately understand
- the information to be expressed by audio and video has high processing efficiency and a wide range of applications.
- Step S0 The user selects the target language according to the subtitles that need to be output finally (the terminal device takes a smart TV as an example, and the smart TV receives the user's operation instructions);
- Step S1 The smart TV collects voice information in playback resources (such as audio or video);
- Step S2 the smart TV sends the target language and the voice information to the cloud server;
- Step S3 The cloud server is used to translate the voice information into the target subtitles corresponding to the target language, and return the target subtitles to the smart TV;
- Step S4 The smart TV receives the target subtitles sent by the cloud server, and matches the target subtitles with the time axis of the playback resource in real time, so that the screen, voice, and subtitles are output simultaneously (step S4 and step S1 can be Synchronize through the background, without affecting the user's viewing);
- Step S5 After the time axis matching of the target subtitle and the playback resource is completed, the target subtitle is synchronously displayed on the display interface of the smart TV during the playback of the playback resource.
- this function can obtain more potential customers, improve the use time of large-screens and customer viscosity, and obtain more commercial value.
- the present disclosure also provides a smart TV correspondingly.
- the smart TV includes a processor 10, a memory 20 and a display 30.
- FIG. 5 only shows part of the components of the smart TV, but it should be understood that it is not required to implement all the shown components, and more or fewer components may be implemented instead.
- the memory 20 may be an internal storage unit of the smart TV, such as a hard disk or memory of the smart TV.
- the memory 20 may also be an external storage device of the smart TV, for example, a plug-in hard disk equipped on the smart TV, a smart memory card (Smart Media Card, SMC), and a secure digital (Secure Digital). Digital, SD) card, flash card (Flash Card), etc.
- the memory 20 may also include both an internal storage unit of the smart TV and an external storage device.
- the memory 20 is used to store application software and various data installed on the smart TV, such as program code for the smart TV installed.
- the memory 20 can also be used to temporarily store data that has been output or will be output.
- a simultaneous subtitle translation program 40 is stored on the memory 20, and the simultaneous subtitle translation program 40 can be executed by the processor 10, so as to realize the simultaneous subtitle translation method in the present application.
- the processor 10 may be a central processing unit (CPU), microprocessor or other data processing chip in some embodiments, and is used to run the program code or process data stored in the memory 20, for example Perform the simultaneous subtitle translation method and so on.
- CPU central processing unit
- microprocessor or other data processing chip in some embodiments, and is used to run the program code or process data stored in the memory 20, for example Perform the simultaneous subtitle translation method and so on.
- the display 30 may be an LED display, a liquid crystal display, a touch liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light-emitting diode) touch device, etc.
- the display 30 is used for displaying information on the smart TV and for displaying a visualized user interface.
- the components 10-30 of the smart TV communicate with each other through a system bus.
- the cloud server is configured to translate the voice information into the target subtitles corresponding to the target language
- the target subtitles are matched with the time axis of the playback resource in real time, and the target subtitles are displayed synchronously during the playback process.
- the receiving the operation instruction of the user to select the target language, collecting the voice information in the playback resource, and sending the target language and the voice information to the cloud server specifically includes:
- the voice information in the collection and playback resources specifically includes:
- the voice information in the video resource is identified and separated.
- the key sampling information is delivered to the cloud server according to the user's needs for the language.
- the subtitle style of the target subtitle selected by the user to be displayed on the display interface is received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size.
- the display position of the target subtitle is above or below the original subtitle.
- the subtitle color of the target subtitle is different from the subtitle color of the original subtitle.
- the real-time matching of the target subtitles with the time axis of the playback resource, and synchronously displaying the target subtitles during the playback process specifically includes:
- the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on the display interface during the playback of the playback resource.
- the synchronously displaying the target subtitles during the playback process further includes:
- the original subtitles and the target subtitles are displayed on the display interface to form a comparison effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language.
- the present disclosure also provides a storage medium, wherein the storage medium stores a simultaneous subtitle translation program, and when the simultaneous subtitle translation program is executed by a processor, the steps of the simultaneous subtitle translation method described above are realized.
- the present disclosure provides a simultaneous subtitle translation method, a smart TV, and a storage medium.
- the method includes: receiving an operation instruction for a user to select a target language, collecting voice information in a playback resource, and combining the target language with The voice information is sent to the cloud server; the target subtitles returned by the cloud server are received, and the cloud server is used to translate the voice information into the target subtitles corresponding to the target language; and the target subtitles are combined with the target subtitles.
- the time axis of the playback resource is matched in real time, and the target subtitles are displayed synchronously during the playback process.
- the present disclosure collects the voice information in the playback resources, according to the language subtitles required by the user, sends the voice information to the cloud server for synchronous translation into the target subtitles corresponding to the target language selected by the user, and performs synchronous display during playback to satisfy the user
- the demand for subtitles in multiple languages enables users to easily understand the content and information expressed by audio and video in different languages, and realize convenient viewing or learning.
- the processes in the methods of the above-mentioned embodiments can be implemented by instructing relevant hardware (such as a processor, a controller, etc.) through a computer program, and the program can be stored in a computer program.
- the program may include the processes of the foregoing method embodiments when executed.
- the storage medium mentioned may be a memory, a magnetic disk, an optical disk, and the like.
- the embodiment of the present disclosure provides a simultaneous subtitle translation method, smart TV, and storage medium.
- the voice information is sent to the cloud server for simultaneous translation into user selection
- the target subtitles corresponding to the target language of the target language are displayed synchronously during playback to meet the needs of users for subtitles in multiple languages, so that users can easily understand the content information expressed by audio and video in different languages, and realize convenient viewing or learning.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Electrically Operated Instructional Devices (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
Claims (15)
- 一种同声字幕翻译方法,其特征在于,所述同声字幕翻译方法包括:接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器;接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。
- 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器,具体包括:接收用户通过遥控器按键或者触摸屏触摸菜单选择目标语言的操作指令,所述目标语言包括多种预先存储的语言;采集当前需要播放的所述播放资源中的所述语音信息,将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器。
- 根据权利要求2所述的同声字幕翻译方法,其特征在于,所述采集播放资源中的语音信息,具体包括:当所述播放资源为音频资源时,直接获取所述音频资源中的语音信息;当所述播放资源为视频资源时,识别并分离出所述视频资源中的语音信息。
- 根据权利要求1或2所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,之后还包括:通过对可解码的音视频资源进行解码和关键信息采样,根据用户对语种的需求,将关键采样信息传递到云端服务器。
- 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,之后还包括:接收用户选择在显示界面上显示所述目标字幕的字幕样式,所述字幕样式的参数包括字幕颜色、字幕透明度、显示位置以及字体大小。
- 根据权利要求5所述的同声字幕翻译方法,其特征在于,所述目标字幕的显示位置在原始字幕的上方或者下方。
- 根据权利要求6所述的同声字幕翻译方法,其特征在于,所述目标字幕的字幕颜色与所述原始字幕的字幕颜色不同。
- 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕,具体包括:获取所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,所述匹配包括画面、语音和字幕的同步匹配;当所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在显示界面上同步显示所述目标字幕。
- 根据权利要求8所述的同声字幕翻译方法,其特征在于,所述在播放过程中同步显示所述目标字幕,还包括:在所述播放资源播放过程中,在显示界面上对比显示原始字幕和所述目标字幕,用于形成所述原始字幕和所述目标字幕的对比效果以进行原始字幕语言的辅助学习。
- 根据权利要求1或2所述的同声字幕翻译方法,其特征在于,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,具体包括:所述云端服务器接收所述语音信息和所述目标语言;所述云端服务器根据所述目标语言对所述语音信息进行识别和翻译,并生成所述目标语言对应的所述目标字幕;所述云端服务器将所述目标字幕回传至智能电视。
- 根据权利要求10所述的同声字幕翻译方法,其特征在于,所述云端服务器通过人工智能进行语音识别和翻译,生成用户想要的目标字幕。
- 根据权利要求10所述的同声字幕翻译方法,其特征在于,所述云端服务器的翻译为在线实时翻译。
- 一种智能电视,其特征在于,所述智能电视包括:存储器、处理器及存储在所 述存储器上并可在所述处理器上运行的同声字幕翻译程序,所述同声字幕翻译程序被所述处理器执行时实现如权利要求1-9任一项所述的同声字幕翻译方法的步骤。
- 一种同声字幕翻译系统,其特征在于,所述同声字幕翻译系统包括如权利要求13所述的智能电视,还包括一与所述智能电视已建立通信连接的云端服务器;所述智能电视用于接收用户选择目标语言的操作指令,采集播放资源中的语音信息并发送至云端服务器;还用于接收所述云端服务器回传的目标字幕,并将所述目标字幕与所述播放资源的播放时间轴进行实时匹配,在播放过程中同步显示所述目标字幕;所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,并将所述目标字幕发送至所述智能电视。
- 一种存储介质,其特征在于,所述存储介质存储有同声字幕翻译程序,所述同声字幕翻译程序被处理器执行时实现如权利要求1-9任一项所述的同声字幕翻译方法的步骤。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910950069.3 | 2019-10-08 | ||
CN201910950069.3A CN110769265A (zh) | 2019-10-08 | 2019-10-08 | 一种同声字幕翻译方法、智能电视及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021068558A1 true WO2021068558A1 (zh) | 2021-04-15 |
Family
ID=69330967
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/098074 WO2021068558A1 (zh) | 2019-10-08 | 2020-06-24 | 一种同声字幕翻译方法、智能电视及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110769265A (zh) |
WO (1) | WO2021068558A1 (zh) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110769265A (zh) * | 2019-10-08 | 2020-02-07 | 深圳创维-Rgb电子有限公司 | 一种同声字幕翻译方法、智能电视及存储介质 |
CN113473238B (zh) * | 2020-04-29 | 2022-10-18 | 海信集团有限公司 | 一种智能设备及视频通话时的同声翻译方法 |
CN111683266A (zh) * | 2020-05-06 | 2020-09-18 | 厦门盈趣科技股份有限公司 | 一种视频同声翻译配置字幕方法及终端 |
CN112055261A (zh) * | 2020-07-14 | 2020-12-08 | 北京百度网讯科技有限公司 | 字幕展示方法、装置、电子设备和存储介质 |
CN113301357B (zh) * | 2020-07-27 | 2022-11-29 | 阿里巴巴集团控股有限公司 | 直播方法、装置及电子设备 |
CN112118309A (zh) * | 2020-09-16 | 2020-12-22 | 三星电子(中国)研发中心 | 音频翻译方法和系统 |
CN112584209A (zh) * | 2020-12-04 | 2021-03-30 | 深圳创维-Rgb电子有限公司 | 一种显示方法、装置、存储介质及智能电视 |
CN112601101B (zh) * | 2020-12-11 | 2023-02-24 | 北京有竹居网络技术有限公司 | 一种字幕显示方法、装置、电子设备及存储介质 |
CN112601102A (zh) * | 2020-12-11 | 2021-04-02 | 北京有竹居网络技术有限公司 | 同声传译字幕的确定方法、装置、电子设备及存储介质 |
CN112995749B (zh) * | 2021-02-07 | 2023-05-26 | 北京字节跳动网络技术有限公司 | 视频字幕的处理方法、装置、设备和存储介质 |
CN113709509A (zh) * | 2021-08-05 | 2021-11-26 | 中移(杭州)信息技术有限公司 | 音视频数据的传输方法、装置以及存储介质 |
CN113709579B (zh) * | 2021-08-05 | 2023-07-21 | 中移(杭州)信息技术有限公司 | 音视频数据的传输方法、装置以及存储介质 |
CN113709558A (zh) * | 2021-10-09 | 2021-11-26 | 立讯电子科技(昆山)有限公司 | 多媒体处理方法及多媒体交互系统 |
CN113992926B (zh) * | 2021-10-19 | 2023-09-12 | 北京有竹居网络技术有限公司 | 界面显示方法、装置、电子设备和存储介质 |
CN115051991A (zh) * | 2022-07-08 | 2022-09-13 | 北京有竹居网络技术有限公司 | 音频处理方法、装置、存储介质与电子设备 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170092274A1 (en) * | 2015-09-24 | 2017-03-30 | Otojoy LLC | Captioning system and/or method |
CN106792097A (zh) * | 2016-12-27 | 2017-05-31 | 深圳Tcl数字技术有限公司 | 音频信号字幕匹配方法及装置 |
CN107277646A (zh) * | 2017-08-08 | 2017-10-20 | 四川长虹电器股份有限公司 | 一种音视频资源的字幕配置系统 |
CN108319590A (zh) * | 2018-01-25 | 2018-07-24 | 芜湖应天光电科技有限责任公司 | 一种基于云服务的自适应翻译机 |
CN207720194U (zh) * | 2017-01-05 | 2018-08-10 | 陈伯妤 | 一种同传手机 |
CN110769265A (zh) * | 2019-10-08 | 2020-02-07 | 深圳创维-Rgb电子有限公司 | 一种同声字幕翻译方法、智能电视及存储介质 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7711543B2 (en) * | 2006-04-14 | 2010-05-04 | At&T Intellectual Property Ii, Lp | On-demand language translation for television programs |
CN101179669B (zh) * | 2006-11-08 | 2011-03-16 | 中兴通讯股份有限公司 | 一种会议电视终端字幕生成叠加方法 |
CN102316277B (zh) * | 2011-09-06 | 2014-10-29 | Tcl新技术(惠州)有限公司 | 字幕系统及其实现方法 |
CN103079039B (zh) * | 2011-10-25 | 2016-08-31 | 康佳集团股份有限公司 | 一种飞播方法和系统 |
CN103327397A (zh) * | 2012-03-22 | 2013-09-25 | 联想(北京)有限公司 | 一种媒体文件的字幕同步显示方法及系统 |
CN103905743A (zh) * | 2012-12-30 | 2014-07-02 | 上海能感物联网有限公司 | 汉语自动实时标注汉语字幕音像录播方法 |
CN103561217A (zh) * | 2013-10-14 | 2014-02-05 | 深圳创维数字技术股份有限公司 | 一种生成字幕的方法及终端 |
CN105681890A (zh) * | 2016-01-26 | 2016-06-15 | 广东欧珀移动通信有限公司 | 一种视频播放终端的字幕显示方法及装置 |
CN105828101B (zh) * | 2016-03-29 | 2019-03-08 | 北京小米移动软件有限公司 | 生成字幕文件的方法及装置 |
CN105848004A (zh) * | 2016-05-16 | 2016-08-10 | 乐视控股(北京)有限公司 | 字幕播放方法、字幕播放装置 |
CN106340291A (zh) * | 2016-09-27 | 2017-01-18 | 广东小天才科技有限公司 | 一种双语字幕制作方法及系统 |
CN106340294A (zh) * | 2016-09-29 | 2017-01-18 | 安徽声讯信息技术有限公司 | 基于同步翻译的新闻直播字幕在线制作系统 |
CN108572950A (zh) * | 2017-03-08 | 2018-09-25 | 捷音特科技股份有限公司 | 耳麦翻译系统 |
CN207264383U (zh) * | 2017-08-23 | 2018-04-20 | 成都铅笔科技有限公司 | 一种实时语音翻译的外语翻译装置 |
CN207530952U (zh) * | 2017-12-13 | 2018-06-22 | 南京洛菲特数码科技有限公司 | 一种多屏图像拼接处理器 |
CN108197123A (zh) * | 2018-02-07 | 2018-06-22 | 云南衍那科技有限公司 | 一种基于智能手表的云翻译系统和方法 |
CN108833810A (zh) * | 2018-06-21 | 2018-11-16 | 珠海金山网络游戏科技有限公司 | 一种三维虚拟偶像直播中实时生成字幕的方法及装置 |
CN108962220B (zh) * | 2018-07-26 | 2021-03-09 | 北京小米移动软件有限公司 | 多媒体文件播放场景下的文本显示方法及装置 |
TWM574267U (zh) * | 2018-08-16 | 2019-02-11 | 張漢龍 | 即時語音自動同步轉譯字幕直播系統 |
-
2019
- 2019-10-08 CN CN201910950069.3A patent/CN110769265A/zh active Pending
-
2020
- 2020-06-24 WO PCT/CN2020/098074 patent/WO2021068558A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170092274A1 (en) * | 2015-09-24 | 2017-03-30 | Otojoy LLC | Captioning system and/or method |
CN106792097A (zh) * | 2016-12-27 | 2017-05-31 | 深圳Tcl数字技术有限公司 | 音频信号字幕匹配方法及装置 |
CN207720194U (zh) * | 2017-01-05 | 2018-08-10 | 陈伯妤 | 一种同传手机 |
CN107277646A (zh) * | 2017-08-08 | 2017-10-20 | 四川长虹电器股份有限公司 | 一种音视频资源的字幕配置系统 |
CN108319590A (zh) * | 2018-01-25 | 2018-07-24 | 芜湖应天光电科技有限责任公司 | 一种基于云服务的自适应翻译机 |
CN110769265A (zh) * | 2019-10-08 | 2020-02-07 | 深圳创维-Rgb电子有限公司 | 一种同声字幕翻译方法、智能电视及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN110769265A (zh) | 2020-02-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021068558A1 (zh) | 一种同声字幕翻译方法、智能电视及存储介质 | |
US20210321157A1 (en) | Special effect processing method and apparatus for live broadcasting, and server | |
KR102246305B1 (ko) | 증강 미디어 서비스 제공 방법, 장치 및 시스템 | |
WO2017181599A1 (zh) | 一种显示视频的方法及装置 | |
CN112163086B (zh) | 多意图的识别方法、显示设备 | |
CN114286173A (zh) | 一种显示设备及音画参数调节方法 | |
WO2017008652A1 (zh) | 信息处理方法及装置 | |
CN112000820A (zh) | 一种媒资推荐方法及显示设备 | |
WO2022032916A1 (zh) | 一种显示系统 | |
CN103984772A (zh) | 文本检索字幕库生成方法和装置、视频检索方法和装置 | |
WO2020135192A1 (zh) | 一种多字幕显示方法、智能终端及存储介质 | |
CN113965813B (zh) | 直播间内的视频播放方法、系统、设备及介质 | |
CN104102683A (zh) | 用于增强视频显示的上下文查询 | |
WO2022134555A1 (zh) | 视频处理方法及终端 | |
CN111625716A (zh) | 媒资推荐方法、服务器及显示设备 | |
CN111654732A (zh) | 一种广告播放方法及显示设备 | |
CN111914134A (zh) | 一种关联推荐方法、智能设备及服务设备 | |
CN111107283B (zh) | 一种信息显示方法、电子设备及存储介质 | |
CN114302221A (zh) | 一种虚拟现实设备及投屏媒资播放方法 | |
CN113468351A (zh) | 一种智能设备及图像处理方法 | |
CN111757187A (zh) | 多语言字幕显示方法、装置、终端设备及存储介质 | |
KR102612580B1 (ko) | 미디어 제공 서버, 트리거 영역을 통해 다른 컨텐츠로 전환하는 방법 및 컴퓨터 프로그램 | |
CN112053688B (zh) | 一种语音交互方法及交互设备、服务器 | |
CN114327033A (zh) | 一种虚拟现实设备及媒资播放方法 | |
CN112256232A (zh) | 显示设备与自然语言生成后处理方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20875068 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20875068 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20875068 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 29-09-22) |