WO2021068558A1 - 一种同声字幕翻译方法、智能电视及存储介质 - Google Patents

一种同声字幕翻译方法、智能电视及存储介质 Download PDF

Info

Publication number
WO2021068558A1
WO2021068558A1 PCT/CN2020/098074 CN2020098074W WO2021068558A1 WO 2021068558 A1 WO2021068558 A1 WO 2021068558A1 CN 2020098074 W CN2020098074 W CN 2020098074W WO 2021068558 A1 WO2021068558 A1 WO 2021068558A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
subtitle
subtitles
cloud server
simultaneous
Prior art date
Application number
PCT/CN2020/098074
Other languages
English (en)
French (fr)
Inventor
邓声扬
孙雷
Original Assignee
深圳创维-Rgb电子有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳创维-Rgb电子有限公司 filed Critical 深圳创维-Rgb电子有限公司
Publication of WO2021068558A1 publication Critical patent/WO2021068558A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42212Specific keyboard arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42224Touch pad or touch panel provided on the remote control
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • H04N21/4415Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration
    • H04N21/4856End-user interface for client configuration for language selection, e.g. for the menu or subtitles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration
    • H04N21/4858End-user interface for client configuration for modifying screen layout parameters, e.g. fonts, size of the windows

Definitions

  • the present disclosure relates to the technical field of smart TVs, such as a simultaneous subtitle translation method, smart TVs, and storage media.
  • TVs large-screen smart display devices
  • TV sets have unique advantages in watching movies and distance education in the current and future of intelligent development.
  • society the process of people's integration into globalization is accelerating.
  • language is the biggest obstacle to integrating into the international society and understanding international culture.
  • it will bring great obstacles to hope to learn and cultural exchanges.
  • the main purpose of the present disclosure is to provide a simultaneous subtitle translation method, a smart TV, and a storage medium, and aims to solve the problem of language barriers in the process of using high-quality audiovisual resources for users in the prior art.
  • the simultaneous subtitle translation method includes the following steps:
  • the cloud server is configured to translate the voice information into the target subtitles corresponding to the target language
  • the target subtitles are matched with the time axis of the playback resource in real time, and the target subtitles are displayed synchronously during the playback.
  • the simultaneous subtitle translation method wherein the receiving operation instruction of the user to select a target language, collecting voice information in the playback resource, and sending the target language and the voice information to a cloud server, specifically include:
  • the collection of voice information in the playback resource specifically includes:
  • the voice information in the video resource is identified and separated.
  • the simultaneous subtitle translation method wherein the receiving an operation instruction of a user to select a target language, afterwards, further includes:
  • the simultaneous subtitle translation method wherein the receiving an operation instruction of a user to select a target language, afterwards, further includes:
  • the subtitle style of the target subtitle selected by the user to be displayed on the display interface is received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size.
  • the display position of the target subtitle is above or below the original subtitle.
  • the subtitle color of the target subtitle is different from the subtitle color of the original subtitle.
  • the simultaneous subtitle translation method wherein the real-time matching of the target subtitles with the time axis of the playback resource, and synchronously displaying the target subtitles during the playback process specifically includes:
  • the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on the display interface during the playback of the playback resource.
  • the simultaneous subtitle translation method wherein the synchronously displaying the target subtitles during the playback process further includes:
  • the original subtitles and the target subtitles are displayed on the display interface to form a comparison effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language.
  • the cloud server is configured to translate the voice information into the target subtitle corresponding to the target language, which specifically includes:
  • the cloud server receives the voice information and the target language
  • the cloud server recognizes and translates the voice information according to the target language, and generates the target subtitles corresponding to the target language;
  • the cloud server returns the target subtitles to the smart TV.
  • the cloud server uses artificial intelligence to perform voice recognition and translation to generate the target subtitles desired by the user.
  • the translation by the cloud server is online real-time translation.
  • the present disclosure also provides a smart TV, wherein the smart TV includes: a memory, a processor, and a simultaneous subtitle translation program stored on the memory and running on the processor When the simultaneous subtitle translation program is executed by the processor, the steps of the simultaneous subtitle translation method described above are implemented.
  • the present disclosure also provides a simultaneous subtitle translation system, wherein the simultaneous subtitle translation system includes the smart TV as described above, and also includes a cloud that has established a communication connection with the smart TV.
  • the smart TV is used to receive the user's operation instructions to select the target language, collect the voice information in the playback resource and send it to the cloud server; also used to receive the target subtitles returned by the cloud server, and the target subtitles Perform real-time matching with the playback time axis of the playback resource, and display the target subtitles synchronously during playback;
  • the cloud server is used to translate the voice information into the target subtitles corresponding to the target language, and The target subtitles are sent to the smart TV.
  • the present disclosure also provides a storage medium, wherein the storage medium stores a simultaneous subtitle translation program, and the simultaneous subtitle translation program is executed by a processor to realize the simultaneous subtitle as described above.
  • the steps of the translation method are described above.
  • the present disclosure collects the voice information in the playback resource by receiving the operation instruction of the user to select the target language, and sends the target language and the voice information to the cloud server; receiving the target subtitles returned by the cloud server, the cloud server It is used to translate the voice information into the target subtitles corresponding to the target language; match the target subtitles with the time axis of the playback resource in real time, and display the target subtitles synchronously during the playback process.
  • the present disclosure collects the voice information in the playback resources, according to the language subtitles required by the user, sends the voice information to the cloud server for synchronous translation into the target subtitles corresponding to the target language selected by the user, and performs synchronous display during playback to satisfy the user
  • the demand for subtitles in multiple languages enables users to easily understand the content and information expressed by audio and video in different languages, and realize convenient viewing or learning.
  • Fig. 1 is a flowchart of a preferred embodiment of the simultaneous subtitle translation method of the present disclosure
  • step S10 is a flowchart of step S10 in the preferred embodiment of the simultaneous subtitle translation method of the present disclosure
  • step S30 is a flowchart of step S30 in the preferred embodiment of the simultaneous subtitle translation method of the present disclosure
  • FIG. 5 is a schematic diagram of the operating environment of the preferred embodiment of the disclosed smart TV.
  • the simultaneous subtitle translation method includes the following steps:
  • Step S10 Receive an operation instruction for the user to select a target language, collect voice information in the playback resource, and send the target language and the voice information to a cloud server.
  • FIG. 2 is a flowchart of step S10 in the simultaneous subtitle translation method provided by the present disclosure.
  • the step S10 includes:
  • S12. Collect the voice information in the playback resource that currently needs to be played, and send the compressed voice information and the target language to the cloud server.
  • the remote control The user can send an operation instruction to select the target language to the TV through the remote controller (such as a smart TV).
  • the remote controller such as a smart TV
  • the user can directly issue an operation instruction to select the target language through the touch screen touch menu
  • the TV or touch device system has pre-stored multiple languages, and you can add other languages that are not stored.
  • the subtitle style of the target subtitle selected by the user to be displayed on the display interface can also be received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size
  • the user selects the color of the final output target subtitle to be red (for example, the original subtitle is black, the two colors are different for comparison and viewing), the subtitle transparency can be selected according to the user’s needs, and the display position can be above or below the original subtitle.
  • the font size can also be selected according to user needs, and other subtitle parameters or related settings can also be selected.
  • the TV or touch device After the TV or touch device receives the operation instruction for the user to select the target language, it recognizes the operation instruction and executes the corresponding function, that is, collects the playback resource that needs to be played (or is currently playing) according to the target language
  • the voice information in the collected playback resource specifically includes: when the playback resource is an audio resource (for example, audio is played through a TV, although there is no changed picture, there will be a display interface, showing When the interface may be a certain playback interface that remains unchanged during the audio playback process, then the voice information in the audio resource can be directly obtained; when the playback resource is a video resource (including audio information and video information), Then it is necessary to identify and separate the voice information in the video resource. Because the target caption generation only needs to obtain the voice information.
  • an audio resource for example, audio is played through a TV, although there is no changed picture, there will be a display interface, showing When the interface may be a certain playback interface that remains unchanged during the audio playback process, then the voice information
  • Compress the acquired voice information (to facilitate data transmission), and finally send the compressed voice information and the target language to the cloud server, so that the cloud server can respond to the needs of the user
  • the voice data is processed.
  • Step S20 Receive the target subtitles returned by the cloud server, where the cloud server is used to translate the voice information into the target subtitles corresponding to the target language.
  • the cloud server mainly performs online real-time translation operations on the voice information according to the target language, for example, the The cloud server can perform voice recognition and translation through artificial intelligence (AI), and generate the target subtitles that the user wants (target subtitles mainly refer to the user's native language).
  • AI artificial intelligence
  • target subtitles mainly refer to the user's native language.
  • the cloud server After the cloud server receives the voice information and the target language, the cloud server according to the target language (for example, the original audio and video resources are French, and the user needs Chinese, then the target language is Chinese, the original The subtitles are French) recognize and translate the voice information, and generate the target subtitles corresponding to the target language (for example, generate Chinese subtitles), and the cloud server transmits the target subtitles back to the TV (for example, a smart TV). ) Or touch the device.
  • the target language for example, the original audio and video resources are French, and the user needs Chinese
  • the target language is Chinese
  • the original The subtitles are French
  • the cloud server transmits the target subtitles back to the TV (for example, a smart TV). ) Or touch the device.
  • the translation of the cloud server is online real-time translation, which does not affect the user's viewing.
  • the TV or touch device can also choose to broadcast and translate according to the user's network environment, or translate first and then broadcast. Provide users with a consistent and transparent viewing experience.
  • Step S30 Match the target subtitles with the time axis of the playback resource in real time, and display the target subtitles synchronously during the playback process.
  • FIG. 3 is a flowchart of step S30 in the simultaneous subtitle translation method provided by the present disclosure.
  • the step S30 includes:
  • the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on a display interface during the playback of the playback resource.
  • the cloud server when the cloud server recognizes and translates the voice information according to the target language, and generates the target subtitles corresponding to the target language, the cloud server transmits the target subtitles back to the TV or After touching the device, the TV or touch device matches the target subtitles and pictures according to the target subtitles, and outputs synchronously according to the time axis, that is, real-time matching of the target subtitles and the time axis of the playback resource, so that the picture, voice and target subtitles Synchronous matching; when the time axis matching is completed, the target subtitles are displayed synchronously on the display interface (i.e. display screen) during audio or video playback. For example, during audio or video playback, the original subtitles are compared and displayed on the display screen.
  • the display interface i.e. display screen
  • the subtitles and the target subtitles are used to form the contrast effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language, which not only preserves the original sound but also allows viewing Understand the meaning to be expressed. For example, if the user wants to learn the language represented by the original subtitles, then the comparison display can assist the user in learning other languages.
  • the simultaneous subtitle translation method of the present disclosure decodes the decodable audio and video resources and samples the key information.
  • the key sampling information is transferred to the cloud server, which is translated by the cloud server and then returned to the smart
  • subtitles are used to cooperate with the original language output and presented to the user, so that the user can easily understand the content information expressed by audio and video in different languages, realize convenient communication or learning, can give the user a real context, and the user can immediately understand
  • the information to be expressed by audio and video has high processing efficiency and a wide range of applications.
  • Step S0 The user selects the target language according to the subtitles that need to be output finally (the terminal device takes a smart TV as an example, and the smart TV receives the user's operation instructions);
  • Step S1 The smart TV collects voice information in playback resources (such as audio or video);
  • Step S2 the smart TV sends the target language and the voice information to the cloud server;
  • Step S3 The cloud server is used to translate the voice information into the target subtitles corresponding to the target language, and return the target subtitles to the smart TV;
  • Step S4 The smart TV receives the target subtitles sent by the cloud server, and matches the target subtitles with the time axis of the playback resource in real time, so that the screen, voice, and subtitles are output simultaneously (step S4 and step S1 can be Synchronize through the background, without affecting the user's viewing);
  • Step S5 After the time axis matching of the target subtitle and the playback resource is completed, the target subtitle is synchronously displayed on the display interface of the smart TV during the playback of the playback resource.
  • this function can obtain more potential customers, improve the use time of large-screens and customer viscosity, and obtain more commercial value.
  • the present disclosure also provides a smart TV correspondingly.
  • the smart TV includes a processor 10, a memory 20 and a display 30.
  • FIG. 5 only shows part of the components of the smart TV, but it should be understood that it is not required to implement all the shown components, and more or fewer components may be implemented instead.
  • the memory 20 may be an internal storage unit of the smart TV, such as a hard disk or memory of the smart TV.
  • the memory 20 may also be an external storage device of the smart TV, for example, a plug-in hard disk equipped on the smart TV, a smart memory card (Smart Media Card, SMC), and a secure digital (Secure Digital). Digital, SD) card, flash card (Flash Card), etc.
  • the memory 20 may also include both an internal storage unit of the smart TV and an external storage device.
  • the memory 20 is used to store application software and various data installed on the smart TV, such as program code for the smart TV installed.
  • the memory 20 can also be used to temporarily store data that has been output or will be output.
  • a simultaneous subtitle translation program 40 is stored on the memory 20, and the simultaneous subtitle translation program 40 can be executed by the processor 10, so as to realize the simultaneous subtitle translation method in the present application.
  • the processor 10 may be a central processing unit (CPU), microprocessor or other data processing chip in some embodiments, and is used to run the program code or process data stored in the memory 20, for example Perform the simultaneous subtitle translation method and so on.
  • CPU central processing unit
  • microprocessor or other data processing chip in some embodiments, and is used to run the program code or process data stored in the memory 20, for example Perform the simultaneous subtitle translation method and so on.
  • the display 30 may be an LED display, a liquid crystal display, a touch liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light-emitting diode) touch device, etc.
  • the display 30 is used for displaying information on the smart TV and for displaying a visualized user interface.
  • the components 10-30 of the smart TV communicate with each other through a system bus.
  • the cloud server is configured to translate the voice information into the target subtitles corresponding to the target language
  • the target subtitles are matched with the time axis of the playback resource in real time, and the target subtitles are displayed synchronously during the playback process.
  • the receiving the operation instruction of the user to select the target language, collecting the voice information in the playback resource, and sending the target language and the voice information to the cloud server specifically includes:
  • the voice information in the collection and playback resources specifically includes:
  • the voice information in the video resource is identified and separated.
  • the key sampling information is delivered to the cloud server according to the user's needs for the language.
  • the subtitle style of the target subtitle selected by the user to be displayed on the display interface is received, and the parameters of the subtitle style include subtitle color, subtitle transparency, display position, and font size.
  • the display position of the target subtitle is above or below the original subtitle.
  • the subtitle color of the target subtitle is different from the subtitle color of the original subtitle.
  • the real-time matching of the target subtitles with the time axis of the playback resource, and synchronously displaying the target subtitles during the playback process specifically includes:
  • the target subtitles are matched with the time axis of the playback resource, the target subtitles are synchronously displayed on the display interface during the playback of the playback resource.
  • the synchronously displaying the target subtitles during the playback process further includes:
  • the original subtitles and the target subtitles are displayed on the display interface to form a comparison effect of the original subtitles and the target subtitles for auxiliary learning of the original subtitle language.
  • the present disclosure also provides a storage medium, wherein the storage medium stores a simultaneous subtitle translation program, and when the simultaneous subtitle translation program is executed by a processor, the steps of the simultaneous subtitle translation method described above are realized.
  • the present disclosure provides a simultaneous subtitle translation method, a smart TV, and a storage medium.
  • the method includes: receiving an operation instruction for a user to select a target language, collecting voice information in a playback resource, and combining the target language with The voice information is sent to the cloud server; the target subtitles returned by the cloud server are received, and the cloud server is used to translate the voice information into the target subtitles corresponding to the target language; and the target subtitles are combined with the target subtitles.
  • the time axis of the playback resource is matched in real time, and the target subtitles are displayed synchronously during the playback process.
  • the present disclosure collects the voice information in the playback resources, according to the language subtitles required by the user, sends the voice information to the cloud server for synchronous translation into the target subtitles corresponding to the target language selected by the user, and performs synchronous display during playback to satisfy the user
  • the demand for subtitles in multiple languages enables users to easily understand the content and information expressed by audio and video in different languages, and realize convenient viewing or learning.
  • the processes in the methods of the above-mentioned embodiments can be implemented by instructing relevant hardware (such as a processor, a controller, etc.) through a computer program, and the program can be stored in a computer program.
  • the program may include the processes of the foregoing method embodiments when executed.
  • the storage medium mentioned may be a memory, a magnetic disk, an optical disk, and the like.
  • the embodiment of the present disclosure provides a simultaneous subtitle translation method, smart TV, and storage medium.
  • the voice information is sent to the cloud server for simultaneous translation into user selection
  • the target subtitles corresponding to the target language of the target language are displayed synchronously during playback to meet the needs of users for subtitles in multiple languages, so that users can easily understand the content information expressed by audio and video in different languages, and realize convenient viewing or learning.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Electrically Operated Instructional Devices (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

本公开涉及一种同声字幕翻译方法、智能电视及存储介质,通过采集播放资源中的语音信息,根据用户需要的语言字幕,将所述语音信息发送至云端服务器同步翻译成用户选择的目标语言对应的目标字幕,并在播放时进行同步显示,使用户能轻易理解不同语种音视频表达的内容信息,实现便捷观看或学习。

Description

一种同声字幕翻译方法、智能电视及存储介质 技术领域
本公开涉及智能电视技术领域,例如涉及一种同声字幕翻译方法、智能电视及存储介质。
背景技术
随着社会国际化程度越来越高,用户对外交流场景不断增加,使得人们对语言学习和新闻理解的需求越来越多,大屏智能显示设备(电视等)在这种交流中拥有便捷的使用环境。电视机作为大屏显示设备,在智能化发展的当下和未来,在观看影视和远程教育上有着独特的优势。随着社会的发展,人们融入全球化的进程不断加快。对普通民众而言,语言是一个融入国际社会,理解国际文化最大的障碍,尤其在一些独特优质的媒体资料上,由于用户外国语水平不高,会给希望学习和文化交流带来极大障碍。
目前人工智能技术不断发展,语音实时传译、全文翻译等技术逐渐成熟,识别准确率不断提升,以及网络技术向着大速率,低延时的方向发展。但是由于在音视频内容领域存在大量的不同语种,会给非母语的人们观看时造成理解上的困难,导致用户无法跨域语种的障碍,在使用非母语的音视频资源存在很大的障碍。
因此,现有技术还有待于改进和发展。
发明内容
本公开的主要目的在于提供一种同声字幕翻译方法、智能电视及存储介质,旨在解决现有技术中用户在使用优质影音资源的过程中存在语言障碍的问题。
为实现上述目的,本公开提供一种同声字幕翻译方法,所述同声字幕翻译方法包括如下步骤:
接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器;
接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;
将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步 显示所述目标字幕。
可选地,所述的同声字幕翻译方法,其中,所述接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器,具体包括:
接收用户通过遥控器按键或者触摸屏触摸菜单选择目标语言的操作指令,所述目标语言包括多种预先存储的语言;
采集当前需要播放的所述播放资源中的所述语音信息,将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器。
可选地,所述的同声字幕翻译方法,其中,所述采集播放资源中的语音信息,具体包括:
当所述播放资源为音频资源时,直接获取所述音频资源中的语音信息;
当所述播放资源为视频资源时,识别并分离出所述视频资源中的语音信息。
可选地,所述的同声字幕翻译方法,其中,所述接收用户选择目标语言的操作指令,之后还包括:
通过对可解码的音视频资源进行解码和关键信息采样,根据用户对语种的需求,将关键采样信息传递到云端服务器。可选地,所述的同声字幕翻译方法,其中,所述接收用户选择目标语言的操作指令,之后还包括:
接收用户选择在显示界面上显示所述目标字幕的字幕样式,所述字幕样式的参数包括字幕颜色、字幕透明度、显示位置以及字体大小。
可选地,所述的同声字幕翻译方法,其中,所述目标字幕的显示位置在原始字幕的上方或者下方。
可选地,所述的同声字幕翻译方法,其中,所述目标字幕的字幕颜色与所述原始字幕的字幕颜色不同。
可选地,所述的同声字幕翻译方法,其中,所述将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕,具体包括:
获取所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,所述匹配包括画面、语音和字幕的同步匹配;
当所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在显示界面上同步显示所述目标字幕。
可选地,所述的同声字幕翻译方法,其中,所述在播放过程中同步显示所述目标字幕,还包括:
在所述播放资源播放过程中,在显示界面上对比显示原始字幕和所述目标字幕,用于形成所述原始字幕和所述目标字幕的对比效果以进行原始字幕语言的辅助学习。
可选地,所述的同声字幕翻译方法,其中,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,具体包括:
所述云端服务器接收所述语音信息和所述目标语言;
所述云端服务器根据所述目标语言对所述语音信息进行识别和翻译,并生成所述目标语言对应的所述目标字幕;
所述云端服务器将所述目标字幕回传至智能电视。
可选地,所述的同声字幕翻译方法,其中,所述云端服务器通过人工智能进行语音识别和翻译,生成用户想要的目标字幕。
可选地,所述的同声字幕翻译方法,其中,所述云端服务器的翻译为在线实时翻译。
此外,为实现上述目的,本公开还提供一种智能电视,其中,所述智能电视包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的同声字幕翻译程序,所述同声字幕翻译程序被所述处理器执行时实现如上所述的同声字幕翻译方法的步骤。
此外,为实现上述目的,本公开还提供一种同声字幕翻译系统,其中,所述同声字幕翻译系统包括如上所述的智能电视,还包括一与所述智能电视已建立通信连接的云端服务器;所述智能电视用于接收用户选择目标语言的操作指令,采集播放资源中的语音信息并发送至云端服务器;还用于接收所述云端服务器回传的目标字幕,并将所述目标字幕与所述播放资源的播放时间轴进行实时匹配,在播放过程中同步显示所述目标字幕;所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,并将所述目标字幕发送至所述智能电视。
此外,为实现上述目的,本公开还提供一种存储介质,其中,所述存储介质存储有同声字幕翻译程序,所述同声字幕翻译程序被处理器执行时实现如上所述的同声字幕翻译方法的步骤。
本公开通过接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器;接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。本公开通过采集播放资源中的语音信息,根据用户需要的语言字幕,将所述语音信息发送至云端服务器同步翻译成用户选择的目标语言对应的目标字幕,并在播放时进行同步显示,满足用户对多种语言字幕的需求,使用户能轻易理解不同语种音视频表达的内容信息,实现便捷观看或学习。
附图说明
图1是本公开同声字幕翻译方法的较佳实施例的流程图;
图2是本公开同声字幕翻译方法的较佳实施例中步骤S10的流程图;
图3是本公开同声字幕翻译方法的较佳实施例中步骤S30的流程图;
图4是本公开同声字幕翻译方法的较佳实施例中进行目标字幕同步翻译的整个过程的流程图;
图5为本公开智能电视的较佳实施例的运行环境示意图。
具体实施方式
为使本公开的目的、技术方案及优点更加清楚、明确,以下参照附图并举实施例对本公开进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本公开,并不用于限定本公开。
实施例1
本公开较佳实施例所述的同声字幕翻译方法,如图1所示,所述同声字幕翻译方法包括以下步骤:
步骤S10、接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器。
具体的过程请参阅图2,其为本公开提供的同声字幕翻译方法中步骤S10的流程图。
如图2所示,所述步骤S10包括:
S11、接收用户通过遥控器按键或者触摸屏触摸菜单选择目标语言的操作指 令,所述目标语言包括多种预先存储的语言;
S12、采集当前需要播放的所述播放资源中的所述语音信息,将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器。
具体地,首先要知道用户想要看到的字母语言是哪一种,那么就要接收用户选择目标语言的操作指令,而用户发出操作指令的方式可以多样化,本公开中,对于配套有遥控器的电视(例如智能电视),用户可以通过遥控器发出选择目标语言的操作指令到电视,对于具有触摸功能(带有触摸屏)的设备,用户可以直接通过触摸屏触摸菜单发出选择目标语言的操作指令,电视或者触摸设备的系统中已经预先存储了多种语言,对于其他未存储的也可以进行添加。
进一步地,在接收用户选择目标语言的操作指令之后,还可以接收用户选择在显示界面上显示所述目标字幕的字幕样式,所述字幕样式的参数包括字幕颜色、字幕透明度、显示位置以及字体大小,例如用户选择最后输出的目标字幕的颜色为红色(例如原始字幕为黑色,两者颜色不同便于对比观看),字幕透明度可根据用户的需要进行选择,显示位置可以在原始字幕的上方或者下方,字体大小也可以根据用户的需求进行选择,还可以选择其他的字幕参数或者相关设置。
当电视或者触摸设备接收到用户选择目标语言的操作指令之后,识别所述操作指令并执行对应的功能,也就是根据所述目标语言采集当前需要播放(或者当前正在播放的)的所述播放资源中的所述语音信息,其中,所述采集播放资源中的语音信息具体包括:当所述播放资源为音频资源(例如通过电视播放音频,虽然没有变动的画面,但是也会有显示界面,显示界面可能是某一个播放界面在音频播放过程中一直不变)时,那么直接获取所述音频资源中的语音信息即可;当所述播放资源为视频资源(包括音频信息和视频信息)时,那么就需要识别并分离出所述视频资源中的语音信息。因为进行目标字幕生成只需要得到语音信息即可。
将获取的所述语音信息进行压缩处理(便于数据的传输),最后将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器,便于所述云端服务器根据用户的需求,对所述语音数据进行处理。
步骤S20、接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕。
具体地,当电视或者触摸设备将所述目标语言和所述语音信息发送至所述云端服务器之后,所述云端服务器主要根据所述目标语言对所述语音信息进行在线实时翻译操作,例如所述云端服务器可以通过人工智能(AI)进行语音识别和翻译,生成用户想要的目标字幕(目标字幕主要是指用户的母语),同时,对于云端服务器AI技术层面,通过大量的音视频资源训练和客户各种需求的增加,可推动机器语言处理的快速发展。
其中,所述云端服务器接收所述语音信息和所述目标语言之后,所述云端服务器根据所述目标语言(例如原始音视频为法语资源,而用户需要的是中文,那么目标语言就是中文,原始字幕就是法文)对所述语音信息进行识别和翻译,并生成所述目标语言对应的所述目标字幕(例如生成中文字幕),所述云端服务器将所述目标字幕回传至电视(例如智能电视)或者触摸设备。
进一步地,所述云端服务器的翻译是在线实时翻译,并不影响用户的观看,同时,电视或者触摸设备也可根据用户的网络环境情况,根据需求选择边播边译,或先译后播,给用户提供连贯透明的观看体验。
步骤S30、将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。
具体过程请参阅图3,其为本公开提供的同声字幕翻译方法中步骤S30的流程图。
如图3所示,所述步骤S30包括:
S31、获取所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,所述匹配包括画面、语音和字幕的同步匹配;
S32、当所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在显示界面上同步显示所述目标字幕。
具体地,当所述云端服务器根据所述目标语言对所述语音信息进行识别和翻译,并生成所述目标语言对应的所述目标字幕,所述云端服务器将所述目标字幕回传至电视或者触摸设备后,电视或者触摸设备根据所述目标字幕匹配目标字幕和画面,根据时间轴进行同步输出,即将所述目标字幕与所述播放资源的时间轴进行实时匹配,使得画面、语音和目标字幕进行同步匹配;当完成时间轴匹配后,在音频或者视频播放过程中在显示界面(即显示屏)上同步显示所述目标字幕, 例如在音频或者视频播放过程中,在显示屏上对比显示原始字幕和所述目标字幕(例如原始字幕显示在目标字幕的上方或者下方),用于形成所述原始字幕和所述目标字幕的对比效果以进行原始字幕语言的辅助学习,既保留原音又能看懂所要表达的意思,例如用户想要学习原始字幕所代表的语言,那么进行对比显示就可以辅助用户进行其他语种的学习。
本公开的同声字幕翻译方法,通过对可解码的音视频资源进行解码和关键信息采样,根据用户对语种的需求,将关键采样信息传递到云端服务器,由云端服务器进行翻译后,返回到智能显示设备中,以字幕的方式配合原语输出,呈现给用户,使用户能轻易理解不同语种音视频表达的内容信息,实现便捷交流或学习,能给用户真实的语境,同时用户能立即明白音视频所要表达的信息,处理效率高,应用范围广。
进一步地,如图4所示,本公开中进行目标字幕同步翻译的整个过程如下:
步骤S0、用户根据最终需要输出的字幕选择目标语言(终端设备以智能电视为例,智能电视接收用户的操作指令);
步骤S1、智能电视采集播放资源(例如音频或者视频)中的语音信息;
步骤S2、,智能电视将所述目标语言和所述语音信息发送至云端服务器;
步骤S3、云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,并将所述目标字幕回传至智能电视;
步骤S4、智能电视接收所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,使得画面、语音和字幕的同步输出(步骤S4和步骤S1可以通过后台同步进行,不影响用户的观看);
步骤S5、所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在智能电视的显示界面上同步显示所述目标字幕。
本公开带来了如下技术好处:
(1)增加了新功能无需额外增加成本,用户在已拥有智能显示终端的情况下,无需再购买其他硬件设备,节省了硬件费用。
(2)用户的使用场景更加广泛,用户可用此功能理解不同语种的新闻或教程等音视频资源所要表达的信息,由于可以同步输出母语与外语的对比字幕,用户也可以做为学习各种外语的辅助工具,在实际场景的语境中,获得水平的提升。
(3)对于大屏智能显示设备的厂商,可通过此功能获得更多的潜在客户,提升大屏的使用时间和客户粘度,获得更多的商业价值。
(4)对于云端服务器的AI技术层面,通过大量的音视频资源训练和客户各种需求的增加,可推动机器语言处理的快速发展。
实施例2
进一步地,如图5所示,基于上述同声字幕翻译方法,本公开还相应提供了一种智能电视,所述智能电视包括处理器10、存储器20及显示器30。图5仅示出了智能电视的部分组件,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。
所述存储器20在一些实施例中可以是所述智能电视的内部存储单元,例如智能电视的硬盘或内存。所述存储器20在另一些实施例中也可以是所述智能电视的外部存储设备,例如所述智能电视上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器20还可以既包括所述智能电视的内部存储单元也包括外部存储设备。所述存储器20用于存储安装于所述智能电视的应用软件及各类数据,例如所述安装智能电视的程序代码等。所述存储器20还可以用于暂时地存储已经输出或者将要输出的数据。在一实施例中,存储器20上存储有同声字幕翻译程序40,该同声字幕翻译程序40可被处理器10所执行,从而实现本申请中同声字幕翻译方法。
所述处理器10在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行所述存储器20中存储的程序代码或处理数据,例如执行所述同声字幕翻译方法等。
所述显示器30在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。所述显示器30用于显示在所述智能电视的信息以及用于显示可视化的用户界面。所述智能电视的部件10-30通过系统总线相互通信。
在一实施例中,当处理器10执行所述存储器20中同声字幕翻译程序40时实现以下步骤:
接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目 标语言和所述语音信息发送至云端服务器;
接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;
将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。
所述接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器,具体包括:
接收用户通过遥控器按键或者触摸屏触摸菜单选择目标语言的操作指令,所述目标语言包括多种预先存储的语言;
采集当前需要播放的所述播放资源中的所述语音信息,将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器。
所述采集播放资源中的语音信息,具体包括:
当所述播放资源为音频资源时,直接获取所述音频资源中的语音信息;
当所述播放资源为视频资源时,识别并分离出所述视频资源中的语音信息。
所述接收用户选择目标语言的操作指令,之后还包括:
通过对可解码的音视频资源进行解码和关键信息采样,根据用户对语种的需求,将关键采样信息传递到云端服务器。
所述接收用户选择目标语言的操作指令,之后还包括:
接收用户选择在显示界面上显示所述目标字幕的字幕样式,所述字幕样式的参数包括字幕颜色、字幕透明度、显示位置以及字体大小。
所述目标字幕的显示位置在原始字幕的上方或者下方。
所述目标字幕的字幕颜色与所述原始字幕的字幕颜色不同。
所述将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕,具体包括:
获取所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,所述匹配包括画面、语音和字幕的同步匹配;
当所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在显示界面上同步显示所述目标字幕。
所述在播放过程中同步显示所述目标字幕,还包括:
在所述播放资源播放过程中,在显示界面上对比显示原始字幕和所述目标字幕,用于形成所述原始字幕和所述目标字幕的对比效果以进行原始字幕语言的辅助学习。
实施例3
本公开还提供一种存储介质,其中,所述存储介质存储有同声字幕翻译程序,所述同声字幕翻译程序被处理器执行时实现如上所述的同声字幕翻译方法的步骤。
综上所述,本公开提供一种同声字幕翻译方法、智能电视及存储介质,所述方法包括:接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器;接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。本公开通过采集播放资源中的语音信息,根据用户需要的语言字幕,将所述语音信息发送至云端服务器同步翻译成用户选择的目标语言对应的目标字幕,并在播放时进行同步显示,满足用户对多种语言字幕的需求,使用户能轻易理解不同语种音视频表达的内容信息,实现便捷观看或学习。
当然,本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关硬件(如处理器,控制器等)来完成,所述的程序可存储于一计算机可读取的存储介质中,所述程序在执行时可包括如上述各方法实施例的流程。其中所述的存储介质可为存储器、磁碟、光盘等。
应当理解的是,本公开的应用不限于上述的举例,对本领域普通技术人员来说,可以根据上述说明加以改进或变换,所有这些改进和变换都应属于本公开所附权利要求的保护范围。
工业实用性
本公开实施例提供的一种同声字幕翻译方法、智能电视及存储介质,通过采集播放资源中的语音信息,根据用户需要的语言字幕,将所述语音信息发送至云端服务器同步翻译成用户选择的目标语言对应的目标字幕,并在播放时进行同步显示,满足用户对多种语言字幕的需求,使用户能轻易理解不同语种音视频表达的内容信息,实现便捷观看或学习。

Claims (15)

  1. 一种同声字幕翻译方法,其特征在于,所述同声字幕翻译方法包括:
    接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器;
    接收所述云端服务器回传的目标字幕,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕;
    将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕。
  2. 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,采集播放资源中的语音信息,将所述目标语言和所述语音信息发送至云端服务器,具体包括:
    接收用户通过遥控器按键或者触摸屏触摸菜单选择目标语言的操作指令,所述目标语言包括多种预先存储的语言;
    采集当前需要播放的所述播放资源中的所述语音信息,将压缩处理后的所述语音信息和所述目标语言发送至所述云端服务器。
  3. 根据权利要求2所述的同声字幕翻译方法,其特征在于,所述采集播放资源中的语音信息,具体包括:
    当所述播放资源为音频资源时,直接获取所述音频资源中的语音信息;
    当所述播放资源为视频资源时,识别并分离出所述视频资源中的语音信息。
  4. 根据权利要求1或2所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,之后还包括:
    通过对可解码的音视频资源进行解码和关键信息采样,根据用户对语种的需求,将关键采样信息传递到云端服务器。
  5. 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述接收用户选择目标语言的操作指令,之后还包括:
    接收用户选择在显示界面上显示所述目标字幕的字幕样式,所述字幕样式的参数包括字幕颜色、字幕透明度、显示位置以及字体大小。
  6. 根据权利要求5所述的同声字幕翻译方法,其特征在于,所述目标字幕的显示位置在原始字幕的上方或者下方。
  7. 根据权利要求6所述的同声字幕翻译方法,其特征在于,所述目标字幕的字幕颜色与所述原始字幕的字幕颜色不同。
  8. 根据权利要求1所述的同声字幕翻译方法,其特征在于,所述将所述目标字幕与所述播放资源的时间轴进行实时匹配,在播放过程中同步显示所述目标字幕,具体包括:
    获取所述云端服务器发送的所述目标字幕,将所述目标字幕与所述播放资源的时间轴进行实时匹配,所述匹配包括画面、语音和字幕的同步匹配;
    当所述目标字幕与所述播放资源的时间轴匹配完成后,在所述播放资源播放过程中在显示界面上同步显示所述目标字幕。
  9. 根据权利要求8所述的同声字幕翻译方法,其特征在于,所述在播放过程中同步显示所述目标字幕,还包括:
    在所述播放资源播放过程中,在显示界面上对比显示原始字幕和所述目标字幕,用于形成所述原始字幕和所述目标字幕的对比效果以进行原始字幕语言的辅助学习。
  10. 根据权利要求1或2所述的同声字幕翻译方法,其特征在于,所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,具体包括:
    所述云端服务器接收所述语音信息和所述目标语言;
    所述云端服务器根据所述目标语言对所述语音信息进行识别和翻译,并生成所述目标语言对应的所述目标字幕;
    所述云端服务器将所述目标字幕回传至智能电视。
  11. 根据权利要求10所述的同声字幕翻译方法,其特征在于,所述云端服务器通过人工智能进行语音识别和翻译,生成用户想要的目标字幕。
  12. 根据权利要求10所述的同声字幕翻译方法,其特征在于,所述云端服务器的翻译为在线实时翻译。
  13. 一种智能电视,其特征在于,所述智能电视包括:存储器、处理器及存储在所 述存储器上并可在所述处理器上运行的同声字幕翻译程序,所述同声字幕翻译程序被所述处理器执行时实现如权利要求1-9任一项所述的同声字幕翻译方法的步骤。
  14. 一种同声字幕翻译系统,其特征在于,所述同声字幕翻译系统包括如权利要求13所述的智能电视,还包括一与所述智能电视已建立通信连接的云端服务器;
    所述智能电视用于接收用户选择目标语言的操作指令,采集播放资源中的语音信息并发送至云端服务器;还用于接收所述云端服务器回传的目标字幕,并将所述目标字幕与所述播放资源的播放时间轴进行实时匹配,在播放过程中同步显示所述目标字幕;
    所述云端服务器用于将所述语音信息翻译成所述目标语言对应的所述目标字幕,并将所述目标字幕发送至所述智能电视。
  15. 一种存储介质,其特征在于,所述存储介质存储有同声字幕翻译程序,所述同声字幕翻译程序被处理器执行时实现如权利要求1-9任一项所述的同声字幕翻译方法的步骤。
PCT/CN2020/098074 2019-10-08 2020-06-24 一种同声字幕翻译方法、智能电视及存储介质 WO2021068558A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910950069.3 2019-10-08
CN201910950069.3A CN110769265A (zh) 2019-10-08 2019-10-08 一种同声字幕翻译方法、智能电视及存储介质

Publications (1)

Publication Number Publication Date
WO2021068558A1 true WO2021068558A1 (zh) 2021-04-15

Family

ID=69330967

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/098074 WO2021068558A1 (zh) 2019-10-08 2020-06-24 一种同声字幕翻译方法、智能电视及存储介质

Country Status (2)

Country Link
CN (1) CN110769265A (zh)
WO (1) WO2021068558A1 (zh)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110769265A (zh) * 2019-10-08 2020-02-07 深圳创维-Rgb电子有限公司 一种同声字幕翻译方法、智能电视及存储介质
CN113473238B (zh) * 2020-04-29 2022-10-18 海信集团有限公司 一种智能设备及视频通话时的同声翻译方法
CN111683266A (zh) * 2020-05-06 2020-09-18 厦门盈趣科技股份有限公司 一种视频同声翻译配置字幕方法及终端
CN112055261A (zh) * 2020-07-14 2020-12-08 北京百度网讯科技有限公司 字幕展示方法、装置、电子设备和存储介质
CN113301357B (zh) * 2020-07-27 2022-11-29 阿里巴巴集团控股有限公司 直播方法、装置及电子设备
CN112118309A (zh) * 2020-09-16 2020-12-22 三星电子(中国)研发中心 音频翻译方法和系统
CN112584209A (zh) * 2020-12-04 2021-03-30 深圳创维-Rgb电子有限公司 一种显示方法、装置、存储介质及智能电视
CN112601101B (zh) * 2020-12-11 2023-02-24 北京有竹居网络技术有限公司 一种字幕显示方法、装置、电子设备及存储介质
CN112601102A (zh) * 2020-12-11 2021-04-02 北京有竹居网络技术有限公司 同声传译字幕的确定方法、装置、电子设备及存储介质
CN112995749B (zh) * 2021-02-07 2023-05-26 北京字节跳动网络技术有限公司 视频字幕的处理方法、装置、设备和存储介质
CN113709509A (zh) * 2021-08-05 2021-11-26 中移(杭州)信息技术有限公司 音视频数据的传输方法、装置以及存储介质
CN113709579B (zh) * 2021-08-05 2023-07-21 中移(杭州)信息技术有限公司 音视频数据的传输方法、装置以及存储介质
CN113709558A (zh) * 2021-10-09 2021-11-26 立讯电子科技(昆山)有限公司 多媒体处理方法及多媒体交互系统
CN113992926B (zh) * 2021-10-19 2023-09-12 北京有竹居网络技术有限公司 界面显示方法、装置、电子设备和存储介质
CN115051991A (zh) * 2022-07-08 2022-09-13 北京有竹居网络技术有限公司 音频处理方法、装置、存储介质与电子设备

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170092274A1 (en) * 2015-09-24 2017-03-30 Otojoy LLC Captioning system and/or method
CN106792097A (zh) * 2016-12-27 2017-05-31 深圳Tcl数字技术有限公司 音频信号字幕匹配方法及装置
CN107277646A (zh) * 2017-08-08 2017-10-20 四川长虹电器股份有限公司 一种音视频资源的字幕配置系统
CN108319590A (zh) * 2018-01-25 2018-07-24 芜湖应天光电科技有限责任公司 一种基于云服务的自适应翻译机
CN207720194U (zh) * 2017-01-05 2018-08-10 陈伯妤 一种同传手机
CN110769265A (zh) * 2019-10-08 2020-02-07 深圳创维-Rgb电子有限公司 一种同声字幕翻译方法、智能电视及存储介质

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7711543B2 (en) * 2006-04-14 2010-05-04 At&T Intellectual Property Ii, Lp On-demand language translation for television programs
CN101179669B (zh) * 2006-11-08 2011-03-16 中兴通讯股份有限公司 一种会议电视终端字幕生成叠加方法
CN102316277B (zh) * 2011-09-06 2014-10-29 Tcl新技术(惠州)有限公司 字幕系统及其实现方法
CN103079039B (zh) * 2011-10-25 2016-08-31 康佳集团股份有限公司 一种飞播方法和系统
CN103327397A (zh) * 2012-03-22 2013-09-25 联想(北京)有限公司 一种媒体文件的字幕同步显示方法及系统
CN103905743A (zh) * 2012-12-30 2014-07-02 上海能感物联网有限公司 汉语自动实时标注汉语字幕音像录播方法
CN103561217A (zh) * 2013-10-14 2014-02-05 深圳创维数字技术股份有限公司 一种生成字幕的方法及终端
CN105681890A (zh) * 2016-01-26 2016-06-15 广东欧珀移动通信有限公司 一种视频播放终端的字幕显示方法及装置
CN105828101B (zh) * 2016-03-29 2019-03-08 北京小米移动软件有限公司 生成字幕文件的方法及装置
CN105848004A (zh) * 2016-05-16 2016-08-10 乐视控股(北京)有限公司 字幕播放方法、字幕播放装置
CN106340291A (zh) * 2016-09-27 2017-01-18 广东小天才科技有限公司 一种双语字幕制作方法及系统
CN106340294A (zh) * 2016-09-29 2017-01-18 安徽声讯信息技术有限公司 基于同步翻译的新闻直播字幕在线制作系统
CN108572950A (zh) * 2017-03-08 2018-09-25 捷音特科技股份有限公司 耳麦翻译系统
CN207264383U (zh) * 2017-08-23 2018-04-20 成都铅笔科技有限公司 一种实时语音翻译的外语翻译装置
CN207530952U (zh) * 2017-12-13 2018-06-22 南京洛菲特数码科技有限公司 一种多屏图像拼接处理器
CN108197123A (zh) * 2018-02-07 2018-06-22 云南衍那科技有限公司 一种基于智能手表的云翻译系统和方法
CN108833810A (zh) * 2018-06-21 2018-11-16 珠海金山网络游戏科技有限公司 一种三维虚拟偶像直播中实时生成字幕的方法及装置
CN108962220B (zh) * 2018-07-26 2021-03-09 北京小米移动软件有限公司 多媒体文件播放场景下的文本显示方法及装置
TWM574267U (zh) * 2018-08-16 2019-02-11 張漢龍 即時語音自動同步轉譯字幕直播系統

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170092274A1 (en) * 2015-09-24 2017-03-30 Otojoy LLC Captioning system and/or method
CN106792097A (zh) * 2016-12-27 2017-05-31 深圳Tcl数字技术有限公司 音频信号字幕匹配方法及装置
CN207720194U (zh) * 2017-01-05 2018-08-10 陈伯妤 一种同传手机
CN107277646A (zh) * 2017-08-08 2017-10-20 四川长虹电器股份有限公司 一种音视频资源的字幕配置系统
CN108319590A (zh) * 2018-01-25 2018-07-24 芜湖应天光电科技有限责任公司 一种基于云服务的自适应翻译机
CN110769265A (zh) * 2019-10-08 2020-02-07 深圳创维-Rgb电子有限公司 一种同声字幕翻译方法、智能电视及存储介质

Also Published As

Publication number Publication date
CN110769265A (zh) 2020-02-07

Similar Documents

Publication Publication Date Title
WO2021068558A1 (zh) 一种同声字幕翻译方法、智能电视及存储介质
US20210321157A1 (en) Special effect processing method and apparatus for live broadcasting, and server
KR102246305B1 (ko) 증강 미디어 서비스 제공 방법, 장치 및 시스템
WO2017181599A1 (zh) 一种显示视频的方法及装置
CN112163086B (zh) 多意图的识别方法、显示设备
CN114286173A (zh) 一种显示设备及音画参数调节方法
WO2017008652A1 (zh) 信息处理方法及装置
CN112000820A (zh) 一种媒资推荐方法及显示设备
WO2022032916A1 (zh) 一种显示系统
CN103984772A (zh) 文本检索字幕库生成方法和装置、视频检索方法和装置
WO2020135192A1 (zh) 一种多字幕显示方法、智能终端及存储介质
CN113965813B (zh) 直播间内的视频播放方法、系统、设备及介质
CN104102683A (zh) 用于增强视频显示的上下文查询
WO2022134555A1 (zh) 视频处理方法及终端
CN111625716A (zh) 媒资推荐方法、服务器及显示设备
CN111654732A (zh) 一种广告播放方法及显示设备
CN111914134A (zh) 一种关联推荐方法、智能设备及服务设备
CN111107283B (zh) 一种信息显示方法、电子设备及存储介质
CN114302221A (zh) 一种虚拟现实设备及投屏媒资播放方法
CN113468351A (zh) 一种智能设备及图像处理方法
CN111757187A (zh) 多语言字幕显示方法、装置、终端设备及存储介质
KR102612580B1 (ko) 미디어 제공 서버, 트리거 영역을 통해 다른 컨텐츠로 전환하는 방법 및 컴퓨터 프로그램
CN112053688B (zh) 一种语音交互方法及交互设备、服务器
CN114327033A (zh) 一种虚拟现实设备及媒资播放方法
CN112256232A (zh) 显示设备与自然语言生成后处理方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20875068

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20875068

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20875068

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 29-09-22)