CN114765703B - Method and device for dyeing TTS voice corresponding subtitle and storage medium - Google Patents

Method and device for dyeing TTS voice corresponding subtitle and storage medium Download PDF

Info

Publication number
CN114765703B
CN114765703B CN202110043812.4A CN202110043812A CN114765703B CN 114765703 B CN114765703 B CN 114765703B CN 202110043812 A CN202110043812 A CN 202110043812A CN 114765703 B CN114765703 B CN 114765703B
Authority
CN
China
Prior art keywords
audio
line
playing progress
dyeing
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110043812.4A
Other languages
Chinese (zh)
Other versions
CN114765703A (en
Inventor
杨磊
蒋宁
夏溧
王洪斌
吴海英
刘腾腾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zhongguancun Kejin Technology Co Ltd
Original Assignee
Beijing Zhongguancun Kejin Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zhongguancun Kejin Technology Co Ltd filed Critical Beijing Zhongguancun Kejin Technology Co Ltd
Priority to CN202110043812.4A priority Critical patent/CN114765703B/en
Publication of CN114765703A publication Critical patent/CN114765703A/en
Application granted granted Critical
Publication of CN114765703B publication Critical patent/CN114765703B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Studio Circuits (AREA)

Abstract

The application discloses a method and device for dyeing a TTS voice corresponding subtitle and a storage medium. Wherein the method comprises the following steps: determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions; determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file; under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file; determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.

Description

Method and device for dyeing TTS voice corresponding subtitle and storage medium
Technical Field
The present invention relates to the field of subtitle processing technologies, and in particular, to a method and apparatus for dyeing a subtitle corresponding to TTS speech, and a storage medium.
Background
TTS is an abbreviation for Text To Speech, i.e. "from Text To Speech", which is part of a man-machine conversation, allowing the machine To speak. TTS speech is colloquially speech obtained via text-to-speech technology. In some special scenes, when playing such TTS voice files, subtitles corresponding to the TTS voice files need to be displayed. However, since TTS voice files in these scenes are temporarily generated according to text conversion, they cannot set the progress time of coloring of subtitles in advance like songs. So that the subtitles corresponding to the TTS voice file cannot be dyed accurately.
Aiming at the technical problems that the TTS voice file is temporarily generated according to text conversion, the dyeing progress time of the caption corresponding to the TTS voice file cannot be set in advance, and therefore the caption corresponding to the TTS voice file cannot be accurately dyed, no effective solution is proposed at present.
Disclosure of Invention
The embodiment of the disclosure provides a method, a device and a storage medium for dyeing a caption corresponding to a TTS voice, which at least solve the technical problem that the dyeing progress time of the caption corresponding to the TTS voice file cannot be set in advance because the TTS voice file is temporarily generated according to text conversion, so that the caption corresponding to the TTS voice file cannot be accurately dyed in the prior art.
According to an aspect of the disclosed embodiments, there is provided a method for dyeing a subtitle corresponding to TTS voice, including: determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions; determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file; under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file; determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.
According to another aspect of the embodiments of the present disclosure, there is also provided a storage medium including a stored program, wherein the method of any one of the above is performed by a processor when the program is run.
According to another aspect of the embodiments of the present disclosure, there is also provided a device for dyeing a subtitle corresponding to TTS voice, including: the first determining module is used for determining the audio fragment corresponding to each caption line from the audio file obtained by the text-to-speech technology according to the proportion of the line width of each caption line to the sum of the line widths of all the captions; the second determining module is used for determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file; a third determining module, configured to determine a current playing progress of the audio file when the screen of the current frame is drawn; a fourth determining module, configured to determine a dyeing range of the subtitle displayed in the screen frame of the current frame according to the audio playing progress and the current playing progress; and the dyeing module is used for dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.
According to another aspect of the embodiments of the present disclosure, there is also provided a device for dyeing a subtitle corresponding to TTS voice, including: a processor; and a memory, coupled to the processor, for providing instructions to the processor for processing the steps of: determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions; determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file; under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file; determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.
In the embodiment of the disclosure, first, according to the ratio of the line width of each line of subtitles to the sum of the line widths of all subtitles, an audio clip corresponding to each line of subtitles is determined from an audio file obtained through a text-to-speech technology. And then determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file. Secondly, the difference between the TTS voice file and the song is fully considered, no transitional sound effect exists, and the compact and coherent voice playing is realized, so that the subtitle can be dyed through the playing progress of the TTS voice file, and therefore, under the condition of drawing a screen picture of the current frame, the current playing progress of the audio file, namely the ratio of the current playing length to the total length of the audio file, is required to be determined. And then, determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined audio playing progress and the current playing progress of the audio file at the moment of completing the display of each line of caption. And finally, dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range. Therefore, the embodiment can accurately dye the caption displayed in the screen picture of the current frame through the current playing progress of the audio file obtained through the text-to-speech technology. By coherent screen pictures of all frames which are accurately dyed, the technical effect of accurately dyeing the captions displayed on the screen according to the playing progress of TTS voice is achieved. Further, the technical problem that the existing technology cannot set the dyeing progress time of the caption corresponding to the TTS voice file in advance because the TTS voice file is temporarily generated according to the text conversion, so that the caption corresponding to the TTS voice file cannot be dyed accurately is solved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the disclosure and are incorporated in and constitute a part of this application, illustrate embodiments of the disclosure and together with the description serve to explain the disclosure and do not constitute an undue limitation on the disclosure. In the drawings:
FIG. 1 is a block diagram of a hardware architecture of a computing device for implementing a method according to embodiment 1 of the present disclosure;
fig. 2 is a flowchart of a method for dyeing a TTS voice corresponding subtitle according to embodiment 1 of the present disclosure;
fig. 3 is a schematic diagram of audio clips corresponding to each line of subtitles and an audio playing progress according to embodiment 1 of the disclosure;
fig. 4 is a general flowchart of a method for dyeing a TTS voice corresponding subtitle according to embodiment 1 of the present disclosure;
fig. 5 is a schematic diagram of a device for dyeing a TTS voice corresponding subtitle according to embodiment 2 of the present disclosure; and
fig. 6 is a schematic diagram of a device for dyeing corresponding subtitles in TTS voice according to embodiment 3 of the disclosure.
Detailed Description
In order to better understand the technical solutions of the present disclosure, the following description will clearly and completely describe the technical solutions of the embodiments of the present disclosure with reference to the drawings in the embodiments of the present disclosure. It will be apparent that the described embodiments are merely embodiments of a portion, but not all, of the present disclosure. All other embodiments, which can be made by one of ordinary skill in the art without inventive effort, based on the embodiments in this disclosure, shall fall within the scope of the present disclosure.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the foregoing figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the disclosure described herein may be capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
First, partial terms or terminology appearing in describing embodiments of the present disclosure are applicable to the following explanation:
terminology 1: TTS is an abbreviation for Text To Speech, i.e. "from Text To Speech", which is part of a man-machine conversation, enabling the machine To speak;
term 2: TTS speech is colloquially speech obtained via text-to-speech technology.
Example 1
According to the present embodiment, there is provided an embodiment of a method of coloring a TTS voice corresponding subtitle, it being noted that the steps shown in the flowchart of the drawings may be performed in a computer system such as a set of computer executable instructions, and that although a logical order is shown in the flowchart, in some cases, the steps shown or described may be performed in an order different from that herein.
The method embodiments provided by the present embodiments may be performed in a mobile terminal, a computer terminal, a server, or similar computing device. Fig. 1 shows a hardware block diagram of a computing device of a method for dyeing a TTS voice corresponding subtitle. As shown in fig. 1, the computing device may include one or more processors (which may include, but are not limited to, a microprocessor MCU, a programmable logic device FPGA, etc., processing means), memory for storing data, and transmission means for communication functions. In addition, the method may further include: a display, an input/output interface (I/O interface), a Universal Serial Bus (USB) port (which may be included as one of the ports of the I/O interface), a network interface, a power supply, and/or a camera. It will be appreciated by those of ordinary skill in the art that the configuration shown in fig. 1 is merely illustrative and is not intended to limit the configuration of the electronic device described above. For example, the computing device may also include more or fewer components than shown in FIG. 1, or have a different configuration than shown in FIG. 1.
It should be noted that the one or more processors and/or other data processing circuits described above may be referred to herein generally as "data processing circuits. The data processing circuit may be embodied in whole or in part in software, hardware, firmware, or any other combination. Furthermore, the data processing circuitry may be a single stand-alone processing module, or incorporated in whole or in part into any of the other elements in the computing device. As referred to in the embodiments of the present disclosure, the data processing circuit acts as a processor control (e.g., selection of the variable resistance termination path to interface with).
The memory may be used to store software programs and modules of application software, such as a program instruction/data storage device corresponding to a method for dyeing a TTS voice corresponding subtitle in the embodiments of the present disclosure, and the processor executes various functional applications and data processing by running the software programs and modules stored in the memory, thereby implementing the method for dyeing a TTS voice corresponding subtitle of an application program. The memory may include high speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid state memory. In some examples, the memory may further include memory remotely located with respect to the processor, which may be connected to the computing device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The transmission means is used for receiving or transmitting data via a network. Specific examples of the network described above may include a wireless network provided by a communications provider of the computing device. In one example, the transmission means comprises a network adapter (Network Interface Controller, NIC) connectable to other network devices via the base station to communicate with the internet. In one example, the transmission device may be a Radio Frequency (RF) module, which is used to communicate with the internet wirelessly.
The display may be, for example, a touch screen type Liquid Crystal Display (LCD) that may enable a user to interact with a user interface of the computing device.
It should be noted herein that in some alternative embodiments, the computing device shown in FIG. 1 described above may include hardware elements (including circuitry), software elements (including computer code stored on a computer-readable medium), or a combination of both hardware and software elements. It should be noted that fig. 1 is only one example of a particular specific example and is intended to illustrate the types of components that may be present in the computing devices described above.
In the above-described operating environment, according to a first aspect of the present embodiment, there is provided a method for dyeing a subtitle corresponding to TTS voice. Fig. 2 shows a schematic flow chart of the method, and referring to fig. 2, the method includes:
S202: determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions;
s204: determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file;
s206: under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file;
s208: determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and
s210: and according to the determined dyeing range, dyeing the caption displayed in the screen picture of the current frame.
As described in the background, TTS is an abbreviation for Text To Speech, i.e. "from Text To Speech", which is part of a man-machine conversation, allowing a machine To speak. TTS speech is colloquially speech obtained via text-to-speech technology. In some special scenes, when playing such TTS voice files, subtitles corresponding to the TTS voice files need to be displayed. However, since TTS voice files in these scenes are temporarily generated according to text conversion, they cannot set the progress time of coloring of subtitles in advance like songs. So that the subtitles corresponding to the TTS voice file cannot be dyed accurately.
In order to solve the technical problems in the background art, in the process of dyeing subtitles corresponding to TTS voice files, the technical solution provided in this embodiment first determines, according to the ratio of the line width of each line of subtitles to the sum of the line widths of all subtitles, an audio clip corresponding to each line of subtitles from an audio file (corresponding to the TTS voice file) obtained through a text-to-voice technology. The line width of each line of the subtitles can represent the characters contained in the line of the subtitles to a certain extent, the wider the line width is, the more characters are contained in the line of the subtitles, the longer the corresponding audio segment is, and therefore the line width of each line of the subtitles is proportional to the length of the corresponding audio segment. The line width of each line of subtitles is accurate to a pixel value, and the line width of each line of subtitles is different, for example, the line width of each line of subtitles is inconsistent due to white symbols and mixed Chinese and English arrangement. Therefore, the position of the audio clip corresponding to each line of subtitles in the entire audio file can be calculated by the ratio of the line width of each line of subtitles to the total line width of all subtitles.
As shown in fig. 3, the whole audio file corresponds to three lines of subtitles, the first line of subtitles is AAAAA, the second line of subtitles is bbb, the third line of subtitles is CC, and according to the proportion of the line width of the first line of subtitles to the sum of the line widths of the total subtitles being 50%, the proportion of the line width of the second line of subtitles to the sum of the line widths of the total subtitles being 36%, the proportion of the line width of the third line of subtitles to the sum of the line widths of the total subtitles being 14%, at this time, the audio segment corresponding to the first line of subtitles is calculated as audio segment 1 in fig. 3, the audio segment corresponding to the second line of subtitles is audio segment 2 in fig. 3, and the audio segment corresponding to the third line of subtitles is audio segment 3 in fig. 3. At this time, it is necessary to determine the audio playback progress of the audio file at the time when the display of each line of subtitles is completed, that is, the audio playback progress of the audio file at the time when the display of the first line of subtitles is completed (for example, 50%), the audio playback progress of the audio file at the time when the display of the second line of subtitles is completed (for example, 86%), and the audio playback progress of the audio file at the time when the display of the first line of subtitles is completed (for example, 100%), according to the audio position of the audio clip corresponding to each line of subtitles in the audio file.
Further, since the screen is drawn in accordance with the persistence of vision, one frame of frame is refreshed to produce the effect of continuous animation, everything (including subtitles) needs to be redrawn when drawing the screen picture of each frame. Under the application scene, the TTS voice file is compact and coherent in voice playing in consideration of the fact that the TTS voice file is different from songs and has no transitional sound effect, so that subtitles can be dyed through the playing progress of the TTS voice file. Therefore, in the case of drawing a screen of the current frame, it is necessary to determine the current playing progress of the audio file, i.e., to determine the ratio of the current playing length to the total length of the audio file. Then, according to the determined audio playing progress and current playing progress of the audio file at the moment of completing the display of each line of subtitles, the dyeing range of the subtitles displayed in the screen picture of the current frame is determined. And finally, dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range. For example: and if the determined current playing progress is 50%, determining that the dyeing range of the subtitles displayed in the screen picture of the current frame is all the subtitles comprising the first row of subtitles according to the determined audio playing progress of the audio file at the moment of completing the display of the subtitles of each row, so that all the subtitles of the first row of subtitles displayed in the screen picture of the current frame are required to be dyed, and other row of subtitles are not required to be dyed.
Thus, in the above manner, the present embodiment can accurately dye the subtitle displayed in the screen of the current frame through the current playing progress of the audio file obtained through the text-to-speech technology. By coherent screen pictures of all frames which are accurately dyed, the technical effect of accurately dyeing the captions displayed on the screen according to the playing progress of TTS voice is achieved. Further, the technical problem that the existing technology cannot set the dyeing progress time of the caption corresponding to the TTS voice file in advance because the TTS voice file is temporarily generated according to the text conversion, so that the caption corresponding to the TTS voice file cannot be dyed accurately is solved.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the audio playing progress and the current playing progress includes: comparing the current playing progress with the audio playing progress, and judging whether the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line captions; and determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined result.
In the operation process of determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress, firstly, comparing the current playing progress with the audio playing progress, and judging whether the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the caption of the target line according to the comparison result. For example: the current playback progress is, for example, 51%, and referring to fig. 3, 51% is compared with 50%, 86% and 100%, respectively, and it is determined whether or not 51% exceeds the target audio playback progress (for example, 50%, 86% and 100%) of the audio file at the time when the display of the target line subtitles (for example, the first line subtitle, the second line subtitle and the third line subtitle) is completed. And finally, determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined result. That is, since the current playing progress "51%" exceeds the target audio playing progress "50%" of the audio file at the time when the display of the first line of subtitles is completed, but does not exceed the target audio playing progress "86%" 100% "of the audio file at the time when the display of the second line of subtitles and the third line of subtitles is completed, it is determined that the coloring range of the subtitles displayed in the screen of the current frame is the entire subtitles of the first line of subtitles plus the partial subtitles of the second line of subtitles according to the result of the determination. In this way, the dyeing range of the subtitle displayed in the screen of the current frame can be quickly and accurately determined.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the result of the determination includes: and when the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line caption as a result of the judgment, the target line caption is fully included in the dyeing range of the caption displayed in the screen picture of the current frame.
Specifically, according to the foregoing, in the case where the result of the determination is that the current play progress (i.e., 51%) exceeds the target audio play progress (i.e., 50%) of the audio file at the time when the display of the target line subtitle (e.g., first line subtitle) is completed, the first line subtitle is entirely incorporated into the coloring range of the subtitles displayed in the screen of the current frame, i.e., the entire subtitles of the first line subtitle are subjected to the coloring process. By the method, the subtitle corresponding to the audio clip which is finished being played can be accurately dyed, so that a user can more intuitively check whether the audio corresponding to the target subtitle is played or not, and the use experience of the user is effectively improved.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the result of the determination includes: and under the condition that the current playing progress is not beyond the target audio playing progress of the audio file at the moment of completing the display of the target line caption, the target line caption part is included in the dyeing range of the caption displayed in the screen picture of the current frame according to the current playing progress.
Specifically, according to the foregoing, the current playing progress is, for example, 51%, and since 51% does not exceed the target audio playing progress (i.e., 86%) of the audio file at the time of completing the display of the second line of subtitles, in this application scenario, it is necessary to incorporate the second line of subtitles partially into the coloring range of the subtitles displayed in the screen of the current frame according to the current playing progress "51%" to achieve accurate coloring of the subtitles.
In a preferred embodiment, in the operation of incorporating the target line subtitle portion into the coloring range of the subtitle displayed in the screen of the current frame according to the current playback progress, first, the audio playback progress (i.e., 50%) of the audio file at the time when the display of the previous line subtitle (i.e., the first line subtitle) of the second line subtitle is completed may be determined, and then the difference between the current playback progress (51%) and the determined audio playback progress (50%) may be found, resulting in a difference of 1%. At this time, it is necessary to determine the dyeing width expressed using pixels from the difference value and the sum of the line widths of all subtitles. Wherein 1% can be used multiplied by the sum of the line widths of all subtitles to give a dyeing width of 3 pixel values. Finally, the corresponding width of the subtitles in the second row is brought into the dyeing range according to the dyeing width (i.e. 3 pixel values), i.e. the width of the 3 pixels in the second row is brought into the dyeing range. Thus, in the process of dyeing, not only the first line of subtitles is dyed completely, but also the width of 3 pixels in the second line of subtitles is dyed, and a part of the first subtitle B, for example, is dyed red. By the mode, accurate dyeing of the caption corresponding to the TTS voice is realized.
Optionally, in a case of drawing a screen of the current frame, determining the current playing progress of the audio file includes: acquiring the current playing length of the current playing audio file and the total length of the audio file from a preset system API tool class for playing the audio file; and determining the current playing progress of the audio file according to the acquired current playing length and total length.
Specifically, when the system API tool class is used to play the audio file, the total length and the current playing length of the loaded audio file can be read, for example, the Android system API tool class MediaPlayer can obtain the playing length and the total length of the currently played audio through the MediaPlayer. Therefore, when the screen picture of the current frame is drawn, in the operation process of determining the current playing progress of the audio file, the current playing length of the currently played audio file and the total length of the audio file can be directly obtained from the system API tool class, and then the current playing progress of the audio file is determined according to the obtained current playing length and total length, namely, the ratio of the current playing length to the total length is determined as the current playing progress of the audio file. By the method, the current playing progress of the audio file can be rapidly and accurately obtained.
Optionally, the method further comprises: simulating a playing scene of the audio file obtained by a text-to-speech technology to generate a simulation scene; and determining the audio clips corresponding to each line of subtitles in the simulated scene and the audio playing progress of the audio file at the moment of completing the display of each line of subtitles.
Specifically, before the subtitle corresponding to the TTS voice is dyed, firstly, scene simulation is performed, namely, scenes related to TTS voice playing are simulated, and general TTS solid-color subtitle display is adopted. Then, an algorithm is written, namely, an audio clip corresponding to each line of subtitles in the simulated scene and an audio playing progress of the audio file at the moment of completing the display of each line of subtitles are determined. The audio clip and the audio playing progress obtained by the method can provide powerful comparison basis for dyeing the caption corresponding to the TTS voice, so that the dyeing range of the caption can be accurately judged.
In addition, fig. 4 is a schematic overall flow chart of a method for dyeing a subtitle corresponding to TTS voice according to the present embodiment. Referring to fig. 4, the method for dyeing the corresponding subtitle of TTS voice provided in this embodiment includes the following steps:
1) And simulating a scene related to TTS playing, and displaying general TTS pure-color subtitles.
2) And (3) writing an algorithm, and calculating the corresponding audio progress of each line of subtitles according to the proportion of the line width of each line of subtitles to the sum of the line widths of all lines of subtitles. The line width of the caption can represent the characters contained in the caption in a line to a certain extent, the wider the line width is, the more characters are contained in the caption in the line, and the longer the corresponding audio is, so that the line width of the caption is proportional to the length of the audio. Therefore, the position of the audio clip corresponding to each line of subtitles in the whole audio file can be calculated by the ratio of the line width of the line of subtitles to the total line width. (note: the accuracy of line width to pixel value, the line width of each line is different, simple e.g. white symbols, the mixed Chinese and English lines will cause the line width of each line to be inconsistent)
3) When the system API tool class is used for playing the audio file, the total length of the loaded audio file and the current playing progress, such as the Android system API tool class MediaPlayer, can be read, and the progress and the total length of the currently played audio can be obtained through the MediaPlayer.
4) The screen is drawn frame by frame according to the persistence of vision to produce the effect of continuous animation. When each frame is drawn, everything is redrawn, the process can be described by drawing a picture, such as "ABCDE" on a first sheet of paper, which is the first frame, you need to draw "ABCDE" on a second sheet of paper when you want to dye them step by step, and then dye a red with a stylus, through which a second frame is generated. Then "ABCDE" is drawn on the third sheet of paper and then AB is dyed red with a dye pen to produce a third frame. And so on. When the frames appear continuously, a pair of continuous dyeing animation is generated due to the persistence of vision.
5) By utilizing the principle, for example, every 150 milliseconds, the progress ratio of the current audio file playing (namely, the ratio of the current playing length to the total length) can be obtained, each row of progress corresponding to the subtitle is compared, if one row of progress is exceeded, all the rows are dyed, no exceeding rows are generated, and partial rendering is performed according to the progress value. All frames are connected, so that the subtitle is dyed according to the progress of TTS voice. Such as: the first line of subtitles is "AAAAA", corresponding to a voice progress value of 50%; the second line of subtitles is "bbb", the corresponding speech progress value is 36%; the third line of subtitles is "CC", and the corresponding speech progress value is 14%. For example, the current voice progress is 50%, then the first line of subtitles are all dyed. When the current speech progress is 51%, we need to draw the black font of three lines of AAAAA bbb.cc, then because the current speech progress is 51%, and exceeds 50%, AAAAA is colored red all over again, the remaining 1% is multiplied by the total width of all lines to get 3 pixel values, then bbb of the second line stains 3 pixels apart, revealing that a portion of the first B is colored red. Thus, every 150 milliseconds, the width of the dyeing increases by a part, and the effect of gradual dyeing is displayed on the screen.
In summary, the method provided by the embodiment achieves the technical effect of gradually dyeing the subtitles through the playing progress of the TTS voice file, can be used in double-recording scenes, has good improvement on user experience, and has friendly presentation prompt effect on subsequent double-recording business audit and playback.
Further, referring to fig. 1, according to a second aspect of the present embodiment, there is provided a storage medium. The storage medium includes a stored program, wherein the method of any of the above is performed by a processor when the program is run.
It should be noted that, for simplicity of description, the foregoing method embodiments are all described as a series of acts, but it should be understood by those skilled in the art that the present invention is not limited by the order of acts described, as some steps may be performed in other orders or concurrently in accordance with the present invention. Further, those skilled in the art will also appreciate that the embodiments described in the specification are all preferred embodiments, and that the acts and modules referred to are not necessarily required for the present invention.
From the description of the above embodiments, it will be clear to a person skilled in the art that the method according to the above embodiments may be implemented by means of software plus the necessary general hardware platform, but of course also by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the method according to the embodiments of the present invention.
Example 2
Fig. 5 shows a device 500 for dyeing a TTS voice-corresponding subtitle according to the present embodiment, the device 500 corresponding to the method according to the first aspect of embodiment 1. Referring to fig. 5, the apparatus 500 includes: a first determining module 510, configured to determine, according to a ratio of a line width of each line of subtitles to a sum of line widths of all subtitles, an audio clip corresponding to each line of subtitles from an audio file obtained through a text-to-speech technology; a second determining module 520, configured to determine, according to the audio position of the audio clip corresponding to each line of subtitles in the audio file, an audio playing progress of the audio file at a time when the display of each line of subtitles is completed; a third determining module 530, configured to determine a current playing progress of the audio file in a case of drawing a screen of the current frame; a fourth determining module 540, configured to determine a dyeing range of the subtitle displayed in the screen of the current frame according to the audio playing progress and the current playing progress; and a dyeing module 550 for dyeing the caption displayed in the screen of the current frame according to the determined dyeing range.
Optionally, the fourth determining module 540 includes: the judging submodule is used for comparing the current playing progress with the audio playing progress and judging whether the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line captions; and a first determination submodule for determining a dyeing range of the caption displayed in the screen picture of the current frame according to the result of the determination.
Optionally, the first determining submodule includes: and the first determining unit is used for completely incorporating the target line captions into the dyeing range of the captions displayed in the screen picture of the current frame in the case that the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line captions as a result of the determination.
Optionally, the first determining submodule includes: and a second determining unit for, in the case where the current playing progress is not more than the target audio playing progress of the audio file at the time of completing the display of the target line subtitle as a result of the determination, incorporating the target line subtitle portion into the coloring range of the subtitle displayed in the screen of the current frame according to the current playing progress.
Optionally, the second determining unit includes: a first determining subunit, configured to determine an audio playing progress of the audio file at a time when display of a caption of a previous line of the target line of captions is completed; a difference value calculating subunit, configured to calculate a difference value between the current playing progress and the determined audio playing progress; a second determining subunit for determining a dyeing width represented by pixels according to the difference and the sum of line widths of all subtitles; and a dyeing subunit, configured to bring the subtitles with corresponding widths in the target line subtitle into a dyeing range according to the dyeing width.
Optionally, the third determining module 530 includes: the acquisition sub-module is used for acquiring the current playing length of the current playing audio file and the total length of the audio file from a preset system API tool class for playing the audio file; and the second determining submodule is used for determining the current playing progress of the audio file according to the acquired current playing length and total length.
Optionally, the dyeing apparatus 500 further includes: the simulation module is used for simulating the playing scene of the audio file obtained through the text-to-speech technology and generating a simulation scene; and the audio playing progress determining module is used for determining the audio clips corresponding to each line of subtitles in the simulated scene and the audio playing progress of the audio file at the moment of completing the display of each line of subtitles.
Thus, according to the present embodiment, first, the audio clip corresponding to each line of subtitles is determined from the audio file obtained via the text-to-speech technique according to the ratio of the line width of each line of subtitles to the sum of the line widths of all the subtitles. And then determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file. Secondly, the difference between the TTS voice file and the song is fully considered, no transitional sound effect exists, and the compact and coherent voice playing is realized, so that the subtitle can be dyed through the playing progress of the TTS voice file, and therefore, under the condition of drawing a screen picture of the current frame, the current playing progress of the audio file, namely the ratio of the current playing length to the total length of the audio file, is required to be determined. And then, determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined audio playing progress and the current playing progress of the audio file at the moment of completing the display of each line of caption. And finally, dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range. Therefore, the embodiment can accurately dye the caption displayed in the screen picture of the current frame through the current playing progress of the audio file obtained through the text-to-speech technology. By coherent screen pictures of all frames which are accurately dyed, the technical effect of accurately dyeing the captions displayed on the screen according to the playing progress of TTS voice is achieved. Further, the technical problem that the existing technology cannot set the dyeing progress time of the caption corresponding to the TTS voice file in advance because the TTS voice file is temporarily generated according to the text conversion, so that the caption corresponding to the TTS voice file cannot be dyed accurately is solved.
Example 3
Fig. 6 shows a device 600 for dyeing a TTS voice-corresponding subtitle according to the present embodiment, the device 600 corresponding to the method according to the first aspect of embodiment 1. Referring to fig. 6, the apparatus 600 includes: a processor 610; and a memory 620 coupled to the processor 610 for providing instructions to the processor 610 for processing the following processing steps: determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions; determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file; under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file; determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the audio playing progress and the current playing progress includes: comparing the current playing progress with the audio playing progress, and judging whether the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line captions; and determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined result.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the result of the determination includes: and when the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line caption as a result of the judgment, the target line caption is fully included in the dyeing range of the caption displayed in the screen picture of the current frame.
Optionally, the operation of determining the dyeing range of the subtitle displayed in the screen of the current frame according to the result of the determination includes: and under the condition that the current playing progress is not beyond the target audio playing progress of the audio file at the moment of completing the display of the target line caption, the target line caption part is included in the dyeing range of the caption displayed in the screen picture of the current frame according to the current playing progress.
Optionally, the operation of incorporating the target line subtitle portion into the dyeing range of the subtitle displayed in the screen of the current frame according to the current playing progress includes: determining the audio playing progress of the audio file at the moment of completing the display of the last line of subtitles of the target line of subtitles; solving a difference value between the current playing progress and the determined audio playing progress; determining a dyeing width represented by pixels according to the difference value and the sum of the line widths of all the subtitles; and according to the dyeing width, the caption with the corresponding width in the caption of the target line is included in the dyeing range.
Optionally, in a case of drawing a screen of the current frame, determining the current playing progress of the audio file includes: acquiring the current playing length of the current playing audio file and the total length of the audio file from a preset system API tool class for playing the audio file; and determining the current playing progress of the audio file according to the acquired current playing length and total length.
Optionally, the memory 620 is further configured to provide instructions for the processor 610 to process the following processing steps: simulating a playing scene of the audio file obtained by a text-to-speech technology to generate a simulation scene; and determining the audio clips corresponding to each line of subtitles in the simulated scene and the audio playing progress of the audio file at the moment of completing the display of each line of subtitles.
Thus, according to the present embodiment, first, the audio clip corresponding to each line of subtitles is determined from the audio file obtained via the text-to-speech technique according to the ratio of the line width of each line of subtitles to the sum of the line widths of all the subtitles. And then determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file. Secondly, the difference between the TTS voice file and the song is fully considered, no transitional sound effect exists, and the compact and coherent voice playing is realized, so that the subtitle can be dyed through the playing progress of the TTS voice file, and therefore, under the condition of drawing a screen picture of the current frame, the current playing progress of the audio file, namely the ratio of the current playing length to the total length of the audio file, is required to be determined. And then, determining the dyeing range of the caption displayed in the screen picture of the current frame according to the determined audio playing progress and the current playing progress of the audio file at the moment of completing the display of each line of caption. And finally, dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range. Therefore, the embodiment can accurately dye the caption displayed in the screen picture of the current frame through the current playing progress of the audio file obtained through the text-to-speech technology. By coherent screen pictures of all frames which are accurately dyed, the technical effect of accurately dyeing the captions displayed on the screen according to the playing progress of TTS voice is achieved. Further, the technical problem that the existing technology cannot set the dyeing progress time of the caption corresponding to the TTS voice file in advance because the TTS voice file is temporarily generated according to the text conversion, so that the caption corresponding to the TTS voice file cannot be dyed accurately is solved.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
In the foregoing embodiments of the present invention, the descriptions of the embodiments are emphasized, and for a portion of this disclosure that is not described in detail in this embodiment, reference is made to the related descriptions of other embodiments.
In the several embodiments provided in the present application, it should be understood that the disclosed technology content may be implemented in other manners. The above-described embodiments of the apparatus are merely exemplary, and the division of the units, such as the division of the units, is merely a logical function division, and may be implemented in another manner, for example, multiple units or components may be combined or may be integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be through some interfaces, units or modules, or may be in electrical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in the embodiments of the present invention may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied essentially or in part or all of the technical solution or in part in the form of a software product stored in a storage medium, including instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a Read-only Memory (ROM), a random access Memory (RAM, randomAccess Memory), a removable hard disk, a magnetic disk, or an optical disk, or other various media capable of storing program codes.
The foregoing is merely a preferred embodiment of the present invention and it should be noted that modifications and adaptations to those skilled in the art may be made without departing from the principles of the present invention, which are intended to be comprehended within the scope of the present invention.

Claims (10)

1. A method for dyeing a subtitle corresponding to TTS speech, comprising:
determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions;
determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio fragment corresponding to each line of subtitles in the audio file;
under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file;
determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and
and according to the determined dyeing range, dyeing the caption displayed in the screen picture of the current frame.
2. The method according to claim 1, wherein the operation of determining a coloring range of a subtitle displayed in a screen of the current frame according to the audio playback progress and the current playback progress, comprises:
Comparing the current playing progress with the audio playing progress, and judging whether the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line captions; and
and determining the dyeing range of the caption displayed in the screen picture of the current frame according to the result of the judgment.
3. The method according to claim 2, wherein the operation of determining the coloring range of the subtitle displayed in the screen of the current frame according to the result of the determination, comprises:
and if the current playing progress exceeds the target audio playing progress of the audio file at the moment of completing the display of the target line caption as a result of the judgment, the target line caption is fully included in the dyeing range of the caption displayed in the screen picture of the current frame.
4. The method according to claim 2, wherein the operation of determining the coloring range of the subtitle displayed in the screen of the current frame according to the result of the determination, comprises:
and under the condition that the current playing progress does not exceed the target audio playing progress of the audio file at the moment of completing the display of the target line caption, the target line caption part is included in the dyeing range of the caption displayed in the screen picture of the current frame according to the current playing progress.
5. The method of claim 4, wherein the operation of incorporating the target line subtitle portion into the coloring range of the subtitle displayed in the screen of the current frame according to the current playback progress includes:
determining the audio playing progress of the audio file at the moment of completing the display of the last line of subtitles of the target line of subtitles;
solving a difference value between the current playing progress and the determined audio playing progress;
determining a dyeing width represented by pixels according to the difference value and the sum of the line widths of all the subtitles; and
and according to the dyeing width, the caption with the corresponding width in the target line caption is included in the dyeing range.
6. The method according to claim 1, wherein the operation of determining the current playing progress of the audio file in the case of drawing a screen of a current frame, comprises:
acquiring the current playing length of the current played audio file and the total length of the audio file from a preset system API tool class for playing the audio file; and
and determining the current playing progress of the audio file according to the acquired current playing length and the total length.
7. The method as recited in claim 1, further comprising:
simulating a playing scene of the audio file obtained by a text-to-speech technology to generate a simulation scene; and
and determining an audio clip corresponding to each line of subtitles in the simulated scene and an audio playing progress of the audio file at the moment of completing the display of each line of subtitles.
8. A storage medium comprising a stored program, wherein the method of any one of claims 1 to 7 is performed by a processor when the program is run.
9. A device for dyeing a subtitle corresponding to TTS speech, comprising:
the first determining module is used for determining the audio fragment corresponding to each caption line from the audio file obtained by the text-to-speech technology according to the proportion of the line width of each caption line to the sum of the line widths of all the captions;
the second determining module is used for determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio clip corresponding to each line of subtitles in the audio file;
a third determining module, configured to determine a current playing progress of the audio file when a screen of the current frame is drawn;
A fourth determining module, configured to determine a dyeing range of a subtitle displayed in a screen of the current frame according to the audio playing progress and the current playing progress; and
and the dyeing module is used for dyeing the caption displayed in the screen picture of the current frame according to the determined dyeing range.
10. A device for dyeing a subtitle corresponding to TTS speech, comprising:
a processor; and
a memory, coupled to the processor, for providing instructions to the processor to process the following processing steps:
determining an audio fragment corresponding to each caption line from an audio file obtained by a text-to-speech technology according to the ratio of the line width of each caption line to the sum of the line widths of all the captions;
determining the audio playing progress of the audio file at the moment of completing the display of each line of subtitles according to the audio position of the audio fragment corresponding to each line of subtitles in the audio file;
under the condition of drawing a screen picture of a current frame, determining the current playing progress of the audio file;
determining the dyeing range of the caption displayed in the screen picture of the current frame according to the audio playing progress and the current playing progress; and
And according to the determined dyeing range, dyeing the caption displayed in the screen picture of the current frame.
CN202110043812.4A 2021-01-13 2021-01-13 Method and device for dyeing TTS voice corresponding subtitle and storage medium Active CN114765703B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110043812.4A CN114765703B (en) 2021-01-13 2021-01-13 Method and device for dyeing TTS voice corresponding subtitle and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110043812.4A CN114765703B (en) 2021-01-13 2021-01-13 Method and device for dyeing TTS voice corresponding subtitle and storage medium

Publications (2)

Publication Number Publication Date
CN114765703A CN114765703A (en) 2022-07-19
CN114765703B true CN114765703B (en) 2023-07-07

Family

ID=82363702

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110043812.4A Active CN114765703B (en) 2021-01-13 2021-01-13 Method and device for dyeing TTS voice corresponding subtitle and storage medium

Country Status (1)

Country Link
CN (1) CN114765703B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102037494A (en) * 2008-06-06 2011-04-27 迪维克斯公司 Systems and methods for font file optimization for multimedia files
CN108055592A (en) * 2017-11-21 2018-05-18 广州视源电子科技股份有限公司 Caption presentation method, device, mobile terminal and storage medium
CN110149548A (en) * 2018-09-26 2019-08-20 腾讯科技(深圳)有限公司 Video dubbing method, electronic device and readable storage medium storing program for executing
US10423660B1 (en) * 2017-12-07 2019-09-24 Amazon Technologies, Inc. System for detecting non-synchronization between audio and subtitle
CN110719518A (en) * 2018-07-12 2020-01-21 阿里巴巴集团控股有限公司 Multimedia data processing method, device and equipment

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014141054A1 (en) * 2013-03-11 2014-09-18 Video Dubber Ltd. Method, apparatus and system for regenerating voice intonation in automatically dubbed videos

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102037494A (en) * 2008-06-06 2011-04-27 迪维克斯公司 Systems and methods for font file optimization for multimedia files
CN108055592A (en) * 2017-11-21 2018-05-18 广州视源电子科技股份有限公司 Caption presentation method, device, mobile terminal and storage medium
US10423660B1 (en) * 2017-12-07 2019-09-24 Amazon Technologies, Inc. System for detecting non-synchronization between audio and subtitle
CN110719518A (en) * 2018-07-12 2020-01-21 阿里巴巴集团控股有限公司 Multimedia data processing method, device and equipment
CN110149548A (en) * 2018-09-26 2019-08-20 腾讯科技(深圳)有限公司 Video dubbing method, electronic device and readable storage medium storing program for executing

Also Published As

Publication number Publication date
CN114765703A (en) 2022-07-19

Similar Documents

Publication Publication Date Title
US20180084302A1 (en) Method and apparatus for content insertion during video playback, and storage medium
EP2871560B1 (en) Image processing method and apparatus
CN105955754B (en) A kind of user interface character displaying method and device
WO2022001553A1 (en) Font adjustment method and apparatus, storage medium, and electronic device
US20150269133A1 (en) Electronic book reading incorporating added environmental feel factors
US20150331558A1 (en) Method for switching pictures of picture galleries and browser
CN109254801A (en) Starting, configuration method, device, equipment, medium and operating system
US11367232B2 (en) Method and device for generating stickers
CN104298423A (en) Method and device for displaying media file on terminal through page
US20170161871A1 (en) Method and electronic device for previewing picture on intelligent terminal
US20200034101A1 (en) Data displaying method and electronic device
CN114765703B (en) Method and device for dyeing TTS voice corresponding subtitle and storage medium
CN110443772A (en) Image processing method, device, computer equipment and storage medium
KR100938992B1 (en) Structure of animation font file and method for displaying text data of handheld terminal
CN115148183A (en) Method and device for scrolling subtitles corresponding to TTS voice and storage medium
CN111782095B (en) Switching display method and device for display objects in visual area
CN112433723B (en) Personalized list development method and device
CN110929471A (en) Method and terminal for displaying rich text and electronic equipment
CN109684010A (en) Information display method and device and electronic equipment
US20060062552A1 (en) System and method of adapting sub-picture data for being displayed on mini-screens
CN112328351A (en) Animation display method, animation display device and terminal equipment
KR100510145B1 (en) On screen displaying apparatus and method capable of supporting proportional font
CN114386369A (en) Text processing method and device, electronic equipment and medium
US20170185270A1 (en) Animation display method for view and terminal device
KR100639328B1 (en) Display method of animation using mobile terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant