CN108924636A - Caption presentation method and device - Google Patents

Caption presentation method and device Download PDF

Info

Publication number
CN108924636A
CN108924636A CN201810700375.7A CN201810700375A CN108924636A CN 108924636 A CN108924636 A CN 108924636A CN 201810700375 A CN201810700375 A CN 201810700375A CN 108924636 A CN108924636 A CN 108924636A
Authority
CN
China
Prior art keywords
content
subtitle
video
dialogue
type
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810700375.7A
Other languages
Chinese (zh)
Inventor
张磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alibaba China Co Ltd
Original Assignee
Beijing Youku Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Youku Technology Co Ltd filed Critical Beijing Youku Technology Co Ltd
Priority to CN201810700375.7A priority Critical patent/CN108924636A/en
Publication of CN108924636A publication Critical patent/CN108924636A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

This disclosure relates to a kind of caption presentation method and device.This method includes:According to the content type of the subtitle of video pictures in target video, the display mode of the subtitle of video pictures is determined, content type includes dialogue type, wherein belonging to has dialogue content in the subtitle of dialogue type;Controlling terminal, according to display mode, shows subtitle during playing target video in video pictures.According to the embodiment of the present disclosure, it can be according to the content type of the subtitle of video pictures in target video, determine the display mode of the subtitle of video pictures, content type includes dialogue type, and controlling terminal, according to display mode, shows subtitle during playing target video in video pictures, so that Subtitle Demonstration mode can be different according to the content type of subtitle, meet user to the subtitling view demand of different content type.

Description

Caption presentation method and device
Technical field
This disclosure relates to field of computer technology more particularly to a kind of caption presentation method and device.
Background technique
With the continuous development of science and technology, user can pass through whenever and wherever possible various terminals (for example, mobile phone) and watch it The video liked, and combination subtitle understands the content of video.However, in the related technology, in video display process, subtitle is shown The mode of showing fixes, is single, is unable to satisfy user to the viewing demand of subtitle in video.
Summary of the invention
In view of this, the present disclosure proposes one kind the present disclosure proposes a kind of caption presentation method and device, can satisfy Subtitling view demand of the user to different content type.
According to the one side of the disclosure, a kind of caption presentation method is provided, the method includes:According in target video The content type of the subtitle of video pictures, determines the display mode of the subtitle of the video pictures, and the content type includes pair White type, wherein belonging to has dialogue content in the subtitle of dialogue type;
Controlling terminal is during playing the target video, according to the display mode, in the video pictures Show the subtitle.
In one possible implementation, the method also includes:
According at least one of the caption content, video frame and audio content of the target video, the mesh is determined Mark the content type of the subtitle of video pictures in video.
In one possible implementation, the display mode includes display area,
According to the content type of the subtitle of video pictures in the target video, the aobvious of the subtitle of the video pictures is determined Show mode, including:
The content type be dialogue type when, determine in the video pictures with the dialogue content phase of the subtitle Corresponding target object;
According to target object region locating in the video pictures, the display area is determined.
In one possible implementation, according to the caption content, video frame and audio content of the target video At least one of, determine the content type of the subtitle of video pictures in the target video, including:
At least one of caption content, video frame and audio content to the target video carry out the knowledge of dialogue content Other places reason, obtains recognition result;
When including dialogue content in the recognition result, determine that the content type of the subtitle is dialogue type.
In one possible implementation, the content type further includes non-dialogue type,
Wherein, according at least one of caption content, video frame and audio content of the target video, institute is determined The content type for stating the subtitle of video pictures in target video further includes:
When not including dialogue content in the recognition result, determine that the content type of the subtitle is non-dialogue type.
In one possible implementation, the target video includes animation video,
Controlling terminal is during playing the target video, according to the display mode, in the video pictures Show the subtitle, including:
Controlling terminal during playing the animation video, in video pictures with the dialogue content phase of the subtitle Region locating for corresponding target object shows the dialogue content in the form of dialog box.
According to another aspect of the present disclosure, a kind of subtitling display equipment is provided, described device includes:
Determining module determines the video pictures for the content type according to the subtitle of video pictures in target video Subtitle display mode, the content type includes dialogue type, wherein belong in the subtitle of dialogue type have dialogue in Hold;
Control module, for controlling terminal during playing the target video, according to the display mode, in institute It states and shows the subtitle in video pictures.
In one possible implementation, described device further includes:
Content type determining module, in the caption content, video frame and audio content according to the target video At least one, determine the content type of the subtitle of video pictures in the target video.
In one possible implementation, the display mode includes display area,
The determining module includes:
First determine submodule, for the content type be dialogue type when, determine in the video pictures with The corresponding target object of dialogue content of the subtitle;
Second determines submodule, for the region locating in the video pictures according to the target object, determines institute State display area.
In one possible implementation, the content type determining module includes:
Recognition result acquisition submodule, in caption content, video frame and the audio content to the target video It is at least one carry out the processing of dialogue content recognition, obtain recognition result;
Third determines submodule, when for including dialogue content in the recognition result, determines the content of the subtitle Type is dialogue type.
In one possible implementation, the content type further includes non-dialogue type,
Wherein, the content type determining module further includes:
4th determines that submodule determines the interior of the subtitle when for not including dialogue content in the recognition result Appearance type is non-dialogue type.
In one possible implementation, the target video includes animation video,
The control module includes:
Control submodule, for controlling terminal during playing the animation video, in video pictures with it is described Region locating for the corresponding target object of dialogue content of subtitle shows the dialogue content in the form of dialog box.
According to another aspect of the present disclosure, a kind of subtitling display equipment is provided, including:Processor;It is handled for storage The memory of device executable instruction;Wherein, the processor is configured to executing the above method.
According to another aspect of the present disclosure, a kind of non-volatile computer readable storage medium storing program for executing is provided, is stored thereon with Computer program instructions, wherein the computer program instructions realize above-mentioned caption presentation method when being executed by processor.
According to the embodiment of the present disclosure, video can be determined according to the content type of the subtitle of video pictures in target video The display mode of the subtitle of picture, content type include dialogue type, and controlling terminal is during playing target video, root According to display mode, subtitle is shown in video pictures so that Subtitle Demonstration mode can according to the content type of subtitle and Difference meets user to the subtitling view demand of different content type.
According to below with reference to the accompanying drawings to detailed description of illustrative embodiments, the other feature and aspect of the disclosure will become It is clear.
Detailed description of the invention
Comprising in the description and constituting the attached drawing of part of specification and specification together illustrates the disclosure Exemplary embodiment, feature and aspect, and for explaining the principles of this disclosure.
Fig. 1 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 2 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 3 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 4 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 5 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 6 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.
Fig. 7 is a kind of schematic diagram of the application scenarios of caption presentation method shown according to an exemplary embodiment.
Fig. 8 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.
Fig. 9 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.
Figure 10 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.
Figure 11 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.
Specific embodiment
Various exemplary embodiments, feature and the aspect of the disclosure are described in detail below with reference to attached drawing.It is identical in attached drawing Appended drawing reference indicate element functionally identical or similar.Although the various aspects of embodiment are shown in the attached drawings, remove It non-specifically points out, it is not necessary to attached drawing drawn to scale.
Dedicated word " exemplary " means " being used as example, embodiment or illustrative " herein.Here as " exemplary " Illustrated any embodiment should not necessarily be construed as preferred or advantageous over other embodiments.
In addition, giving numerous details in specific embodiment below to better illustrate the disclosure. It will be appreciated by those skilled in the art that without certain details, the disclosure equally be can be implemented.In some instances, for Method, means, element and circuit well known to those skilled in the art are not described in detail, in order to highlight the purport of the disclosure.
Fig. 1 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.This method can be applied to In terminal device (for example, mobile phone, tablet computer etc.), server, the disclosure to this with no restriction.As shown in Figure 1, according to this public affairs The caption presentation method for opening embodiment includes:
In step s 11, according to the content type of the subtitle of video pictures in target video, the video pictures are determined The display mode of subtitle, the content type include dialogue type, wherein belonging in the subtitle of dialogue type has in dialogue Hold;
In step s 12, controlling terminal is during playing the target video, according to the display mode, in institute It states and shows the subtitle in video pictures.
According to the embodiment of the present disclosure, video can be determined according to the content type of the subtitle of video pictures in target video The display mode of the subtitle of picture, content type include dialogue type, and controlling terminal is during playing target video, root According to display mode, subtitle is shown in video pictures so that Subtitle Demonstration mode can according to the content type of subtitle and Difference meets user to the subtitling view demand of different content type.
Wherein, target video can be the view that film, TV play, variety show, animation video etc. can be watched for user Frequently.The content type of subtitle can be according to caption content classify as a result, for example, can according in subtitle whether include The content type of subtitle is divided into dialogue type and non-dialogue type by dialogue content (for example, human dialog).Wherein, belong to There is dialogue content in the subtitle of dialogue type.Dialogue content can refer in video (for example, film, TV, short-sighted frequency Deng), the lines said.Wherein it is determined that the video pictures subtitle display mode may include be not limited to display when Between, display area, display font, display color, display size, arrangement mode, display background and display background transparency etc., this It discloses to this with no restriction.
In illustrative application scenarios, user wishes to play target video A, for example, passing through its terminal (for example, mobile phone) Play target video A.The content type of the subtitle of video pictures may include dialogue type in target video A.User's Mobile phone is during playing target video A, for example, the subtitle in a certain video pictures includes the dialogue (example of hero and heroine Such as, respectively " I likes you " and " I also likes you " are the subtitle of dialogue type).The subtitle of the dialogue type is drawn in video In face beside hero and heroine, shown in a manner of dialog box.Subtitle in the video pictures can also include other content type, For example, aside (non-dialogue type), then the aside may be displayed on the lower zone in video pictures.
In this way, subtitle is combined with the scene of video pictures, the display mode of subtitle, also, Subtitle Demonstration are enriched Mode can be different according to the content type of subtitle, meet user to the viewing demand of the subtitle of different content type.
For example, the video pictures can be determined according to the content type of the subtitle of video pictures in target video Subtitle display mode, the content type includes dialogue type, wherein belong in the subtitle of dialogue type have dialogue in Hold.
It should be understood that can be before or while playing target video A, according in the subtitle of video pictures in target video Hold type, determines the display mode of the subtitle of the video pictures.
Wherein, in target video A the content type of the subtitle of video pictures can be in advance manually determine (for example, During production subtitle offline, according to the content of the subtitle in video pictures, the content type of subtitle, example are determined Such as, it is determined as dialogue type or non-dialogue type).Alternatively, can also be preparatory or in real time automatically according to the view of target video The content of subtitle in frequency picture, determining content type, as long as in the subtitle according to video pictures in target video Hold type, determines and determined before the display mode of the subtitle of the video pictures, the disclosure is to the content class for determining subtitle The time of type determines main body, method of determination etc. with no restriction.
Fig. 2 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.In a kind of possible reality In existing mode, as shown in Fig. 2, the method also includes:
In step s 13, according at least one in the caption content, video frame and audio content of the target video Kind, determine the content type of the subtitle of video pictures in the target video.
It for example, can at least one of caption content, video frame and audio content to the target video It is analyzed and processed, according to analysis and processing result, determines the content type of the subtitle of video pictures in the target video.
For example, including the dialogue of hero and heroine in the caption content of the target video, then this subtitle (men and women can be determined The dialogue of leading role) content type be dialogue type.Alternatively, the video frame of target video can be analyzed and processed, for example, It, then can be with by carrying out Text region (for example, the conversation content for identifying the hero and heroine in video frame subtitle) to video frame According to the text recognition result, the interior of the subtitle (for example, the conversation content of hero and heroine is dialogue content) of video pictures is determined Appearance type is dialogue type.Alternatively, speech recognition can also be carried out to the audio content of target video, for example, identifying the mesh The dialogue in the corresponding audio content of video pictures of video including hero and heroine is marked, then can determine the content type of the subtitle For dialogue type.It should be understood that can be combined with more in the caption content of the target video, video frame and audio content It is a, determine the content type of the subtitle of video pictures, the disclosure to according to the caption content of the target video, video frame and At least one of audio content determines that the mode of the content type of the subtitle of video pictures in the target video does not limit System.
Fig. 3 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.In a kind of possible reality In existing mode, as shown in figure 3, step S13 may include:
In step S131, at least one of the caption content, video frame and audio content of the target video The processing of dialogue content recognition is carried out, recognition result is obtained;
In step S132, when in the recognition result including dialogue content, determine that the content type of the subtitle is Dialogue type.
Wherein, dialogue content recognition processing can refer to by all kinds of identification methods (for example, Text region, image recognition, One of speech recognition is a variety of) carry out dialogue content identifying processing.
For example, it can be the speech recognition to audio content, for example, automatic speech recognition technology ASR can be passed through (Automatic Speech Recognition) carries out speech recognition to audio content.Alternatively, can be calculated by deep learning Method, training obtain the neural network that can be used for carrying out speech recognition.Language is carried out to audio content using trained neural network Sound identification.For example, identification obtains the corresponding text results of the audio content.Can also be in caption content, video frame extremely A kind of few carry out Text region.For example, with by utilizing optical character identification OCR (Optical Character Recognition) technology carries out Text region at least one of caption content, video frame.Furthermore it is also possible to be to video Frame carry out image recognition, for example, can by based on the good human face recognition model of neural metwork training, article identification model to view Frequency frame carries out recognition of face, article identification, obtains corresponding recognition result.It can be combined with various identification methods, in subtitle At least one of appearance, video frame and audio content carry out the processing of dialogue content recognition, obtain recognition result.
For example, multiple recognition results (for example, for multiple subtitles) for each video pictures of target video can be obtained.Institute When stating in recognition result including dialogue content, determine that the content type of the subtitle is dialogue type.For example, a certain recognition result For female's main clause say " I dislikes you!", then the content type of the subtitle is dialogue type.
In this way, the content type of the subtitle of video pictures can be determined quickly and accurately.The disclosure is to progress dialogue content The mode of identifying processing is with no restriction.
Fig. 4 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.In a kind of possible reality In existing mode, as shown in figure 4, step S13 can also include:
In step S133, when in the recognition result not including dialogue content, the content type of the subtitle is determined For non-dialogue type.
For example, as it was noted above, content type can also include non-dialogue type.For example, at dialogue content recognition It manages in obtained multiple recognition results, does not include dialogue content in partial recognition result.For example, a certain identification content is aside (for example, story background).Do not include dialogue content in the recognition result, can determine that the content type of the subtitle (aside) is non- Dialogue type.
In this way, the content type of the subtitle of video pictures can be determined quickly and accurately.
As shown in Figure 1, in step s 11, according to the content type of the subtitle of video pictures in target video, determine described in The display mode of the subtitle of video pictures, the content type include dialogue type, wherein belong in the subtitle of dialogue type and have There is dialogue content.
For example, the content type of the part subtitle of video pictures is non-dialogue type, other subtitles in target video Content type be dialogue type.The display mode of the subtitle of non-dialogue type can be determined (for example, being shown in video pictures Lower section).The content type that can determine subtitle in video pictures respectively is the display mode of each subtitle of dialogue type.
For example, including the subtitle that two content types are dialogue type, the respectively dialogue of hero and heroine in video pictures. The display mode of the two subtitles can be then determined respectively.For example, the Subtitle Demonstration of the dialogue type of leading man is being located at male master In the dialog box of angle mouth, the subtitle of the dialogue type of leading lady is shown in beside leading lady in a manner of tracery or strip.
Fig. 5 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.In a kind of possible reality In existing mode, the display mode includes display area, as shown in figure 5, step S11 may include:
In step S111, the content type be dialogue type when, determine in the video pictures with the word The corresponding target object of dialogue content of curtain;
In step S112, according to target object region locating in the video pictures, the display is determined Region.
As it was noted above, display mode may include display area.Wherein, target object can be in video pictures with institute The corresponding object of dialogue content of subtitle is stated, for example, it may be the sounding object etc. of dialogue content.Target object is in the view Locating region can be the region of target object in frequency picture, be also possible to target object peripheral region, the disclosure to this not It is restricted.
For example, it when content type is dialogue type, can determine in video pictures and in the dialogue of the subtitle Hold corresponding target object.For example, a certain subtitle is the dialogue from leading man, then the corresponding target object of the subtitle is Leading man.Region that can be locating in the video pictures according to leading man, for example, the beside of leading man, is determined as The display area of the subtitle.
In this way, the subtitle of dialogue type can be dissolved into video scene, the mode of Subtitle Demonstration is enriched.The disclosure pair Determine target object corresponding with the dialogue content of the subtitle in the video pictures and according to the target object Locating region, determines the mode of the display area with no restriction in the video pictures.
As shown in Figure 1, in step s 12, controlling terminal is during playing the target video, according to the display Mode shows the subtitle in the video pictures.
For example, user is during passing through terminal plays target video, for example, as it was noted above, a certain subtitle For " I likes you ", come from leading man, the display mode of the subtitle determined be beside leading man in a manner of strip, font It is black display for No. 4, color.Can beside leading man in the way of strip, font be No. 4, color is black display " I Like you ".
Fig. 6 is a kind of flow chart of caption presentation method shown according to an exemplary embodiment.In a kind of possible reality In existing mode, the target video includes animation video, as shown in fig. 6, step S12 may include:
In step S121, controlling terminal during playing the animation video, in video pictures with the word Region locating for the corresponding target object of dialogue content of curtain shows the dialogue content in the form of dialog box.
For example, as it was noted above, target video may include animation video.It, can in animation video display process With region locating for target object corresponding with the dialogue content of the subtitle in video pictures, shown in the form of dialog box Show the dialogue content.Wherein, dialog box may include various forms, for example, can be bubble frame, rectangle frame etc., dialog box There can also be different display modes, for example, can have with the dialogue text in different colors, dialog box different Font, font size etc..For example, the dialogue of hero and heroine is shown in the form of rectangle frame respectively beside hero and heroine.
In this way, the video pictures that user class apparent movement overflows style can be supplied to, the display mode of video pictures is enriched.This public affairs It opens to the form and display mode of dialog box with no restriction.
Using example
Below in conjunction with " user play animation video " property application scenarios as an example, provide according to the embodiment of the present disclosure Application example, in order to understand the process of caption presentation method.It will be understood by those skilled in the art that below only using example It is to be not construed as the limitation to the embodiment of the present disclosure for the purpose for being easy to understand the embodiment of the present disclosure.
Fig. 7 is a kind of schematic diagram of the application scenarios of caption presentation method shown according to an exemplary embodiment.At this Using in example, user wishes to play a video, and the playing request for being directed to video is initiated by its mobile phone.Server is receiving The playing request determines that user mobile phone wishes the video played.Server can return to the video file of the video, for example, control User mobile phone processed plays the video.
This using in example, server can be in advance or in real time according in the video in the subtitle of video pictures Hold type, determines the display mode of the subtitle of video pictures.For example, the video is drawn by taking video pictures a certain in the video as an example Face is that two animation roles are talking with, and subtitle includes:" it is well suited for you!" and " thanks ".
It is applied in example at this, server can determine the content type of the subtitle of video pictures in advance or in real time.Example Such as, you " are well suited for according to caption content determination!" and " thanks " be dialogue content, determine the content class of the two subtitles Type is dialogue type.
It is applied in example at this, server can control mobile phone during playing the target video, according to described Display mode shows the subtitle in the video pictures.For example, server can be according to determining in the video pictures Target object corresponding with the dialogue content of the subtitle, and it is locating in the video pictures according to the target object Region determines the display area (display mode includes display area).For example, determination " is well suited for you!" it is video pictures The dialogue of middle right side animation role, " thanks " are the dialogue of left side animation role in video pictures.Server can be according to each Target object region locating in video pictures, determines display area.For example, determination " is well suited for you!" display area Beside right side animation role, the display area of " thanks " is beside left side animation role.
It is applied in example at this, server can control user mobile phone during playing the animation video, regard Region locating for target object corresponding with the dialogue content of the subtitle in frequency picture, in the form of dialog box described in display Dialogue content.For example, as shown in fig. 7, showing dialogue content in the form of dialog box beside two animation roles respectively.
In this way, user's viewing can be facilitated, understand dialogue source, it is immersed in video scene convenient for user.
Fig. 8 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.As shown in figure 8, the dress Set including:
Determining module 21 determines that the video is drawn for the content type according to the subtitle of video pictures in target video The display mode of the subtitle in face, the content type include dialogue type, wherein belonging to has dialogue in the subtitle of dialogue type Content;
Control module 22, for controlling terminal during playing the target video, according to the display mode, The subtitle is shown in the video pictures.
Fig. 9 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.As shown in figure 9, in one kind In possible implementation, described device further includes:
Content type determining module 23, for the caption content, video frame and audio content according to the target video At least one of, determine the content type of the subtitle of video pictures in the target video.
As shown in figure 9, in one possible implementation, the display mode includes display area,
The determining module 21 includes:
First determines submodule 211, for determining in the video pictures when the content type is dialogue type Target object corresponding with the dialogue content of the subtitle;
Second determines submodule 212, for the region locating in the video pictures according to the target object, determines The display area.
As shown in figure 9, in one possible implementation, the content type determining module 23 includes:
Recognition result acquisition submodule 231, for caption content, video frame and the audio content to the target video At least one of carry out the processing of dialogue content recognition, obtain recognition result;
Third determines submodule 232, when for including dialogue content in the recognition result, determines the interior of the subtitle Appearance type is dialogue type.
As shown in figure 9, in one possible implementation, the content type further includes non-dialogue type,
Wherein, the content type determining module 23 further includes:
4th determines that submodule 233 determines the subtitle when for not including dialogue content in the recognition result Content type is non-dialogue type.
As shown in figure 9, in one possible implementation, the target video includes animation video,
The control module 22 includes:
Control submodule 221, for controlling terminal during playing the animation video, in video pictures with institute Region locating for the corresponding target object of dialogue content of subtitle is stated, the dialogue content is shown in the form of dialog box.
Figure 10 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.For example, device 800 can be with It is mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, Medical Devices, body-building Equipment, personal digital assistant etc..
Referring to Fig.1 0, device 800 may include following one or more components:Processing component 802, memory 804, power supply Component 806, multimedia component 808, audio component 810, the interface 812 of input/output (I/O), sensor module 814, and Communication component 816.
The integrated operation of the usual control device 800 of processing component 802, such as with display, telephone call, data communication, phase Machine operation and record operate associated operation.Processing component 802 may include that one or more processors 820 refer to execute It enables, to perform all or part of the steps of the methods described above.In addition, processing component 802 may include one or more modules, just Interaction between processing component 802 and other assemblies.For example, processing component 802 may include multi-media module, it is more to facilitate Interaction between media component 808 and processing component 802.
Memory 804 is configured as storing various types of data to support the operation in device 800.These data are shown Example includes the instruction of any application or method for operating on device 800, contact data, and telephone book data disappears Breath, picture, video etc..Memory 804 can be by any kind of volatibility or non-volatile memory device or their group It closes and realizes, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM) is erasable to compile Journey read-only memory (EPROM), programmable read only memory (PROM), read-only memory (ROM), magnetic memory, flash Device, disk or CD.
Power supply module 806 provides electric power for the various assemblies of device 800.Power supply module 806 may include power management system System, one or more power supplys and other with for device 800 generate, manage, and distribute the associated component of electric power.
Multimedia component 808 includes the screen of one output interface of offer between described device 800 and user.One In a little embodiments, screen may include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, screen Curtain may be implemented as touch screen, to receive input signal from the user.Touch panel includes one or more touch sensings Device is to sense the gesture on touch, slide, and touch panel.The touch sensor can not only sense touch or sliding action Boundary, but also detect duration and pressure associated with the touch or slide operation.In some embodiments, more matchmakers Body component 808 includes a front camera and/or rear camera.When device 800 is in operation mode, such as screening-mode or When video mode, front camera and/or rear camera can receive external multi-medium data.Each front camera and Rear camera can be a fixed optical lens system or have focusing and optical zoom capabilities.
Audio component 810 is configured as output and/or input audio signal.For example, audio component 810 includes a Mike Wind (MIC), when device 800 is in operation mode, when such as call mode, recording mode, and voice recognition mode, microphone is matched It is set to reception external audio signal.The received audio signal can be further stored in memory 804 or via communication set Part 816 is sent.In some embodiments, audio component 810 further includes a loudspeaker, is used for output audio signal.
I/O interface 812 provides interface between processing component 802 and peripheral interface module, and above-mentioned peripheral interface module can To be keyboard, click wheel, button etc..These buttons may include, but are not limited to:Home button, volume button, start button and lock Determine button.
Sensor module 814 includes one or more sensors, and the state for providing various aspects for device 800 is commented Estimate.For example, sensor module 814 can detecte the state that opens/closes of device 800, and the relative positioning of component, for example, it is described Component is the display and keypad of device 800, and sensor module 814 can be with 800 1 components of detection device 800 or device Position change, the existence or non-existence that user contacts with device 800,800 orientation of device or acceleration/deceleration and device 800 Temperature change.Sensor module 814 may include proximity sensor, be configured to detect without any physical contact Presence of nearby objects.Sensor module 814 can also include optical sensor, such as CMOS or ccd image sensor, at As being used in application.In some embodiments, which can also include acceleration transducer, gyro sensors Device, Magnetic Sensor, pressure sensor or temperature sensor.
Communication component 816 is configured to facilitate the communication of wired or wireless way between device 800 and other equipment.Device 800 can access the wireless network based on communication standard, such as WiFi, 2G or 3G or their combination.In an exemplary implementation In example, communication component 816 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel. In one exemplary embodiment, the communication component 816 further includes near-field communication (NFC) module, to promote short range communication.Example Such as, NFC module can be based on radio frequency identification (RFID) technology, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology, Bluetooth (BT) technology and other technologies are realized.
In the exemplary embodiment, device 800 can be believed by one or more application specific integrated circuit (ASIC), number Number processor (DSP), digital signal processing appts (DSPD), programmable logic device (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are realized, for executing the above method.
In the exemplary embodiment, a kind of non-volatile computer readable storage medium storing program for executing is additionally provided, for example including calculating The memory 804 of machine program instruction, above-mentioned computer program instructions can be executed above-mentioned to complete by the processor 820 of device 800 Method.
Figure 11 is a kind of block diagram of subtitling display equipment shown according to an exemplary embodiment.For example, device 1900 can To be provided as a server.Referring to Fig.1 1, it further comprises one or more places that device 1900, which includes processing component 1922, Manage device and memory resource represented by a memory 1932, for store can by the instruction of the execution of processing component 1922, Such as application program.The application program stored in memory 1932 may include it is one or more each correspond to one The module of group instruction.In addition, processing component 1922 is configured as executing instruction, to execute the above method.
Device 1900 can also include that a power supply module 1926 be configured as the power management of executive device 1900, and one Wired or wireless network interface 1950 is configured as device 1900 being connected to network and input and output (I/O) interface 1958.Device 1900 can be operated based on the operating system for being stored in memory 1932, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or similar.
In the exemplary embodiment, a kind of non-volatile computer readable storage medium storing program for executing is additionally provided, for example including calculating The memory 1932 of machine program instruction, above-mentioned computer program instructions can be executed by the processing component 1922 of device 1900 to complete The above method.
The disclosure can be system, method and/or computer program product.Computer program product may include computer Readable storage medium storing program for executing, containing for making processor realize the computer-readable program instructions of various aspects of the disclosure.
Computer readable storage medium, which can be, can keep and store the tangible of the instruction used by instruction execution equipment Equipment.Computer readable storage medium for example can be-- but it is not limited to-- storage device electric, magnetic storage apparatus, optical storage Equipment, electric magnetic storage apparatus, semiconductor memory apparatus or above-mentioned any appropriate combination.Computer readable storage medium More specific example (non exhaustive list) includes:Portable computer diskette, random access memory (RAM), read-only is deposited hard disk It is reservoir (ROM), erasable programmable read only memory (EPROM or flash memory), static random access memory (SRAM), portable Compact disk read-only memory (CD-ROM), digital versatile disc (DVD), memory stick, floppy disk, mechanical coding equipment, for example thereon It is stored with punch card or groove internal projection structure and the above-mentioned any appropriate combination of instruction.Calculating used herein above Machine readable storage medium storing program for executing is not interpreted that instantaneous signal itself, the electromagnetic wave of such as radio wave or other Free propagations lead to It crosses the electromagnetic wave (for example, the light pulse for passing through fiber optic cables) of waveguide or the propagation of other transmission mediums or is transmitted by electric wire Electric signal.
Computer-readable program instructions as described herein can be downloaded to from computer readable storage medium it is each calculate/ Processing equipment, or outer computer or outer is downloaded to by network, such as internet, local area network, wide area network and/or wireless network Portion stores equipment.Network may include copper transmission cable, optical fiber transmission, wireless transmission, router, firewall, interchanger, gateway Computer and/or Edge Server.Adapter or network interface in each calculating/processing equipment are received from network to be counted Calculation machine readable program instructions, and the computer-readable program instructions are forwarded, for the meter being stored in each calculating/processing equipment In calculation machine readable storage medium storing program for executing.
Computer program instructions for executing disclosure operation can be assembly instruction, instruction set architecture (ISA) instructs, Machine instruction, machine-dependent instructions, microcode, firmware instructions, condition setup data or with one or more programming languages The source code or object code that any combination is write, the programming language include the programming language-of object-oriented such as Smalltalk, C++ etc., and conventional procedural programming languages-such as " C " language or similar programming language.Computer Readable program instructions can be executed fully on the user computer, partly execute on the user computer, be only as one Vertical software package executes, part executes on the remote computer or completely in remote computer on the user computer for part Or it is executed on server.In situations involving remote computers, remote computer can pass through network-packet of any kind It includes local area network (LAN) or wide area network (WAN)-is connected to subscriber computer, or, it may be connected to outer computer (such as benefit It is connected with ISP by internet).In some embodiments, by utilizing computer-readable program instructions Status information carry out personalized customization electronic circuit, such as programmable logic circuit, field programmable gate array (FPGA) or can Programmed logic array (PLA) (PLA), the electronic circuit can execute computer-readable program instructions, to realize each side of the disclosure Face.
Referring herein to according to the flow chart of the method, apparatus (system) of the embodiment of the present disclosure and computer program product and/ Or block diagram describes various aspects of the disclosure.It should be appreciated that flowchart and or block diagram each box and flow chart and/ Or in block diagram each box combination, can be realized by computer-readable program instructions.
These computer-readable program instructions can be supplied to general purpose computer, special purpose computer or other programmable datas The processor of processing unit, so that a kind of machine is produced, so that these instructions are passing through computer or other programmable datas When the processor of processing unit executes, function specified in one or more boxes in implementation flow chart and/or block diagram is produced The device of energy/movement.These computer-readable program instructions can also be stored in a computer-readable storage medium, these refer to It enables so that computer, programmable data processing unit and/or other equipment work in a specific way, thus, it is stored with instruction Computer-readable medium then includes a manufacture comprising in one or more boxes in implementation flow chart and/or block diagram The instruction of the various aspects of defined function action.
Computer-readable program instructions can also be loaded into computer, other programmable data processing units or other In equipment, so that series of operation steps are executed in computer, other programmable data processing units or other equipment, to produce Raw computer implemented process, so that executed in computer, other programmable data processing units or other equipment Instruct function action specified in one or more boxes in implementation flow chart and/or block diagram.
The flow chart and block diagram in the drawings show system, method and the computer journeys according to multiple embodiments of the disclosure The architecture, function and operation in the cards of sequence product.In this regard, each box in flowchart or block diagram can generation One module of table, program segment or a part of instruction, the module, program segment or a part of instruction include one or more use The executable instruction of the logic function as defined in realizing.In some implementations as replacements, function marked in the box It can occur in a different order than that indicated in the drawings.For example, two continuous boxes can actually be held substantially in parallel Row, they can also be executed in the opposite order sometimes, and this depends on the function involved.It is also noted that block diagram and/or The combination of each box in flow chart and the box in block diagram and or flow chart, can the function as defined in executing or dynamic The dedicated hardware based system made is realized, or can be realized using a combination of dedicated hardware and computer instructions.
The presently disclosed embodiments is described above, above description is exemplary, and non-exclusive, and It is not limited to disclosed each embodiment.Without departing from the scope and spirit of illustrated each embodiment, for this skill Many modifications and changes are obvious for the those of ordinary skill in art field.The selection of term used herein, purport In the principle, practical application or technological improvement to the technology in market for best explaining each embodiment, or lead this technology Other those of ordinary skill in domain can understand each embodiment disclosed herein.

Claims (14)

1. a kind of caption presentation method, which is characterized in that the method includes:
According to the content type of the subtitle of video pictures in target video, the display mode of the subtitle of the video pictures is determined, The content type includes dialogue type, wherein belonging to has dialogue content in the subtitle of dialogue type;
Controlling terminal, according to the display mode, is shown during playing the target video in the video pictures The subtitle.
2. the method according to claim 1, wherein the method also includes:
According at least one of the caption content, video frame and audio content of the target video, the target view is determined The content type of the subtitle of video pictures in frequency.
3. the method according to claim 1, wherein the display mode includes display area,
According to the content type of the subtitle of video pictures in the target video, the display side of the subtitle of the video pictures is determined Formula, including:
When the content type is dialogue type, determination is corresponding with the dialogue content of the subtitle in the video pictures Target object;
According to target object region locating in the video pictures, the display area is determined.
4. according to the method described in claim 2, it is characterized in that, according to the caption content of the target video, video frame with And at least one of audio content, determine the content type of the subtitle of video pictures in the target video, including:
At least one of caption content, video frame and audio content to the target video carry out at dialogue content recognition Reason obtains recognition result;
When including dialogue content in the recognition result, determine that the content type of the subtitle is dialogue type.
5. according to the method described in claim 4, it is characterized in that, the content type further includes non-dialogue type,
Wherein, according at least one of caption content, video frame and audio content of the target video, the mesh is determined Mark video in video pictures subtitle content type, further include:
When not including dialogue content in the recognition result, determine that the content type of the subtitle is non-dialogue type.
6. the method according to claim 1, wherein the target video includes animation video,
Controlling terminal, according to the display mode, is shown during playing the target video in the video pictures The subtitle, including:
Controlling terminal is corresponding with the dialogue content of the subtitle in video pictures during playing the animation video Target object locating for region, the dialogue content is shown in the form of dialog box.
7. a kind of subtitling display equipment, which is characterized in that described device includes:
Determining module determines the word of the video pictures for the content type according to the subtitle of video pictures in target video The display mode of curtain, the content type includes dialogue type, wherein belonging to has dialogue content in the subtitle of dialogue type;
Control module, for controlling terminal during playing the target video, according to the display mode, in the view The subtitle is shown in frequency picture.
8. device according to claim 7, which is characterized in that described device further includes:
Content type determining module, in the caption content, video frame and audio content according to the target video extremely Few one kind, determines the content type of the subtitle of video pictures in the target video.
9. device according to claim 7, which is characterized in that the display mode includes display area,
The determining module includes:
First determine submodule, for the content type be dialogue type when, determine in the video pictures with it is described The corresponding target object of dialogue content of subtitle;
Second determines submodule, for the region locating in the video pictures according to the target object, determines described aobvious Show region.
10. device according to claim 8, which is characterized in that the content type determining module includes:
Recognition result acquisition submodule, in caption content, video frame and the audio content to the target video extremely A kind of few progress dialogue content recognition processing, obtains recognition result;
Third determines submodule, when for including dialogue content in the recognition result, determines the content type of the subtitle For dialogue type.
11. device according to claim 10, which is characterized in that the content type further includes non-dialogue type,
Wherein, the content type determining module further includes:
4th determines submodule, when for not including dialogue content in the recognition result, determines the content class of the subtitle Type is non-dialogue type.
12. device according to claim 7, which is characterized in that the target video includes animation video,
The control module includes:
Control submodule, for controlling terminal during playing the animation video, in video pictures with the subtitle The corresponding target object of dialogue content locating for region, the dialogue content is shown in the form of dialog box.
13. a kind of subtitling display equipment, which is characterized in that including:
Processor;
Memory for storage processor executable instruction;
Wherein, the processor is configured to:Method described in any one of perform claim requirement 1 to 6.
14. a kind of non-volatile computer readable storage medium storing program for executing, is stored thereon with computer program instructions, which is characterized in that institute It states and realizes method described in any one of claim 1 to 6 when computer program instructions are executed by processor.
CN201810700375.7A 2018-06-29 2018-06-29 Caption presentation method and device Pending CN108924636A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810700375.7A CN108924636A (en) 2018-06-29 2018-06-29 Caption presentation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810700375.7A CN108924636A (en) 2018-06-29 2018-06-29 Caption presentation method and device

Publications (1)

Publication Number Publication Date
CN108924636A true CN108924636A (en) 2018-11-30

Family

ID=64424244

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810700375.7A Pending CN108924636A (en) 2018-06-29 2018-06-29 Caption presentation method and device

Country Status (1)

Country Link
CN (1) CN108924636A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110198468A (en) * 2019-05-15 2019-09-03 北京奇艺世纪科技有限公司 A kind of video caption display methods, device and electronic equipment

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101064177A (en) * 2006-04-26 2007-10-31 松下电器产业株式会社 Caption display control apparatus
CN101483723A (en) * 2008-01-11 2009-07-15 新奥特(北京)视频技术有限公司 Method for performance guarantee of television subtitle playing apparatus based on diversity application
CN102111601A (en) * 2009-12-23 2011-06-29 大猩猩科技股份有限公司 Content-based adaptive multimedia processing system and method
CN102292977A (en) * 2009-01-27 2011-12-21 艾科星科技公司 Systems and methods for providing closed captioning in three-dimensional imagery
CN103139375A (en) * 2011-12-02 2013-06-05 Lg电子株式会社 Mobile terminal and control method thereof
CN103647909A (en) * 2013-12-16 2014-03-19 宇龙计算机通信科技(深圳)有限公司 Caption adjusting method and caption adjusting device
CN103856689A (en) * 2013-10-31 2014-06-11 北京中科模识科技有限公司 Character dialogue subtitle extraction method oriented to news video
US20140201631A1 (en) * 2013-01-15 2014-07-17 Viki, Inc. System and method for captioning media
CN104199885A (en) * 2014-08-22 2014-12-10 广州三星通信技术研究有限公司 Device and method for acquiring relevant content of video files
CN104540275A (en) * 2014-12-17 2015-04-22 欧普照明股份有限公司 Method, device and system for adjusting site lighting device
CN104915433A (en) * 2015-06-24 2015-09-16 宁波工程学院 Method for searching for film and television video

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101064177A (en) * 2006-04-26 2007-10-31 松下电器产业株式会社 Caption display control apparatus
CN101483723A (en) * 2008-01-11 2009-07-15 新奥特(北京)视频技术有限公司 Method for performance guarantee of television subtitle playing apparatus based on diversity application
CN102292977A (en) * 2009-01-27 2011-12-21 艾科星科技公司 Systems and methods for providing closed captioning in three-dimensional imagery
CN102111601A (en) * 2009-12-23 2011-06-29 大猩猩科技股份有限公司 Content-based adaptive multimedia processing system and method
CN103139375A (en) * 2011-12-02 2013-06-05 Lg电子株式会社 Mobile terminal and control method thereof
US20140201631A1 (en) * 2013-01-15 2014-07-17 Viki, Inc. System and method for captioning media
CN103856689A (en) * 2013-10-31 2014-06-11 北京中科模识科技有限公司 Character dialogue subtitle extraction method oriented to news video
CN103647909A (en) * 2013-12-16 2014-03-19 宇龙计算机通信科技(深圳)有限公司 Caption adjusting method and caption adjusting device
CN104199885A (en) * 2014-08-22 2014-12-10 广州三星通信技术研究有限公司 Device and method for acquiring relevant content of video files
CN104540275A (en) * 2014-12-17 2015-04-22 欧普照明股份有限公司 Method, device and system for adjusting site lighting device
CN104915433A (en) * 2015-06-24 2015-09-16 宁波工程学院 Method for searching for film and television video

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
郑建鹏,李建萍: "《广告文案写作》", 31 July 2017, 中国传媒大学出版社 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110198468A (en) * 2019-05-15 2019-09-03 北京奇艺世纪科技有限公司 A kind of video caption display methods, device and electronic equipment

Similar Documents

Publication Publication Date Title
CN109089170A (en) Barrage display methods and device
CN109618184A (en) Method for processing video frequency and device, electronic equipment and storage medium
CN104918107B (en) The identification processing method and device of video file
CN110210310B (en) Video processing method and device for video processing
CN109872297A (en) Image processing method and device, electronic equipment and storage medium
CN108174247A (en) Video interaction method and device
CN106792170A (en) Method for processing video frequency and device
CN107948708A (en) Barrage methods of exhibiting and device
CN106993229A (en) Interactive attribute methods of exhibiting and device
CN108900888A (en) Control method for playing back and device
CN109963200A (en) Video broadcasting method and device
CN109151356A (en) video recording method and device
CN108260020A (en) The method and apparatus that interactive information is shown in panoramic video
CN108924644A (en) Video clip extracting method and device
CN108985176A (en) image generating method and device
CN108259991A (en) Method for processing video frequency and device
CN110519655A (en) Video clipping method and device
CN109005352A (en) It is in step with the method and device of video
CN106960014A (en) Association user recommends method and device
CN109407944A (en) Multimedia resource plays adjusting method and device
CN107832746A (en) Expression recognition method and device
CN108540850A (en) Barrage display methods and device
CN109803158A (en) Video broadcasting method and device
CN108174269A (en) Visualize audio frequency playing method and device
CN110209877A (en) Video analysis method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20200420

Address after: 310052 room 508, floor 5, building 4, No. 699, Wangshang Road, Changhe street, Binjiang District, Hangzhou City, Zhejiang Province

Applicant after: Alibaba (China) Co.,Ltd.

Address before: 100000 room 26, 9 Building 9, Wangjing east garden four, Chaoyang District, Beijing.

Applicant before: BEIJING YOUKU TECHNOLOGY Co.,Ltd.

TA01 Transfer of patent application right
RJ01 Rejection of invention patent application after publication

Application publication date: 20181130

RJ01 Rejection of invention patent application after publication