WO2017076171A1 - 一种在视频图像上显示信息的方法和装置 - Google Patents

一种在视频图像上显示信息的方法和装置 Download PDF

Info

Publication number
WO2017076171A1
WO2017076171A1 PCT/CN2016/102640 CN2016102640W WO2017076171A1 WO 2017076171 A1 WO2017076171 A1 WO 2017076171A1 CN 2016102640 W CN2016102640 W CN 2016102640W WO 2017076171 A1 WO2017076171 A1 WO 2017076171A1
Authority
WO
WIPO (PCT)
Prior art keywords
image frame
information
facial feature
image
location
Prior art date
Application number
PCT/CN2016/102640
Other languages
English (en)
French (fr)
Inventor
黄利华
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Priority to JP2017560519A priority Critical patent/JP6438598B2/ja
Publication of WO2017076171A1 publication Critical patent/WO2017076171A1/zh
Priority to US15/787,908 priority patent/US10425679B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/475End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • G06F16/784Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content the detected or recognised objects being people
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/812Monomedia components thereof involving advertisement data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • H04N21/8133Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation

Definitions

  • the present application relates to the field of video technologies, and in particular, to a method and apparatus for displaying information on a video image.
  • the barrage refers to a comment that appears directly on the video. It can appear on the video as a scroll, pause, or even more action special effect, which is a short comment sent by the person watching the video.
  • the barrage video is a video with a "barrage".
  • many websites provide video transmission barrage functions, such as niconico, acfun, bilibili, dilili, tucao, and explosion point TV.
  • such websites can allow viewers to watch comments or impressions, but they are different from ordinary video sharing websites only in the special comment area under the player. They will appear on the video screen in real time by sliding subtitles to ensure all viewing.
  • Teen can notice that the interaction between viewers can be realized, and even the praise or criticism of the work can be expressed together to increase the fun of watching.
  • the barrage information randomly appears on the video screen, it will appear at any position of the video screen, so that there is a drawback that the barrage obscures the important content of the video, such as occluding key information such as a face in the video image. .
  • the present application provides a method and apparatus for displaying information on a video image, which can avoid facial feature display information on a video image, and improve a video image processing function of the playback device.
  • a method for displaying information on a video image, applied to a playback device, the method package include:
  • the playback device receives an information play request
  • the received information is displayed at a position other than the determined position of the facial feature in the image frame.
  • An apparatus for displaying information on a video image is applied to a playback device, the apparatus comprising at least a processor and a memory, wherein the memory stores a receiving unit, a parsing unit, an obtaining unit, a determining unit, and a display unit, when When the processor executes,
  • the receiving unit is configured to receive an information play request
  • the parsing unit is configured to parse an image frame of a currently playing video
  • the identifying unit is configured to identify a facial feature in the image frame obtained by the parsing
  • the determining unit is configured to determine a location of the recognized facial feature in the image frame
  • the display unit is configured to display the received information in a position other than the determined position of the facial feature in the image frame.
  • a non-volatile computer storage medium having stored therein a computer program for performing the method described above.
  • FIG. 1 is a schematic flow chart of displaying information on a video image in an embodiment of the present application
  • FIG. 2 is a schematic flow chart of displaying information on a video image in an embodiment of the present application
  • FIG. 3 is a schematic flow chart of displaying information on a video image in an embodiment of the present application.
  • FIG. 4 is a schematic flow chart of displaying information on a video image in an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a video playing system according to an embodiment of the present application.
  • FIG. 6 is a schematic diagram of displaying barrographic information in a video image in a prior implementation
  • FIG. 7 is a schematic diagram of displaying barrographic information in a video image according to an embodiment of the present application.
  • FIG. 8 is a schematic structural diagram of an apparatus applied to the above technology according to the present application.
  • FIG. 9 is a schematic diagram of a hardware architecture composition for displaying information on a video image in a specific embodiment of the present application.
  • the embodiment of the present application provides a method for displaying information on a video image, which is applied to a playback device, such as an application for playing a media file (also called a media player) running on a user terminal such as a mobile phone or a personal computer.
  • a playback device such as an application for playing a media file (also called a media player) running on a user terminal such as a mobile phone or a personal computer.
  • the playback device parses the image frame in the currently played video, and recognizes the facial feature of the frame image, and displays the received information at a position other than the position of the facial feature. on.
  • the ability to avoid facial feature display information on the video image improves the video image processing function of the playback device.
  • the facial feature in the embodiment of the present application may be a human face, or may be an animal's head or the like.
  • the facial feature may be preset as a key information according to actual needs, and if the subtitle information is not blocked, the subtitle information may be Set to the corresponding key information.
  • the subtitle information may be Set to the corresponding key information.
  • the information received by the playback device may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
  • subtitle information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
  • advertisement information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
  • picture information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
  • FIG. 1 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
  • Step 101 When the playback device receives the information play request, parsing the image frame of the currently played video.
  • Step 102 The playing device identifies the facial feature in the image frame obtained by the parsing.
  • the image recognition method is pre-configured according to the feature of the facial feature to be identified. For example, when the facial feature is a human face, the facial recognition method is used to identify the facial feature.
  • step 103 the playback device determines the location of the recognized facial feature in the image frame.
  • the position of the facial feature identified in this step in the frame image is identified by the region corresponding to the pixel point of the edge position of the facial feature in the frame image.
  • the image of this interval corresponds to the facial feature.
  • the representation of the location is only an example.
  • the identification of the location mode is not limited.
  • an area identifier corresponding to a plurality of pixel points on the horizontal and vertical coordinates may be used.
  • step 102 and step 103 may be performed on the playback device, or a server may be added to implement the function, and the implementation of the function may be added to the existing server.
  • step 104 the playback device displays the received information in a location other than the determined location of the facial feature in the image frame.
  • the specific manner in which the playback device displays the received information in the image frame may be displayed according to the characteristics of the information. If the received information is the barrage information, the barrage information may be drawn at a position other than the facial feature. Outside the location.
  • the playback device displays the received information on the video image, it is inconsistent with the existing implementation, that is, to avoid the location of the recognized facial feature, and display the received information.
  • the embodiment of the present application is not limited, and may be displayed in a horizontal direction, a vertical display, or the like, and displayed according to actual needs or aesthetics.
  • FIG. 2 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
  • Step 201 When the playback device receives the information play request, parsing the image frame of the currently played video.
  • Step 202 The playback device determines whether the local feature of the image frame is recorded locally. If yes, go to step 205; otherwise, go to step 203.
  • Step 203 The playing device identifies the facial feature in the image frame obtained by the parsing.
  • the received information can be displayed at any position on the frame image, and The received information is displayed in the same way.
  • Step 204 The playback device determines the location of the recognized facial feature in the image frame and records the location of the facial feature of the image frame.
  • Step 205 The playback device displays the received information in a position other than the location of the facial feature in the image frame.
  • the position of the recognized facial feature in the image frame is recorded, and when the frame image is played, when other information is received again, it is not necessary to By performing facial feature recognition, the position of the facial features for the image frame can be directly obtained, which can save time and equipment resources.
  • FIG. 3 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
  • Step 301 When receiving the information play request, the playback device parses the image frame of the currently played video, and records the information for the image frame.
  • Step 302 The playback device identifies the facial feature in the image frame obtained by the parsing.
  • step 303 the playback device determines the location of the recognized facial feature in the image frame.
  • Step 304 The playback device determines whether information other than the received information is recorded for the image frame, and if yes, step 305 is performed; otherwise, step 306 is performed.
  • Step 305 The playback device displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame.
  • Step 306 the playing device displays the received information in a position other than the determined position of the facial feature in the image frame.
  • the received information is recorded for an image frame, and when the image frame is played again, if the information play request is received again, all the information recorded for the image frame is displayed on the image frame except the face.
  • FIG. 4 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
  • Step 401 When receiving the information play request, the playback device parses the image frame of the currently played video, and records the information for the image frame.
  • Step 402 The playing device determines whether the location of the facial feature of the image frame is recorded locally, and if yes, step 405 is performed; otherwise, step 403 is performed.
  • Step 403 The playing device identifies the facial feature in the image frame obtained by the parsing.
  • Step 404 the playing device determines the position of the recognized facial feature in the image frame, and records the position of the facial feature of the image frame.
  • Step 405 The playback device determines whether information other than the received information is recorded for the image frame, and if yes, step 406 is performed; otherwise, step 407 is performed.
  • Step 406 The playing device displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame, and ends the flow.
  • step 407 the playback device displays the received information in a position other than the determined position of the facial feature in the image frame.
  • This embodiment not only saves facial feature recognition time, device resources, but also increases user experience.
  • FIG. 5 is a schematic diagram of a video playing system according to an embodiment of the present application.
  • the video storage server provides video data for the playback device;
  • the information server provides the playback device. Seek information for playback.
  • the video storage server and the information server may be one server or multiple servers. When the embodiment of the present application is specifically implemented, the present invention is not limited thereto, and may be implemented according to the existing implementation.
  • the video storage server sends the video to be played to the playback device.
  • the playback device When the playback device receives the video sent by the video storage server, if it is determined that the video is audio and video, the video data corresponding to the received video is first separated into audio data and video data, and the audio data is decoded to output a sound; After decoding, the video image is output. Regardless of whether or not the received information is displayed, the synchronization of the sound and the image is not changed here, as in the prior implementation.
  • the information server when the information server receives the barrage information sent by the user through the client, the information server sends the barrage information to the playback device.
  • the playback device When the playback device receives the barrage information sent by the barrage server, it parses the image frame of the currently played video; and records the received barrage information for the image frame.
  • the playback device determines whether the location of the facial feature on the image frame is recorded for the image frame, and if so, the playback device determines whether information other than the received information is recorded for the image frame.
  • the playback device When the playback device records the location of the facial features on the image frame, the corresponding barrographic information to be recorded for that image frame is displayed at a location other than the location of the facial features.
  • the playback device does not record the location of the facial feature on the image frame
  • the facial feature is identified in the parsed image frame; the position of the identified facial feature in the image frame is determined; and the facial feature on the image frame is recorded s position.
  • the corresponding barrographic information to be recorded for the image frame is then displayed at a location other than the location of the facial feature.
  • FIG. 6 is a schematic diagram of displaying barrographic information in a video image in a prior implementation.
  • FIG. 7 is a schematic diagram of displaying barrographic information in a video image according to an embodiment of the present application.
  • FIG 6 and 7 are schematic diagrams showing the same barr information for the same frame video image.
  • the facial feature is an example of face information.
  • FIG. 8 is a schematic structural diagram of an apparatus applied to the above technology according to the present application.
  • the device includes: a receiving unit 801, a parsing unit 802, an identifying unit 803, a determining unit 804, and a display unit 805;
  • the receiving unit 801 is configured to receive an information play request.
  • the parsing unit 802 is configured to parse the image frame of the currently played video when the receiving unit 801 receives the information playing request;
  • the identifying unit 803 is configured to identify a facial feature in the image frame obtained by the parsing unit 802;
  • a determining unit 804 configured to determine a location of the facial feature identified by the identifying unit 803 in the image frame
  • the display unit 805 is configured to display the information received by the receiving unit 801 in a position other than the position of the facial feature determined by the determining unit 804 in the image frame.
  • the device further comprises: a recording unit 806;
  • a recording unit 806, configured to: when the determining unit 804 determines the location of the recognized facial feature in the image frame, record the location of the facial feature of the image frame;
  • the determining unit 804 is further configured to: after the parsing unit 802 parses the image frame of the currently playing video, determine whether the recording unit records the location of the facial feature of the image frame, and if so, the trigger display unit 805 displays the received information in the A position other than the determined position of the facial feature in the image frame; otherwise, the trigger recognition unit 803 recognizes the facial feature in the image frame obtained by the analysis.
  • the device further includes: a recording unit 806;
  • the recording unit 806 is further configured to: when the parsing unit 802 parses the image frame of the currently played video, record the information for the image frame;
  • the display unit 805 is further configured to: before displaying the received information in a position other than the determined position of the facial feature in the image frame, determining whether the recording unit records the received information for the image frame The information, if so, the trigger display unit 805 displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame; otherwise, the trigger display unit 805 displays the received information at The position of the image frame other than the position of the determined facial feature.
  • the position of the facial feature in the frame image is identified using an area corresponding to a pixel point of an edge position of the facial feature in the frame image.
  • the received information is one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
  • the units of the above embodiments may be integrated into one, or may be deployed separately; may be combined into one unit, or may be further split into multiple sub-units.
  • a hardware module can include specially designed permanent circuits or logic devices (such as dedicated processors such as FPGAs or ASICs) for performing specific operations.
  • Hardware modules can also be included by Programmable logic devices or circuits (such as general purpose processors or other programmable processors) that are temporarily configured by software are used to perform specific operations.
  • Hardware implementations can be made based on cost and time considerations, either by mechanical means, by dedicated permanent circuits, or by temporarily configured circuits (as configured by software).
  • FIG. 9 is a schematic structural diagram of hardware architecture for displaying information on a video image in a specific embodiment of the present application.
  • the playback device can include a processor 910, a memory 920, a port 930, and a bus 940.
  • Processor 910 and memory 920 are interconnected by a bus 940.
  • the processor 910 can acquire and output data through the port 1130;
  • the receiving unit 801 When the receiving unit 801 is executed by the processor 910, it may be: receiving an information play request;
  • parsing unit 802 When the parsing unit 802 is executed by the processor 910, it may be: parsing an image frame of the currently played video;
  • the identification unit 803, when executed by the processor 910, may be: identifying a facial feature in an image frame;
  • the determining unit 804 when executed by the processor 910, may be: determining a location of the recognized facial feature in the image frame;
  • embodiments of the present application may be implemented by a data processing program executed by a device such as a computer.
  • the data processing program constitutes the present application.
  • a data processing program that is usually stored in one storage medium is executed by directly reading the program out of the storage medium or by installing or copying the program to a storage device (such as a hard disk and/or a memory) of the data processing device. Therefore, such a storage medium also constitutes the present application.
  • the storage medium can use any type of recording method, such as paper storage medium (such as paper tape, etc.), magnetic storage medium (such as floppy disk, hard disk, flash memory, etc.), optical storage medium (such as CD-ROM, etc.), magneto-optical storage medium ( Such as MO, etc.).
  • the present application also discloses a storage medium, such as a non-volatile computer storage medium, in which a data processing program is stored, which is used to execute the above technical solution of the present application.
  • the playback device shown in FIG. 9 is only a specific example, and may be implemented by other structures different from those described in the embodiment, for example, operations performed when the above instruction code is executed, or may be performed by a specific application.
  • the above-mentioned processor 910 may be one or more, and if there are multiple, the plurality of processors are jointly responsible for reading and executing the instruction code. Therefore, the specific structure of the test device is not specifically limited in the present application.
  • the present application when receiving an information play request by the playback device, displays the received information on the currently played image frame while avoiding the facial feature.
  • the realization of the scheme can avoid the facial feature display information on the video image and improve the video image processing function of the playback device.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Marketing (AREA)
  • Business, Economics & Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

本申请提供了一种在视频图像上显示信息的方法和装置,应用于播放设备上,该播放设备接收信息播放请求;解析当前播放视频的图像帧;在解析获得的图像帧中识别出面部特征;确定识别出的面部特征在该图像帧中的位置;将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。

Description

一种在视频图像上显示信息的方法和装置
本申请要求于2015年11月16日提交中国专利局、申请号为201510751308.4、发明名称为“一种在视频图像上显示信息的方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及视频技术领域,特别涉及一种在视频图像上显示信息的方法和装置。
发明背景
弹幕指直接显现在视频上的评论,可以以滚动、停留甚至更多动作特效方式出现在视频上,是观看视频的人发送的简短评论。弹幕视频即带有“弹幕”的视频。目前很多网站提供视频发送弹幕的功能,如niconico、acfun、bilibili、dilili、tucao、爆点TV等。
一般情况,该类网站能允许观看视频者发表评论或感想,但与普通视频分享网站只在播放器下专用点评区显示不同,其会以滑动字幕的方式实时出现在视频画面上,保证所有观看者都能注意到,从而实现观看者间的互动,甚至可以一起表达对作品的赞叹或批评,增加观看乐趣。
现有技术中,由于弹幕信息随机出现在视频画面上,即会出现在视频画面的任意位置,这样会存在弹幕遮挡视频重要播放内容的弊病,如遮挡视频图像中的人脸等关键信息。
发明内容
有鉴于此,本申请提供一种在视频图像上显示信息的方法和装置,能够避开视频图像上的面部特征显示信息,提高了播放设备视频图像处理功能。
为解决上述技术问题,本申请的技术方案是这样实现的:
一种在视频图像上显示信息的方法,应用于播放设备上,该方法包 括:
该播放设备接收信息播放请求;
解析当前播放视频的图像帧;
在解析获得的图像帧中识别出面部特征;
确定识别出的面部特征在该图像帧中的位置;
将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
一种在视频图像上显示信息的装置,应用于播放设备上,该装置至少包括处理器和存储器,其中所述存储器中存储接收单元、解析单元、获取单元、确定单元和显示单元,当被所述处理器执行时,
所述接收单元用于接收信息播放请求;
所述解析单元用于解析当前播放视频的图像帧;
所述识别单元用于在解析获得的图像帧中识别出面部特征;
所述确定单元用于确定识别出的面部特征在该图像帧中的位置;
所述显示单元用于将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
一种非易失性计算机存储介质,其中存储有计算机程序,该计算机程序用于执行上述的方法。
附图简要说明
图1为本申请一个实施例中在视频图像上显示信息的流程示意图;
图2为本申请一个实施例中在视频图像上显示信息的流程示意图;
图3为本申请一个实施例中在视频图像上显示信息的流程示意图;
图4为本申请一个实施例中在视频图像上显示信息的流程示意图;
图5为本申请实施例中视频播放系统示意图;
图6为现有实现中在视频图像中显示弹幕信息的示意图;
图7为本申请实施例中在视频图像中显示弹幕信息的示意图。
图8为本申请应用于上述技术的装置结构示意图;
图9为本申请具体实施例中在视频图像上显示信息的硬件架构组成示意图。
实施方式
为了使本申请的目的、技术方案及优点更加清楚明白,下面结合附图并举实施例,对本申请的技术方案进行详细说明。
本申请实施例中提供一种在视频图像上显示信息的方法,应用于播放设备上,如应用于诸如手机、个人电脑等用户终端上运行的用于播放媒体文件的应用(也叫媒体播放器);该播放设备在接收到信息播放请求时,在当前播放的视频中解析出图像帧,并识别出该帧图像的面部特征,将接收到的信息显示在除面部特征的位置之外的位置上。能够避开视频图像上的面部特征显示信息,提高了播放设备视频图像处理功能。
本申请实施例中的面部特征可以为人脸,也可以为动物的头部等,具体实现时,可以根据实际需要预先设置该面部特征为关键信息,如希望不遮挡字幕信息,则可以将字幕信息设置为对应的关键信息。上述只是一种举例,并不限于上述提到的面部特征。
播放设备接收到的信息可以为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。本申请实施例具体实现时,不限于上述列举信息。
下面结合附图,详细说明本申请实施例中在视频图像上显示信息的过程。
参见图1,图1为本申请一个实施例中在视频图像上显示信息的流程示意图。具体步骤为:
步骤101,播放设备接收到信息播放请求时,解析当前播放视频的图像帧。
步骤102,该播放设备在解析获得的图像帧中识别出面部特征。
本申请实施例中根据要识别的面部特征的特点预先配置图像识别方法,如面部特征为人脸时,则使用人脸识别方法进行面部特征的识别。
步骤103,该播放设备确定识别出的面部特征在该图像帧中的位置。
本步骤中识别出的面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
如针对x轴为第n像素到第m像素,针对y轴为第p像素到第q像素,这一区间的图像对应面部特征。这里位置的表示方式只是一种举例,具体实现时,对位置方式的标识不作限制。在具体实现时,为了更准确确定面部特征的位置,可以使用横纵坐标多个像素点所对应的区域标识。
本申请在具体实现时,步骤102和步骤103的实现可以在播放设备上,也可以增加一台服务器实现该功能,还可以在已有的服务器上增加该功能的实现。
步骤104,该播放设备将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
该播放设备将接收到的信息显示在图像帧中的具体方式,可以根据所述信息的特点进行显示,如接收到的信息为弹幕信息,则可以将弹幕信息绘制在除面部特征的位置之外的位置上。
播放设备在视频图像上显示接收到的信息时,同现有实现不一致的地方,即为避开识别出的面部特征的位置,显示接收到的信息。至于接收到的信息的具体显示方式,本申请实施例不作限制,可以横向显示,纵向显示等,根据实际需要,或者美观来显示。
参见图2,图2为本申请一个实施例中在视频图像上显示信息的流程示意图。具体步骤为:
步骤201,播放设备接收到信息播放请求时,解析当前播放视频的图像帧。
步骤202,该播放设备确定本地是否记录该图像帧的面部特征的位 置,如果是,执行步骤205;否则,执行步骤203。
步骤203,该播放设备在解析获得的图像帧中识别出面部特征。
在具体实现时,若在解析获得的图像中未识别出面部特征,则针对该图像帧记录无面部特征,则在显示接收到的信息可以在该帧图像上任意位置显示,即可和现有实现方式相同的方式显示接收到的信息。
步骤204,该播放设备确定识别出的面部特征在该图像帧中的位置,并记录该图像帧的面部特征的位置。
步骤205,该播放设备将接收到的信息显示在该图像帧中除面部特征的位置之外的位置。
本实施例中,在首次针对一图像帧进行面部特征识别时,将识别出的面部特征在该图像帧中的位置进行记录,当播放该帧图像时,再次接收到其他信息时,不需要再次进行面部特征识别,直接获取针对该图像帧的面部特征的位置,即可,能够节省时间,以及设备资源。
参见图3,图3为本申请一个实施例中在视频图像上显示信息的流程示意图。具体步骤为:
步骤301,播放设备接收到信息播放请求时,解析当前播放视频的图像帧,并针对该图像帧记录所述信息。
步骤302,该播放设备在解析获得的图像帧中识别出面部特征。
步骤303,该播放设备确定识别出的面部特征在该图像帧中的位置。
步骤304,该播放设备确定针对该图像帧是否记录接收的信息之外的信息,如果是,执行步骤305;否则,执行步骤306。
步骤305,该播放设备将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
步骤306,该播放设备将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
本实施例中,针对一图像帧记录接收到的信息,再次在播放该图像帧时,若再次接收到信息播放请求,则将针对该图像帧记录的所有信息都显示在该图像帧上除面部特征的位置之外的位置。这样的技术方案,能够增加用户体验,如针对一个视频上的弹幕信息,不同的用户在观看过程中针对某一帧图像播放时,显示在该帧图像发送的所有弹幕。
参见图4,图4为本申请一个实施例中在视频图像上显示信息的流程示意图。具体步骤为:
步骤401,播放设备接收到信息播放请求时,解析当前播放视频的图像帧,并针对该图像帧记录所述信息。
步骤402,该播放设备确定本地是否记录该图像帧的面部特征的位置,如果是,执行步骤405;否则,执行步骤403。
步骤403,该播放设备在解析获得的图像帧中识别出面部特征。
步骤404,该播放设备确定识别出的面部特征在该图像帧中的位置,并记录该图像帧的面部特征的位置。
步骤405,该播放设备确定针对该图像帧是否记录接收的信息之外的信息,如果是,执行步骤406;否则,执行步骤407。
步骤406,该播放设备将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置,结束本流程。
步骤407,该播放设备将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
本实施例既节省面部特征识别时间,设备资源,又能增加用户体验。
下面以接收到的信息为弹幕信息为例,结合附图详细说明在视频图像上显示信息的过程。
参见图5,图5为本申请实施例中视频播放系统示意图。图5中视频存储服务器为播放设备提供视频数据;信息服务器为播放设备提供请 求播放的信息。在具体实现时,视频存储服务器和信息服务器可以为一台服务器,也可以为多台服务器,本申请实施例具体实现时,对此不作限制,按现有实现即可。
视频存储服务器将要播放的视频发送给播放设备。
播放设备接收到视频存储服务器发送的视频时,若确定该视频为音视频,需先将接收到的视频对应的视频数据分离为音频数据和视频数据,对音频数据解码后输出声音;对视频数据解码后,输出视频图像。无论是否显示接收到的信息,声音和图像的同步这里不作改变,按现有实现方式即可。
本申请实施例中,信息服务器接收到用户通过客户端发送的弹幕信息时,将该弹幕信息发送给播放设备。
播放设备接收到弹幕服务器发送的弹幕信息时,解析当前播放视频的图像帧;并针对该图像帧记录接收到的弹幕信息。
播放设备确定针对该图像帧是否记录该图像帧上的面部特征的位置,如果是,则该播放设备确定针对该图像帧是否记录接收的信息之外的信息。
当播放设备记录该图像帧上的面部特征的位置,则将针对该图像帧将记录的对应弹幕信息都显示在除面部特征的位置之外的位置。
当播放设备未记录该图像帧上的面部特征的位置,则在解析获得的图像帧中识别出面部特征;确定识别出的面部特征在该图像帧中的位置;并记录该图像帧上面部特征的位置。然后将针对该图像帧将记录的对应弹幕信息都显示在除面部特征的位置之外的位置。
这只是列举的一个在准备播放的视频图像上显示弹幕信息的例子,本申请具体实现时,对此不作限制,按现有方式实现确定在准备播放的视频图像帧上是否需要显示弹幕信息。
参见图6,图6为现有实现中在视频图像中显示弹幕信息的示意图。
参见图7,图7为本申请实施例中在视频图像中显示弹幕信息的示意图。
图6和图7为针对同一帧视频图像显示相同的弹幕信息的示意图。面部特征为人脸信息为例。
由图6和图7可见,现有实现过程中,图6中的弹幕信息(战争一触即发、国产大英雄和666666666666)会显示在人脸上,不利于用户对视频的观看;而通过本申请实施例的实现方式处理后,图7中的上述弹幕信息不会显示在人脸上,提高了用户观看视频的体验;因此,本申请提供的技术方案实现了不影响用户观影体验的信息显示,提高了播放设备视频图像处理功能。
基于同样的发明构思,本申请还提出一种在视频图像上显示信息的装置,应用于播放设备上。参见图8,图8为本申请应用于上述技术的装置结构示意图。该装置包括:接收单元801、解析单元802、识别单元803、确定单元804和显示单元805;
接收单元801,用于接收信息播放请求;
解析单元802,用于当接收单元801接收到信息播放请求时,解析当前播放视频的图像帧;
识别单元803,用于在解析单元802解析获得的图像帧中识别出面部特征;
确定单元804,用于确定识别单元803识别出的面部特征在该图像帧中的位置;
显示单元805,用于将接收单元801接收到的信息显示在该图像帧中除确定单元804确定出的面部特征的位置之外的位置。
较佳地,该装置进一步包括:记录单元806;
记录单元806,用于当确定单元804确定识别出的面部特征在该图像帧中的位置时,记录该图像帧的面部特征的位置;
确定单元804,进一步用于当解析单元802解析当前播放视频的图像帧之后,确定所述记录单元是否记录该图像帧的面部特征的位置,如果是,触发显示单元805将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触发识别单元803在解析获得的图像帧中识别出面部特征。
该装置进一步包括:记录单元806;
记录单元806,进一步用于当解析单元802解析当前播放视频的图像帧时,针对该图像帧记录所述信息;
显示单元805,进一步用于当将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置之前,确定所述记录单元中针对该图像帧是否记录接收的信息之外的信息,如果是,触发显示单元805将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触发显示单元805将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
所述面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
所述接收到的信息为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。
上述实施例的单元可以集成于一体,也可以分离部署;可以合并为一个单元,也可以进一步拆分成多个子单元。
本申请各实施例中的各单元可以以机械方式或电子方式实现。例如,一个硬件模块可以包括专门设计的永久性电路或逻辑器件(如专用处理器,如FPGA或ASIC)用于完成特定的操作。硬件模块也可以包括由 软件临时配置的可编程逻辑器件或电路(如包括通用处理器或其它可编程处理器)用于执行特定操作。至于具体采用机械方式,或是采用专用的永久性电路,或是采用临时配置的电路(如由软件进行配置)来实现硬件模块,可以根据成本和时间上的考虑来决定。
以上对本申请具体实施例中的在视频图像上显示弹幕的装置进行了说明,下面给出本申请具体实施例中播放设备的硬件架构组成,该设备是可以软硬件结合的可编程设备,具体参见图9,图9为本申请具体实施例中在视频图像上显示信息的硬件架构组成示意图。该播放设备可包括:处理器910,存储器920,端口930以及总线940。处理器910和存储器920通过总线940互联。处理器910可通过端口1130获取和输出数据;其中,
接收单元801被处理器910执行时可以为:接收信息播放请求;
解析单元802被处理器910执行时可以为:解析当前播放视频的图像帧;
识别单元803被处理器910执行时可以为:在图像帧中识别出面部特征;
确定单元804被处理器910执行时可以为:确定识别出的面部特征在该图像帧中的位置;
显示单元805被处理器910执行时可以为:将接收到的信息显示在该图像帧中除确定单元804确定出的面部特征的位置之外的位置。
记录单元806被处理器910执行时可以为:记录接收到的信息,记录图像帧中面部特征的位置。
由此可以看出,当存储在存储器920中的指令模块被处理器910执行时,可实现前述技术方案中接收单元、解析单元、识别单元、确定单元、显示单元和记录单元的各种功能。
另外,本申请的实施例中可以通过由设备如计算机执行的数据处理程序来实现。显然,数据处理程序构成了本申请。此外,通常存储在一个存储介质中的数据处理程序通过直接将程序读取出存储介质或者通过将程序安装或复制到数据处理设备的存储设备(如硬盘和/或内存)中执行。因此,这样的存储介质也构成了本申请。存储介质可以使用任何类型的记录方式,例如纸张存储介质(如纸带等)、磁存储介质(如软盘、硬盘、闪存等)、光存储介质(如CD-ROM等)、磁光存储介质(如MO等)等。
因此,本申请还公开了一种存储介质,比如非易失性计算机存储介质,其中存储有数据处理程序,该数据处理程序用于执行本申请上述技术方案。
需要说明的是,图9所示的播放设备只是一个具体的例子,也可以通过其他的与本实施例描述不同的结构实现,例如,执行上述指令代码时所完成的操作,也可以由特定应用专用集成电路(ASIC)实现。另外,上述的处理器910可以是一个或多个,如果是多个,则由多个处理器共同负责读取和执行所述指令代码。因此,本申请对测试设备的具体结构不作具体限定。
综上所述,本申请通过播放设备在接收到信息播放请求时,在当前播放的图像帧上避开面部特征显示接收到的信息。该方案的实现,能够避开视频图像上的面部特征显示信息,提高了播放设备视频图像处理功能。
以上所述仅为本申请的较佳实施例而已,并不用以限制本申请,凡在本申请的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本申请保护的范围之内。

Claims (11)

  1. 一种在视频图像上显示信息的方法,应用于播放设备上,其特征在于,该方法包括:
    该播放设备接收信息播放请求;
    解析当前播放视频的图像帧;
    在解析获得的图像帧中识别出面部特征;
    确定识别出的面部特征在该图像帧中的位置;
    将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
  2. 根据权利要求1所述的方法,其特征在于,所述确定识别出的面部特征在该图像帧中的位置时,所述方法进一步包括:记录该图像帧的面部特征的位置;
    所述解析当前播放视频的图像帧之后,所述在解析获得的图像帧中识别出面部特征之前,所述方法进一步包括:
    确定本地是否记录该图像帧的面部特征的位置,如果是,将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,执行所述在解析获得的图像帧中识别出面部特征步骤。
  3. 根据权利要求1或2所述的方法,其特征在于,所述解析当前播放视频的图像帧时,所述方法进一步包括:针对该图像帧记录所述信息;
    所述将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置之前,所述方法进一步包括:
    确定针对该图像帧是否记录接收的信息之外的信息,如果是,将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,将接收到的信息显示在该图像帧中除确定出的面部 特征的位置之外的位置。
  4. 根据权利要求1或2所述的方法,其特征在于,
    所述面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
  5. 根据权利要求1或2所述的方法,其特征在于,
    所述接收到的信息为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。
  6. 一种在视频图像上显示信息的装置,应用于播放设备上,其特征在于,该装置至少包括处理器和存储器,其中所述存储器中存储接收单元、解析单元、识别单元、确定单元和显示单元,当被所述处理器执行时,
    所述接收单元用于接收信息播放请求;
    所述解析单元用于解析当前播放视频的图像帧;
    所述识别单元用于在所述图像帧中识别出面部特征;
    所述确定单元用于确定所述识别出的面部特征在该图像帧中的位置;
    所述显示单元用于将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
  7. 根据权利要求6所述的装置,其特征在于,该存储器中进一步存储记录单元,当被所述处理器执行时,
    所述记录单元用于根据确定识别出的面部特征在该图像帧中的位置,记录该图像帧的面部特征的位置;
    所述确定单元进一步用于当解析当前播放视频的图像帧之后,确定是否已经记录该图像帧的面部特征的位置,如果是,触发将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触 发在解析获得的图像帧中识别出面部特征。
  8. 根据权利要求6或7所述的装置,其特征在于,该存储器中进一步存储记录单元,当被所述处理器执行时,
    所述记录单元进一步用于当解析当前播放视频的图像帧时,针对该图像帧记录所述信息;
    所述显示单元进一步用于将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置之前,确定针对该图像帧是否记录接收的信息之外的信息,如果是,触发将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触发将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
  9. 根据权利要求6或7所述的装置,其特征在于,
    所述面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
  10. 根据权利要求6或7所述的装置,其特征在于,
    所述接收到的信息为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。
  11. 一种非易失性计算机存储介质,其特征在于,其中存储有计算机程序,该计算机程序用于执行所述权利要求1至5任一项所述的方法。
PCT/CN2016/102640 2015-11-06 2016-10-20 一种在视频图像上显示信息的方法和装置 WO2017076171A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2017560519A JP6438598B2 (ja) 2015-11-06 2016-10-20 ビデオ画像の上に情報を表示するための方法及びデバイス
US15/787,908 US10425679B2 (en) 2015-11-06 2017-10-19 Method and device for displaying information on video image

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510751308.4 2015-11-06
CN201510751308.4A CN105430512A (zh) 2015-11-06 2015-11-06 一种在视频图像上显示信息的方法和装置

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/787,908 Continuation US10425679B2 (en) 2015-11-06 2017-10-19 Method and device for displaying information on video image

Publications (1)

Publication Number Publication Date
WO2017076171A1 true WO2017076171A1 (zh) 2017-05-11

Family

ID=55508395

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/102640 WO2017076171A1 (zh) 2015-11-06 2016-10-20 一种在视频图像上显示信息的方法和装置

Country Status (4)

Country Link
US (1) US10425679B2 (zh)
JP (1) JP6438598B2 (zh)
CN (1) CN105430512A (zh)
WO (1) WO2017076171A1 (zh)

Families Citing this family (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8721571B2 (en) 2010-11-22 2014-05-13 Siwa Corporation Selective removal of cells having accumulated agents
US10358502B2 (en) 2014-12-18 2019-07-23 Siwa Corporation Product and method for treating sarcopenia
CN105430512A (zh) 2015-11-06 2016-03-23 腾讯科技(北京)有限公司 一种在视频图像上显示信息的方法和装置
CN105635848A (zh) * 2015-12-24 2016-06-01 深圳市金立通信设备有限公司 一种弹幕显示方法及终端
CN106210855B (zh) * 2016-07-11 2019-12-13 网易(杭州)网络有限公司 对象显示方法和装置
CN106303731A (zh) * 2016-08-01 2017-01-04 北京奇虎科技有限公司 弹幕的显示方法及装置
CN108124185B (zh) * 2016-11-28 2019-06-21 广州华多网络科技有限公司 一种弹幕显示方法、装置及终端
US10858449B1 (en) 2017-01-06 2020-12-08 Siwa Corporation Methods and compositions for treating osteoarthritis
US10925937B1 (en) 2017-01-06 2021-02-23 Siwa Corporation Vaccines for use in treating juvenile disorders associated with inflammation
CN107181976B (zh) * 2017-04-28 2021-01-29 华为技术有限公司 一种弹幕显示方法及电子设备
CN107147941A (zh) * 2017-05-27 2017-09-08 努比亚技术有限公司 视频播放的弹幕显示方法、装置及计算机可读存储介质
CN107454255B (zh) * 2017-07-28 2020-07-17 维沃移动通信有限公司 一种歌词显示方法、移动终端及计算机可读存储介质
CN107943964A (zh) * 2017-11-27 2018-04-20 腾讯音乐娱乐科技(深圳)有限公司 歌词显示方法、装置及计算机可读存储介质
CN107948760B (zh) * 2017-11-30 2021-01-29 上海哔哩哔哩科技有限公司 弹幕播放控制方法、服务器及弹幕播放控制系统
US11518801B1 (en) 2017-12-22 2022-12-06 Siwa Corporation Methods and compositions for treating diabetes and diabetic complications
CN109302619A (zh) * 2018-09-18 2019-02-01 北京奇艺世纪科技有限公司 一种信息处理方法及装置
CN109618213B (zh) * 2018-12-17 2020-05-19 华中科技大学 一种防止弹幕遮挡目标对象的方法
CN109905775A (zh) * 2019-01-16 2019-06-18 北京奇艺世纪科技有限公司 一种涂鸦弹幕生成及显示方法、装置、终端设备及存储介质
CN109905757A (zh) * 2019-02-13 2019-06-18 安徽励图信息科技股份有限公司 通过人脸识别控制视频字幕播出的方法
CN109862414B (zh) * 2019-03-22 2021-10-15 武汉斗鱼鱼乐网络科技有限公司 一种蒙版弹幕显示方法、装置及服务器
CN116916080A (zh) * 2019-05-17 2023-10-20 上海哔哩哔哩科技有限公司 视频数据处理方法、装置、计算机设备及可读存储介质
CN110362373A (zh) * 2019-06-29 2019-10-22 华为技术有限公司 一种控制屏幕小窗口的方法及相关设备
CN110351596B (zh) * 2019-07-17 2021-07-27 上海播呗网络科技有限公司 一种互联网流媒体大数据弹幕处理系统及处理方法
CN110636259B (zh) * 2019-09-10 2021-03-02 杭州亿圣信息技术有限公司 一种基于视频检测的osd叠加方法
CN110958480A (zh) * 2019-11-28 2020-04-03 湖南快乐阳光互动娱乐传媒有限公司 一种弹幕显示方法
CN113593567B (zh) * 2021-06-23 2022-09-09 荣耀终端有限公司 视频声音转文本的方法及相关设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1423896A (zh) * 2000-12-15 2003-06-11 皇家菲利浦电子有限公司 基于视频内容分析的画中画重新定位和/或大小调整
JP2004128614A (ja) * 2002-09-30 2004-04-22 Toshiba Corp 画像表示制御装置及び画像表示制御プログラム
CN102845067A (zh) * 2010-04-01 2012-12-26 汤姆森许可贸易公司 三维(3d)呈现中的字幕
US20140196082A1 (en) * 2012-07-17 2014-07-10 Panasonic Corporation Comment information generating apparatus and comment information generating method
CN105430512A (zh) * 2015-11-06 2016-03-23 腾讯科技(北京)有限公司 一种在视频图像上显示信息的方法和装置

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7184100B1 (en) * 1999-03-24 2007-02-27 Mate - Media Access Technologies Ltd. Method of selecting key-frames from a video sequence
US6778224B2 (en) * 2001-06-25 2004-08-17 Koninklijke Philips Electronics N.V. Adaptive overlay element placement in video
JP2003271954A (ja) * 2002-03-12 2003-09-26 Canon I-Tech Inc 画像処理装置、画像処理方法、プログラムおよび記録媒体
JP2007251273A (ja) * 2006-03-13 2007-09-27 Oki Electric Ind Co Ltd 画像処理装置,画像伝送システムおよび画像処理方法
JP4507281B2 (ja) * 2006-03-30 2010-07-21 富士フイルム株式会社 画像表示装置、撮像装置および画像表示方法
US20120323704A1 (en) * 2008-02-27 2012-12-20 Chad Steelberg Enhanced world wide web-based communications
JP2010226351A (ja) * 2009-03-23 2010-10-07 Fujitsu Ten Ltd 映像再生装置
JP5371574B2 (ja) * 2009-06-23 2013-12-18 株式会社第一興商 背景映像中の顔画像を避けるように歌詞字幕を表示するカラオケ装置
JP5465620B2 (ja) * 2010-06-25 2014-04-09 Kddi株式会社 映像コンテンツに重畳する付加情報の領域を決定する映像出力装置、プログラム及び方法
JP2013247611A (ja) * 2012-05-29 2013-12-09 Nippon Telegr & Teleph Corp <Ntt> 字幕放送出力装置、字幕放送出力方法及び字幕放送出力プログラム
US9124765B2 (en) * 2012-12-27 2015-09-01 Futurewei Technologies, Inc. Method and apparatus for performing a video conference

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1423896A (zh) * 2000-12-15 2003-06-11 皇家菲利浦电子有限公司 基于视频内容分析的画中画重新定位和/或大小调整
JP2004128614A (ja) * 2002-09-30 2004-04-22 Toshiba Corp 画像表示制御装置及び画像表示制御プログラム
CN102845067A (zh) * 2010-04-01 2012-12-26 汤姆森许可贸易公司 三维(3d)呈现中的字幕
US20140196082A1 (en) * 2012-07-17 2014-07-10 Panasonic Corporation Comment information generating apparatus and comment information generating method
CN105430512A (zh) * 2015-11-06 2016-03-23 腾讯科技(北京)有限公司 一种在视频图像上显示信息的方法和装置

Also Published As

Publication number Publication date
JP6438598B2 (ja) 2018-12-19
JP2018519727A (ja) 2018-07-19
US20180041796A1 (en) 2018-02-08
US10425679B2 (en) 2019-09-24
CN105430512A (zh) 2016-03-23

Similar Documents

Publication Publication Date Title
WO2017076171A1 (zh) 一种在视频图像上显示信息的方法和装置
WO2019109643A1 (zh) 视频推荐方法、装置、计算机设备和存储介质
CN109089154B (zh) 一种视频提取方法、装置、设备及介质
CN109089127B (zh) 一种视频拼接方法、装置、设备及介质
WO2017092343A1 (zh) 一种视频数据的检测方法和装置
US20160307371A1 (en) Synchronizing an augmented reality video stream with a displayed video stream
WO2015081776A1 (zh) 视频画面的处理方法及装置
WO2017181598A1 (zh) 视频播放方法及装置
JP4267649B2 (ja) ビデオ番組の処理方法、関連装置及び関連媒体
US20190147865A1 (en) Content recognizing method and apparatus, device, and computer storage medium
CN109729429B (zh) 视频播放方法、装置、设备和介质
JP2009177411A (ja) 電子機器および画像表示方法
CN104967903A (zh) 一种视频播放的检测方法及装置
JP2010257509A (ja) オーサリング装置、オーサリング方法およびプログラム
WO2017202060A1 (zh) 一种视频播放方法及装置
US9264646B2 (en) Electronic device and video playing method
WO2017096898A1 (zh) 广告展示方法及装置
JP2010509830A (ja) ビデオデータストリームの要約を生成するための方法及び装置
WO2019114330A1 (zh) 一种视频播放方法、装置和终端设备
CN112055254A (zh) 视频播放的方法、装置、终端及存储介质
WO2017101414A1 (zh) 基于安卓平台的多媒体播放方法、装置及移动终端设备
US20090193355A1 (en) Information processing apparatus and display control method
TWI535278B (zh) 影片播放方法以及系統
CN112188221B (zh) 播放控制方法、装置、计算机设备及存储介质
CN109840406B (zh) 活体验证方法、装置和计算机设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16861439

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2017560519

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 01.10.2018)

122 Ep: pct application non-entry in european phase

Ref document number: 16861439

Country of ref document: EP

Kind code of ref document: A1