WO2017076171A1 - 一种在视频图像上显示信息的方法和装置 - Google Patents
一种在视频图像上显示信息的方法和装置 Download PDFInfo
- Publication number
- WO2017076171A1 WO2017076171A1 PCT/CN2016/102640 CN2016102640W WO2017076171A1 WO 2017076171 A1 WO2017076171 A1 WO 2017076171A1 CN 2016102640 W CN2016102640 W CN 2016102640W WO 2017076171 A1 WO2017076171 A1 WO 2017076171A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image frame
- information
- facial feature
- image
- location
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 28
- 230000001815 facial effect Effects 0.000 claims abstract description 103
- 238000004590 computer program Methods 0.000 claims description 2
- 238000010586 diagram Methods 0.000 description 11
- 230000006870 function Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 241001465754 Metazoa Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/475—End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7837—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
- G06F16/784—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content the detected or recognised objects being people
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
- H04N21/4316—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4788—Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/812—Monomedia components thereof involving advertisement data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
- H04N21/8133—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
Definitions
- the present application relates to the field of video technologies, and in particular, to a method and apparatus for displaying information on a video image.
- the barrage refers to a comment that appears directly on the video. It can appear on the video as a scroll, pause, or even more action special effect, which is a short comment sent by the person watching the video.
- the barrage video is a video with a "barrage".
- many websites provide video transmission barrage functions, such as niconico, acfun, bilibili, dilili, tucao, and explosion point TV.
- such websites can allow viewers to watch comments or impressions, but they are different from ordinary video sharing websites only in the special comment area under the player. They will appear on the video screen in real time by sliding subtitles to ensure all viewing.
- Teen can notice that the interaction between viewers can be realized, and even the praise or criticism of the work can be expressed together to increase the fun of watching.
- the barrage information randomly appears on the video screen, it will appear at any position of the video screen, so that there is a drawback that the barrage obscures the important content of the video, such as occluding key information such as a face in the video image. .
- the present application provides a method and apparatus for displaying information on a video image, which can avoid facial feature display information on a video image, and improve a video image processing function of the playback device.
- a method for displaying information on a video image, applied to a playback device, the method package include:
- the playback device receives an information play request
- the received information is displayed at a position other than the determined position of the facial feature in the image frame.
- An apparatus for displaying information on a video image is applied to a playback device, the apparatus comprising at least a processor and a memory, wherein the memory stores a receiving unit, a parsing unit, an obtaining unit, a determining unit, and a display unit, when When the processor executes,
- the receiving unit is configured to receive an information play request
- the parsing unit is configured to parse an image frame of a currently playing video
- the identifying unit is configured to identify a facial feature in the image frame obtained by the parsing
- the determining unit is configured to determine a location of the recognized facial feature in the image frame
- the display unit is configured to display the received information in a position other than the determined position of the facial feature in the image frame.
- a non-volatile computer storage medium having stored therein a computer program for performing the method described above.
- FIG. 1 is a schematic flow chart of displaying information on a video image in an embodiment of the present application
- FIG. 2 is a schematic flow chart of displaying information on a video image in an embodiment of the present application
- FIG. 3 is a schematic flow chart of displaying information on a video image in an embodiment of the present application.
- FIG. 4 is a schematic flow chart of displaying information on a video image in an embodiment of the present application.
- FIG. 5 is a schematic diagram of a video playing system according to an embodiment of the present application.
- FIG. 6 is a schematic diagram of displaying barrographic information in a video image in a prior implementation
- FIG. 7 is a schematic diagram of displaying barrographic information in a video image according to an embodiment of the present application.
- FIG. 8 is a schematic structural diagram of an apparatus applied to the above technology according to the present application.
- FIG. 9 is a schematic diagram of a hardware architecture composition for displaying information on a video image in a specific embodiment of the present application.
- the embodiment of the present application provides a method for displaying information on a video image, which is applied to a playback device, such as an application for playing a media file (also called a media player) running on a user terminal such as a mobile phone or a personal computer.
- a playback device such as an application for playing a media file (also called a media player) running on a user terminal such as a mobile phone or a personal computer.
- the playback device parses the image frame in the currently played video, and recognizes the facial feature of the frame image, and displays the received information at a position other than the position of the facial feature. on.
- the ability to avoid facial feature display information on the video image improves the video image processing function of the playback device.
- the facial feature in the embodiment of the present application may be a human face, or may be an animal's head or the like.
- the facial feature may be preset as a key information according to actual needs, and if the subtitle information is not blocked, the subtitle information may be Set to the corresponding key information.
- the subtitle information may be Set to the corresponding key information.
- the information received by the playback device may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
- subtitle information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
- advertisement information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
- picture information may be one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
- FIG. 1 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
- Step 101 When the playback device receives the information play request, parsing the image frame of the currently played video.
- Step 102 The playing device identifies the facial feature in the image frame obtained by the parsing.
- the image recognition method is pre-configured according to the feature of the facial feature to be identified. For example, when the facial feature is a human face, the facial recognition method is used to identify the facial feature.
- step 103 the playback device determines the location of the recognized facial feature in the image frame.
- the position of the facial feature identified in this step in the frame image is identified by the region corresponding to the pixel point of the edge position of the facial feature in the frame image.
- the image of this interval corresponds to the facial feature.
- the representation of the location is only an example.
- the identification of the location mode is not limited.
- an area identifier corresponding to a plurality of pixel points on the horizontal and vertical coordinates may be used.
- step 102 and step 103 may be performed on the playback device, or a server may be added to implement the function, and the implementation of the function may be added to the existing server.
- step 104 the playback device displays the received information in a location other than the determined location of the facial feature in the image frame.
- the specific manner in which the playback device displays the received information in the image frame may be displayed according to the characteristics of the information. If the received information is the barrage information, the barrage information may be drawn at a position other than the facial feature. Outside the location.
- the playback device displays the received information on the video image, it is inconsistent with the existing implementation, that is, to avoid the location of the recognized facial feature, and display the received information.
- the embodiment of the present application is not limited, and may be displayed in a horizontal direction, a vertical display, or the like, and displayed according to actual needs or aesthetics.
- FIG. 2 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
- Step 201 When the playback device receives the information play request, parsing the image frame of the currently played video.
- Step 202 The playback device determines whether the local feature of the image frame is recorded locally. If yes, go to step 205; otherwise, go to step 203.
- Step 203 The playing device identifies the facial feature in the image frame obtained by the parsing.
- the received information can be displayed at any position on the frame image, and The received information is displayed in the same way.
- Step 204 The playback device determines the location of the recognized facial feature in the image frame and records the location of the facial feature of the image frame.
- Step 205 The playback device displays the received information in a position other than the location of the facial feature in the image frame.
- the position of the recognized facial feature in the image frame is recorded, and when the frame image is played, when other information is received again, it is not necessary to By performing facial feature recognition, the position of the facial features for the image frame can be directly obtained, which can save time and equipment resources.
- FIG. 3 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
- Step 301 When receiving the information play request, the playback device parses the image frame of the currently played video, and records the information for the image frame.
- Step 302 The playback device identifies the facial feature in the image frame obtained by the parsing.
- step 303 the playback device determines the location of the recognized facial feature in the image frame.
- Step 304 The playback device determines whether information other than the received information is recorded for the image frame, and if yes, step 305 is performed; otherwise, step 306 is performed.
- Step 305 The playback device displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame.
- Step 306 the playing device displays the received information in a position other than the determined position of the facial feature in the image frame.
- the received information is recorded for an image frame, and when the image frame is played again, if the information play request is received again, all the information recorded for the image frame is displayed on the image frame except the face.
- FIG. 4 is a schematic flowchart of displaying information on a video image in an embodiment of the present application. The specific steps are:
- Step 401 When receiving the information play request, the playback device parses the image frame of the currently played video, and records the information for the image frame.
- Step 402 The playing device determines whether the location of the facial feature of the image frame is recorded locally, and if yes, step 405 is performed; otherwise, step 403 is performed.
- Step 403 The playing device identifies the facial feature in the image frame obtained by the parsing.
- Step 404 the playing device determines the position of the recognized facial feature in the image frame, and records the position of the facial feature of the image frame.
- Step 405 The playback device determines whether information other than the received information is recorded for the image frame, and if yes, step 406 is performed; otherwise, step 407 is performed.
- Step 406 The playing device displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame, and ends the flow.
- step 407 the playback device displays the received information in a position other than the determined position of the facial feature in the image frame.
- This embodiment not only saves facial feature recognition time, device resources, but also increases user experience.
- FIG. 5 is a schematic diagram of a video playing system according to an embodiment of the present application.
- the video storage server provides video data for the playback device;
- the information server provides the playback device. Seek information for playback.
- the video storage server and the information server may be one server or multiple servers. When the embodiment of the present application is specifically implemented, the present invention is not limited thereto, and may be implemented according to the existing implementation.
- the video storage server sends the video to be played to the playback device.
- the playback device When the playback device receives the video sent by the video storage server, if it is determined that the video is audio and video, the video data corresponding to the received video is first separated into audio data and video data, and the audio data is decoded to output a sound; After decoding, the video image is output. Regardless of whether or not the received information is displayed, the synchronization of the sound and the image is not changed here, as in the prior implementation.
- the information server when the information server receives the barrage information sent by the user through the client, the information server sends the barrage information to the playback device.
- the playback device When the playback device receives the barrage information sent by the barrage server, it parses the image frame of the currently played video; and records the received barrage information for the image frame.
- the playback device determines whether the location of the facial feature on the image frame is recorded for the image frame, and if so, the playback device determines whether information other than the received information is recorded for the image frame.
- the playback device When the playback device records the location of the facial features on the image frame, the corresponding barrographic information to be recorded for that image frame is displayed at a location other than the location of the facial features.
- the playback device does not record the location of the facial feature on the image frame
- the facial feature is identified in the parsed image frame; the position of the identified facial feature in the image frame is determined; and the facial feature on the image frame is recorded s position.
- the corresponding barrographic information to be recorded for the image frame is then displayed at a location other than the location of the facial feature.
- FIG. 6 is a schematic diagram of displaying barrographic information in a video image in a prior implementation.
- FIG. 7 is a schematic diagram of displaying barrographic information in a video image according to an embodiment of the present application.
- FIG 6 and 7 are schematic diagrams showing the same barr information for the same frame video image.
- the facial feature is an example of face information.
- FIG. 8 is a schematic structural diagram of an apparatus applied to the above technology according to the present application.
- the device includes: a receiving unit 801, a parsing unit 802, an identifying unit 803, a determining unit 804, and a display unit 805;
- the receiving unit 801 is configured to receive an information play request.
- the parsing unit 802 is configured to parse the image frame of the currently played video when the receiving unit 801 receives the information playing request;
- the identifying unit 803 is configured to identify a facial feature in the image frame obtained by the parsing unit 802;
- a determining unit 804 configured to determine a location of the facial feature identified by the identifying unit 803 in the image frame
- the display unit 805 is configured to display the information received by the receiving unit 801 in a position other than the position of the facial feature determined by the determining unit 804 in the image frame.
- the device further comprises: a recording unit 806;
- a recording unit 806, configured to: when the determining unit 804 determines the location of the recognized facial feature in the image frame, record the location of the facial feature of the image frame;
- the determining unit 804 is further configured to: after the parsing unit 802 parses the image frame of the currently playing video, determine whether the recording unit records the location of the facial feature of the image frame, and if so, the trigger display unit 805 displays the received information in the A position other than the determined position of the facial feature in the image frame; otherwise, the trigger recognition unit 803 recognizes the facial feature in the image frame obtained by the analysis.
- the device further includes: a recording unit 806;
- the recording unit 806 is further configured to: when the parsing unit 802 parses the image frame of the currently played video, record the information for the image frame;
- the display unit 805 is further configured to: before displaying the received information in a position other than the determined position of the facial feature in the image frame, determining whether the recording unit records the received information for the image frame The information, if so, the trigger display unit 805 displays all the information for the image frame in a position other than the determined position of the facial feature in the image frame; otherwise, the trigger display unit 805 displays the received information at The position of the image frame other than the position of the determined facial feature.
- the position of the facial feature in the frame image is identified using an area corresponding to a pixel point of an edge position of the facial feature in the frame image.
- the received information is one or any combination of the following: subtitle information, advertisement information, barrage information, and picture information.
- the units of the above embodiments may be integrated into one, or may be deployed separately; may be combined into one unit, or may be further split into multiple sub-units.
- a hardware module can include specially designed permanent circuits or logic devices (such as dedicated processors such as FPGAs or ASICs) for performing specific operations.
- Hardware modules can also be included by Programmable logic devices or circuits (such as general purpose processors or other programmable processors) that are temporarily configured by software are used to perform specific operations.
- Hardware implementations can be made based on cost and time considerations, either by mechanical means, by dedicated permanent circuits, or by temporarily configured circuits (as configured by software).
- FIG. 9 is a schematic structural diagram of hardware architecture for displaying information on a video image in a specific embodiment of the present application.
- the playback device can include a processor 910, a memory 920, a port 930, and a bus 940.
- Processor 910 and memory 920 are interconnected by a bus 940.
- the processor 910 can acquire and output data through the port 1130;
- the receiving unit 801 When the receiving unit 801 is executed by the processor 910, it may be: receiving an information play request;
- parsing unit 802 When the parsing unit 802 is executed by the processor 910, it may be: parsing an image frame of the currently played video;
- the identification unit 803, when executed by the processor 910, may be: identifying a facial feature in an image frame;
- the determining unit 804 when executed by the processor 910, may be: determining a location of the recognized facial feature in the image frame;
- embodiments of the present application may be implemented by a data processing program executed by a device such as a computer.
- the data processing program constitutes the present application.
- a data processing program that is usually stored in one storage medium is executed by directly reading the program out of the storage medium or by installing or copying the program to a storage device (such as a hard disk and/or a memory) of the data processing device. Therefore, such a storage medium also constitutes the present application.
- the storage medium can use any type of recording method, such as paper storage medium (such as paper tape, etc.), magnetic storage medium (such as floppy disk, hard disk, flash memory, etc.), optical storage medium (such as CD-ROM, etc.), magneto-optical storage medium ( Such as MO, etc.).
- the present application also discloses a storage medium, such as a non-volatile computer storage medium, in which a data processing program is stored, which is used to execute the above technical solution of the present application.
- the playback device shown in FIG. 9 is only a specific example, and may be implemented by other structures different from those described in the embodiment, for example, operations performed when the above instruction code is executed, or may be performed by a specific application.
- the above-mentioned processor 910 may be one or more, and if there are multiple, the plurality of processors are jointly responsible for reading and executing the instruction code. Therefore, the specific structure of the test device is not specifically limited in the present application.
- the present application when receiving an information play request by the playback device, displays the received information on the currently played image frame while avoiding the facial feature.
- the realization of the scheme can avoid the facial feature display information on the video image and improve the video image processing function of the playback device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Marketing (AREA)
- Business, Economics & Management (AREA)
- General Engineering & Computer Science (AREA)
- Library & Information Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims (11)
- 一种在视频图像上显示信息的方法,应用于播放设备上,其特征在于,该方法包括:该播放设备接收信息播放请求;解析当前播放视频的图像帧;在解析获得的图像帧中识别出面部特征;确定识别出的面部特征在该图像帧中的位置;将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
- 根据权利要求1所述的方法,其特征在于,所述确定识别出的面部特征在该图像帧中的位置时,所述方法进一步包括:记录该图像帧的面部特征的位置;所述解析当前播放视频的图像帧之后,所述在解析获得的图像帧中识别出面部特征之前,所述方法进一步包括:确定本地是否记录该图像帧的面部特征的位置,如果是,将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,执行所述在解析获得的图像帧中识别出面部特征步骤。
- 根据权利要求1或2所述的方法,其特征在于,所述解析当前播放视频的图像帧时,所述方法进一步包括:针对该图像帧记录所述信息;所述将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置之前,所述方法进一步包括:确定针对该图像帧是否记录接收的信息之外的信息,如果是,将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,将接收到的信息显示在该图像帧中除确定出的面部 特征的位置之外的位置。
- 根据权利要求1或2所述的方法,其特征在于,所述面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
- 根据权利要求1或2所述的方法,其特征在于,所述接收到的信息为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。
- 一种在视频图像上显示信息的装置,应用于播放设备上,其特征在于,该装置至少包括处理器和存储器,其中所述存储器中存储接收单元、解析单元、识别单元、确定单元和显示单元,当被所述处理器执行时,所述接收单元用于接收信息播放请求;所述解析单元用于解析当前播放视频的图像帧;所述识别单元用于在所述图像帧中识别出面部特征;所述确定单元用于确定所述识别出的面部特征在该图像帧中的位置;所述显示单元用于将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
- 根据权利要求6所述的装置,其特征在于,该存储器中进一步存储记录单元,当被所述处理器执行时,所述记录单元用于根据确定识别出的面部特征在该图像帧中的位置,记录该图像帧的面部特征的位置;所述确定单元进一步用于当解析当前播放视频的图像帧之后,确定是否已经记录该图像帧的面部特征的位置,如果是,触发将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触 发在解析获得的图像帧中识别出面部特征。
- 根据权利要求6或7所述的装置,其特征在于,该存储器中进一步存储记录单元,当被所述处理器执行时,所述记录单元进一步用于当解析当前播放视频的图像帧时,针对该图像帧记录所述信息;所述显示单元进一步用于将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置之前,确定针对该图像帧是否记录接收的信息之外的信息,如果是,触发将针对该图像帧的所有信息显示在该图像帧中除确定出的面部特征的位置之外的位置;否则,触发将接收到的信息显示在该图像帧中除确定出的面部特征的位置之外的位置。
- 根据权利要求6或7所述的装置,其特征在于,所述面部特征在帧图像中的位置使用该帧图像中所述面部特征的边缘位置的像素点所对应的区域标识。
- 根据权利要求6或7所述的装置,其特征在于,所述接收到的信息为下述之一或任意组合:字幕信息、广告信息、弹幕信息、图片信息。
- 一种非易失性计算机存储介质,其特征在于,其中存储有计算机程序,该计算机程序用于执行所述权利要求1至5任一项所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017560519A JP6438598B2 (ja) | 2015-11-06 | 2016-10-20 | ビデオ画像の上に情報を表示するための方法及びデバイス |
US15/787,908 US10425679B2 (en) | 2015-11-06 | 2017-10-19 | Method and device for displaying information on video image |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510751308.4 | 2015-11-06 | ||
CN201510751308.4A CN105430512A (zh) | 2015-11-06 | 2015-11-06 | 一种在视频图像上显示信息的方法和装置 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/787,908 Continuation US10425679B2 (en) | 2015-11-06 | 2017-10-19 | Method and device for displaying information on video image |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017076171A1 true WO2017076171A1 (zh) | 2017-05-11 |
Family
ID=55508395
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2016/102640 WO2017076171A1 (zh) | 2015-11-06 | 2016-10-20 | 一种在视频图像上显示信息的方法和装置 |
Country Status (4)
Country | Link |
---|---|
US (1) | US10425679B2 (zh) |
JP (1) | JP6438598B2 (zh) |
CN (1) | CN105430512A (zh) |
WO (1) | WO2017076171A1 (zh) |
Families Citing this family (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8721571B2 (en) | 2010-11-22 | 2014-05-13 | Siwa Corporation | Selective removal of cells having accumulated agents |
US10358502B2 (en) | 2014-12-18 | 2019-07-23 | Siwa Corporation | Product and method for treating sarcopenia |
CN105430512A (zh) | 2015-11-06 | 2016-03-23 | 腾讯科技(北京)有限公司 | 一种在视频图像上显示信息的方法和装置 |
CN105635848A (zh) * | 2015-12-24 | 2016-06-01 | 深圳市金立通信设备有限公司 | 一种弹幕显示方法及终端 |
CN106210855B (zh) * | 2016-07-11 | 2019-12-13 | 网易(杭州)网络有限公司 | 对象显示方法和装置 |
CN106303731A (zh) * | 2016-08-01 | 2017-01-04 | 北京奇虎科技有限公司 | 弹幕的显示方法及装置 |
CN108124185B (zh) * | 2016-11-28 | 2019-06-21 | 广州华多网络科技有限公司 | 一种弹幕显示方法、装置及终端 |
US10858449B1 (en) | 2017-01-06 | 2020-12-08 | Siwa Corporation | Methods and compositions for treating osteoarthritis |
US10925937B1 (en) | 2017-01-06 | 2021-02-23 | Siwa Corporation | Vaccines for use in treating juvenile disorders associated with inflammation |
CN107181976B (zh) * | 2017-04-28 | 2021-01-29 | 华为技术有限公司 | 一种弹幕显示方法及电子设备 |
CN107147941A (zh) * | 2017-05-27 | 2017-09-08 | 努比亚技术有限公司 | 视频播放的弹幕显示方法、装置及计算机可读存储介质 |
CN107454255B (zh) * | 2017-07-28 | 2020-07-17 | 维沃移动通信有限公司 | 一种歌词显示方法、移动终端及计算机可读存储介质 |
CN107943964A (zh) * | 2017-11-27 | 2018-04-20 | 腾讯音乐娱乐科技(深圳)有限公司 | 歌词显示方法、装置及计算机可读存储介质 |
CN107948760B (zh) * | 2017-11-30 | 2021-01-29 | 上海哔哩哔哩科技有限公司 | 弹幕播放控制方法、服务器及弹幕播放控制系统 |
US11518801B1 (en) | 2017-12-22 | 2022-12-06 | Siwa Corporation | Methods and compositions for treating diabetes and diabetic complications |
CN109302619A (zh) * | 2018-09-18 | 2019-02-01 | 北京奇艺世纪科技有限公司 | 一种信息处理方法及装置 |
CN109618213B (zh) * | 2018-12-17 | 2020-05-19 | 华中科技大学 | 一种防止弹幕遮挡目标对象的方法 |
CN109905775A (zh) * | 2019-01-16 | 2019-06-18 | 北京奇艺世纪科技有限公司 | 一种涂鸦弹幕生成及显示方法、装置、终端设备及存储介质 |
CN109905757A (zh) * | 2019-02-13 | 2019-06-18 | 安徽励图信息科技股份有限公司 | 通过人脸识别控制视频字幕播出的方法 |
CN109862414B (zh) * | 2019-03-22 | 2021-10-15 | 武汉斗鱼鱼乐网络科技有限公司 | 一种蒙版弹幕显示方法、装置及服务器 |
CN116916080A (zh) * | 2019-05-17 | 2023-10-20 | 上海哔哩哔哩科技有限公司 | 视频数据处理方法、装置、计算机设备及可读存储介质 |
CN110362373A (zh) * | 2019-06-29 | 2019-10-22 | 华为技术有限公司 | 一种控制屏幕小窗口的方法及相关设备 |
CN110351596B (zh) * | 2019-07-17 | 2021-07-27 | 上海播呗网络科技有限公司 | 一种互联网流媒体大数据弹幕处理系统及处理方法 |
CN110636259B (zh) * | 2019-09-10 | 2021-03-02 | 杭州亿圣信息技术有限公司 | 一种基于视频检测的osd叠加方法 |
CN110958480A (zh) * | 2019-11-28 | 2020-04-03 | 湖南快乐阳光互动娱乐传媒有限公司 | 一种弹幕显示方法 |
CN113593567B (zh) * | 2021-06-23 | 2022-09-09 | 荣耀终端有限公司 | 视频声音转文本的方法及相关设备 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1423896A (zh) * | 2000-12-15 | 2003-06-11 | 皇家菲利浦电子有限公司 | 基于视频内容分析的画中画重新定位和/或大小调整 |
JP2004128614A (ja) * | 2002-09-30 | 2004-04-22 | Toshiba Corp | 画像表示制御装置及び画像表示制御プログラム |
CN102845067A (zh) * | 2010-04-01 | 2012-12-26 | 汤姆森许可贸易公司 | 三维(3d)呈现中的字幕 |
US20140196082A1 (en) * | 2012-07-17 | 2014-07-10 | Panasonic Corporation | Comment information generating apparatus and comment information generating method |
CN105430512A (zh) * | 2015-11-06 | 2016-03-23 | 腾讯科技(北京)有限公司 | 一种在视频图像上显示信息的方法和装置 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7184100B1 (en) * | 1999-03-24 | 2007-02-27 | Mate - Media Access Technologies Ltd. | Method of selecting key-frames from a video sequence |
US6778224B2 (en) * | 2001-06-25 | 2004-08-17 | Koninklijke Philips Electronics N.V. | Adaptive overlay element placement in video |
JP2003271954A (ja) * | 2002-03-12 | 2003-09-26 | Canon I-Tech Inc | 画像処理装置、画像処理方法、プログラムおよび記録媒体 |
JP2007251273A (ja) * | 2006-03-13 | 2007-09-27 | Oki Electric Ind Co Ltd | 画像処理装置,画像伝送システムおよび画像処理方法 |
JP4507281B2 (ja) * | 2006-03-30 | 2010-07-21 | 富士フイルム株式会社 | 画像表示装置、撮像装置および画像表示方法 |
US20120323704A1 (en) * | 2008-02-27 | 2012-12-20 | Chad Steelberg | Enhanced world wide web-based communications |
JP2010226351A (ja) * | 2009-03-23 | 2010-10-07 | Fujitsu Ten Ltd | 映像再生装置 |
JP5371574B2 (ja) * | 2009-06-23 | 2013-12-18 | 株式会社第一興商 | 背景映像中の顔画像を避けるように歌詞字幕を表示するカラオケ装置 |
JP5465620B2 (ja) * | 2010-06-25 | 2014-04-09 | Kddi株式会社 | 映像コンテンツに重畳する付加情報の領域を決定する映像出力装置、プログラム及び方法 |
JP2013247611A (ja) * | 2012-05-29 | 2013-12-09 | Nippon Telegr & Teleph Corp <Ntt> | 字幕放送出力装置、字幕放送出力方法及び字幕放送出力プログラム |
US9124765B2 (en) * | 2012-12-27 | 2015-09-01 | Futurewei Technologies, Inc. | Method and apparatus for performing a video conference |
-
2015
- 2015-11-06 CN CN201510751308.4A patent/CN105430512A/zh active Pending
-
2016
- 2016-10-20 JP JP2017560519A patent/JP6438598B2/ja active Active
- 2016-10-20 WO PCT/CN2016/102640 patent/WO2017076171A1/zh active Application Filing
-
2017
- 2017-10-19 US US15/787,908 patent/US10425679B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1423896A (zh) * | 2000-12-15 | 2003-06-11 | 皇家菲利浦电子有限公司 | 基于视频内容分析的画中画重新定位和/或大小调整 |
JP2004128614A (ja) * | 2002-09-30 | 2004-04-22 | Toshiba Corp | 画像表示制御装置及び画像表示制御プログラム |
CN102845067A (zh) * | 2010-04-01 | 2012-12-26 | 汤姆森许可贸易公司 | 三维(3d)呈现中的字幕 |
US20140196082A1 (en) * | 2012-07-17 | 2014-07-10 | Panasonic Corporation | Comment information generating apparatus and comment information generating method |
CN105430512A (zh) * | 2015-11-06 | 2016-03-23 | 腾讯科技(北京)有限公司 | 一种在视频图像上显示信息的方法和装置 |
Also Published As
Publication number | Publication date |
---|---|
JP6438598B2 (ja) | 2018-12-19 |
JP2018519727A (ja) | 2018-07-19 |
US20180041796A1 (en) | 2018-02-08 |
US10425679B2 (en) | 2019-09-24 |
CN105430512A (zh) | 2016-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2017076171A1 (zh) | 一种在视频图像上显示信息的方法和装置 | |
WO2019109643A1 (zh) | 视频推荐方法、装置、计算机设备和存储介质 | |
CN109089154B (zh) | 一种视频提取方法、装置、设备及介质 | |
CN109089127B (zh) | 一种视频拼接方法、装置、设备及介质 | |
WO2017092343A1 (zh) | 一种视频数据的检测方法和装置 | |
US20160307371A1 (en) | Synchronizing an augmented reality video stream with a displayed video stream | |
WO2015081776A1 (zh) | 视频画面的处理方法及装置 | |
WO2017181598A1 (zh) | 视频播放方法及装置 | |
JP4267649B2 (ja) | ビデオ番組の処理方法、関連装置及び関連媒体 | |
US20190147865A1 (en) | Content recognizing method and apparatus, device, and computer storage medium | |
CN109729429B (zh) | 视频播放方法、装置、设备和介质 | |
JP2009177411A (ja) | 電子機器および画像表示方法 | |
CN104967903A (zh) | 一种视频播放的检测方法及装置 | |
JP2010257509A (ja) | オーサリング装置、オーサリング方法およびプログラム | |
WO2017202060A1 (zh) | 一种视频播放方法及装置 | |
US9264646B2 (en) | Electronic device and video playing method | |
WO2017096898A1 (zh) | 广告展示方法及装置 | |
JP2010509830A (ja) | ビデオデータストリームの要約を生成するための方法及び装置 | |
WO2019114330A1 (zh) | 一种视频播放方法、装置和终端设备 | |
CN112055254A (zh) | 视频播放的方法、装置、终端及存储介质 | |
WO2017101414A1 (zh) | 基于安卓平台的多媒体播放方法、装置及移动终端设备 | |
US20090193355A1 (en) | Information processing apparatus and display control method | |
TWI535278B (zh) | 影片播放方法以及系統 | |
CN112188221B (zh) | 播放控制方法、装置、计算机设备及存储介质 | |
CN109840406B (zh) | 活体验证方法、装置和计算机设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16861439 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2017560519 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 01.10.2018) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16861439 Country of ref document: EP Kind code of ref document: A1 |