CN106254939A - Information cuing method and device - Google Patents

Information cuing method and device Download PDF

Info

Publication number
CN106254939A
CN106254939A CN201610875775.2A CN201610875775A CN106254939A CN 106254939 A CN106254939 A CN 106254939A CN 201610875775 A CN201610875775 A CN 201610875775A CN 106254939 A CN106254939 A CN 106254939A
Authority
CN
China
Prior art keywords
video
frame
audio
audio frame
people information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610875775.2A
Other languages
Chinese (zh)
Other versions
CN106254939B (en
Inventor
张亮
伍亮雄
刘鸣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201610875775.2A priority Critical patent/CN106254939B/en
Publication of CN106254939A publication Critical patent/CN106254939A/en
Application granted granted Critical
Publication of CN106254939B publication Critical patent/CN106254939B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06KRECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K9/00Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
    • G06K9/00624Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
    • G06K9/00711Recognising video content, e.g. extracting audiovisual features from movies, extracting representative key-frames, discriminating news vs. sport content
    • G06K9/00758Matching video sequences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4722End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting additional data associated with the content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4882Data services, e.g. news ticker for displaying messages, e.g. warnings, reminders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • H04N21/8133Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program

Abstract

The disclosure is directed to information cuing method and device.The method includes: when playing video, obtain the audio frame in described video;Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;People information corresponding for described audio frame is pointed out.This technical scheme, when playing this video, by obtaining the audio frame in this video, and this audio frame is automatically analyzed, the information of voice-over actor corresponding to this audio frame can be got, and then people information corresponding for audio frame is pointed out, so that user is when watching this video, the specifying information of the voice-over actor can being fully understood by this video, this viewing experience being conducive to improving user further, can overcome and not identify the information of voice-over actor in correlation technique in video and cause user cannot understand the defect of voice-over actor.

Description

Information cuing method and device
Technical field
It relates to field of terminal technology, particularly relate to information cuing method and device.
Background technology
At present, user, when watching video, is typically only capable to see the part people information of this video, such as the master of this video Drill, director etc., and these information imperfect, it is impossible to make user be fully understood by the various people informations of this video.
Summary of the invention
Disclosure embodiment provides information cuing method and device.Described technical scheme is as follows:
First aspect according to disclosure embodiment, it is provided that a kind of information cuing method, including:
When playing video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
In one embodiment, described method also includes:
When playing described video, obtain frame of video corresponding with described audio frame in described video;
Described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding;
Described people information corresponding for described audio frame is pointed out, including:
When the people information corresponding with described audio frame when the people information that described frame of video is corresponding does not mates, by described sound Frequently the people information that frame is corresponding is pointed out.
In one embodiment, described described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding, Including:
The image of described frame of video is mated with at least one pre-set image;
When the image of described frame of video matches with the target image at least one pre-set image described, by described mesh People information corresponding to logo image is defined as the people information that described frame of video is corresponding.
In one embodiment, described described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding, Including:
Obtain the audio parameter of described audio frame;
The audio parameter of described audio frame is mated with the audio parameter of at least one preset sound;
Audio parameter phase when audio parameter and the target sound at least one preset sound described of described audio frame During coupling, the people information of described target sound is defined as the people information that described audio frame is corresponding.
In one embodiment, described audio parameter includes: at least one in loudness, tone, tone color.
In one embodiment, the people information that described audio frame is corresponding includes: the title of the personage that described audio frame is corresponding Exhale, at least one information in position, contact method.
Second aspect according to disclosure embodiment, it is provided that a kind of information presentation device, including:
First acquisition module, for when playing video, obtaining the audio frame in described video;
Second acquisition module, for being analyzed described audio frame, it is thus achieved that the people information that described audio frame is corresponding;
Reminding module, for pointing out people information corresponding for described audio frame.
In one embodiment, described device also includes:
3rd acquisition module, for when playing described video, obtaining corresponding with described audio frame in described video regarding Frequently frame;
4th acquisition module, for being identified described frame of video, it is thus achieved that the people information that described frame of video is corresponding;
Described reminding module includes:
First prompting submodule, for believing as the people information that described frame of video the is corresponding personage corresponding with described audio frame When breath does not mates, people information corresponding for described audio frame is pointed out.
In one embodiment, described 4th acquisition module includes:
First matched sub-block, for mating the image of described frame of video with at least one pre-set image;
First determines submodule, for when the image of described frame of video and the target figure at least one pre-set image described During as matching, people information corresponding for described target image is defined as the people information that described frame of video is corresponding.
In one embodiment, described second acquisition module includes:
Obtain submodule, for obtaining the audio parameter of described audio frame;
Second matched sub-block, for the audio parameter by the audio parameter of described audio frame with at least one preset sound Mate;
Second determines submodule, for when the audio parameter of described audio frame and the mesh at least one preset sound described When the audio parameter of mark sound matches, the people information of described target sound is defined as personage's letter that described audio frame is corresponding Breath.
In one embodiment, described audio parameter includes: at least one in loudness, tone, tone color.
In one embodiment, the people information that described audio frame is corresponding includes: the title of the personage that described audio frame is corresponding Exhale, at least one information in position, contact method.
The third aspect according to disclosure embodiment, it is provided that a kind of information presentation device, including:
Processor;
For storing the memorizer of processor executable;
Wherein, described processor is configured to:
When playing described video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
Embodiment of the disclosure that the technical scheme of offer can include following beneficial effect:
The technical scheme of offer is provided, when playing this video, by obtaining the audio frame in this video, And this audio frame is automatically analyzed, the information of voice-over actor corresponding to this audio frame can be got, and then by audio frame Corresponding people information is pointed out, so that user is when watching this video, can be fully understood by dubbing in this video and drill The specifying information of member, this viewing experience being conducive to improving user further, can overcome in correlation technique and video is not marked Know and the information of voice-over actor and cause user cannot understand the defect of voice-over actor.
It should be appreciated that it is only exemplary and explanatory, not that above general description and details hereinafter describe The disclosure can be limited.
Accompanying drawing explanation
Accompanying drawing herein is merged in description and constitutes the part of this specification, it is shown that meet the enforcement of the disclosure Example, and for explaining the principle of the disclosure together with description.
Fig. 1 is the flow chart according to a kind of information cuing method shown in an exemplary embodiment.
Fig. 2 is the flow chart according to the another kind of information cuing method shown in an exemplary embodiment.
Fig. 3 is the flow chart according to another information cuing method shown in an exemplary embodiment.
Fig. 4 is the flow chart according to another information cuing method shown in an exemplary embodiment.
Fig. 5 is the block diagram according to a kind of information presentation device shown in an exemplary embodiment.
Fig. 6 is the block diagram according to the another kind of information presentation device shown in an exemplary embodiment.
Fig. 7 is the block diagram according to another information presentation device shown in an exemplary embodiment.
Fig. 8 is the block diagram according to another information presentation device shown in an exemplary embodiment.
Fig. 9 is according to the block diagram being applicable to information presentation device shown in an exemplary embodiment.
Detailed description of the invention
Here will illustrate exemplary embodiment in detail, its example represents in the accompanying drawings.Explained below relates to During accompanying drawing, unless otherwise indicated, the same numbers in different accompanying drawings represents same or analogous key element.Following exemplary embodiment Described in embodiment do not represent all embodiments consistent with the disclosure.On the contrary, they are only with the most appended The example of the apparatus and method that some aspects that described in detail in claims, the disclosure are consistent.
At present, user is when watching the video such as film or TV play, owing to the video in correlation technique only identifies The protagonist of this video, director etc., do not identify the information of voice-over actor, and this results in user and can only see the part of this video People information, such as the protagonist of this video, director etc., and due to these information imperfect, thus, user when watching video, May be not aware that the voice-over actor in current video, more not know whom voice-over actor is, this just carries to user and serves puzzlement, User cannot be made to be fully understood by the various people informations of this video, be also unfavorable for the development of voice-over actor simultaneously.
In order to solve above-mentioned technical problem, disclosure embodiment provides a kind of information cuing method, and the method can be used for In information alert program, system or device, and executive agent corresponding to the method can be the terminals such as mobile phone, flat board, computer, As it is shown in figure 1, the method comprising the steps of S101 is to step S103:
In step S101, when playing video, obtain the audio frame in video;
This video can be film, TV play etc..
When obtaining the audio frame in this video, can obtain according to prefixed time interval, or according to other conditions Obtain, such as: when the character image in monitoring frame of video corresponding with this audio frame is for presetting character image, then acquisition should Audio frame.
In step s 102, audio frame is analyzed, it is thus achieved that the people information that audio frame is corresponding;
The people information that this audio frame is corresponding may include that the address of the personnel that dub that this audio frame is corresponding, position, contact At least one information such as mode.
Due to the possible corresponding different voice-over actor of different audio frames, thus, when this audio frame includes multiframe, corresponding People information may also have multiple.
It addition, people information corresponding to this audio frame and people information corresponding to frame of video corresponding with this audio frame may Coupling, it is likely to not mate, the personnel that dub that i.e. this audio frame is corresponding and performer corresponding to frame of video corresponding with audio frame Can be same person or be not same person.
In step s 103, people information corresponding for audio frame is pointed out.
When playing this video, by obtaining the audio frame in this video, and this audio frame is automatically analyzed, permissible Get the information of voice-over actor corresponding to this audio frame, and then people information corresponding for audio frame is pointed out, so that User is when watching this video, and the specifying information of the voice-over actor can being fully understood by this video, this is conducive to carrying further The viewing experience of high user, can overcome correlation technique does not identifies in video voice-over actor information and cause user without Method understands the defect of voice-over actor, such as: when watching " Kung Fu Panda " video, after being analyzed audio frame, can obtain The voice-over actor getting father A Bao is Cheng Long, then voice-over actor Cheng Long can carry out pointing out to increase the viewing of user Experience.
Secondly, when people information corresponding for this audio frame is pointed out, can be by personage's letter corresponding for this audio frame Breath is pointed out in credit, end credits, or points out in the picture of frame of video corresponding to this audio frame.
As in figure 2 it is shown, in one embodiment, method also includes:
In step s 201, when playing video, obtain frame of video corresponding with audio frame in video;
Frame of video corresponding with this audio frame is the video that need synchronize play identical with the reproduction time of this audio frame Frame (such as: if the reproduction time that this audio frame is in this video is t1 t2, then the broadcasting of frame of video corresponding with this audio frame Put the time also for t1 t2), or for pre-set the frame of video of incidence relation (such as: when in this video with this audio frame Audio frame and frame of video between the most synchronize time, in order to avoid playing disorder, pre-between its corresponding frame of video of audio frame First may be provided with incidence relation).
In step S202, frame of video is identified, it is thus achieved that the people information that frame of video is corresponding;
It addition, step S201 and step S202 can Tong Bu with step S101 and step S102 perform, it is also possible to behind Perform.
Above-mentioned steps S103 can be performed as:
In step A1, when the people information corresponding with audio frame when the people information that frame of video is corresponding does not mates, by sound Frequently the people information that frame is corresponding is pointed out.
When playing this video, it is also possible to obtain frame of video corresponding with audio frame in video, and frame of video is known Not, thus obtain the people information that frame of video is corresponding, and when the personage that the people information that frame of video is corresponding is corresponding with audio frame believes Breath is when mating, and illustrates that voice-over actor that this audio frame the is corresponding performer corresponding with frame of video is not same person, and then The sound source illustrating audio frame corresponding to this frame of video is derived from other voice-over actors rather than performer corresponding to this frame of video, Therefore, it is possible to people information corresponding for audio frame is pointed out, so that user is when watching this video, can be fully understood by The specifying information of the voice-over actor in this video, to improve the viewing experience of user further.
During it addition, the people information corresponding with audio frame when the people information that frame of video is corresponding matches, this video is described Performer that frame the is corresponding performer corresponding with frame of video is same performer, has further related to audio frame corresponding to this frame of video not It is derived from other voice-over actors, but the performer that this frame of video is corresponding, again due to performer corresponding to the frame of video in video Information is carried in video, therefore, it is possible to no longer pointed out by people information corresponding for audio frame, to avoid repeating Prompting.
As it is shown on figure 3, in one embodiment, step S202 shown in above-mentioned Fig. 2 can be performed as:
In step bl is determined., the image of frame of video is mated with at least one pre-set image;
At least one pre-set image can come from being pre-stored in the image of this locality, or the image of network side, and these In image, the people information of each image is the most identified.
It addition, pre-set image can come from the image library of feature, for example, it is possible to from movie and television play actor image storehouse, stage Acute actor image storehouse, TV play actor image storehouse, conventional welcome guest's image library etc..By the image of this frame of video and at least one When pre-set image mates, it is right the image of this frame of video and each image at least one pre-set image to be carried out Ratio, such as: can be by the image parameter of the image parameter of the image of frame of video with each image at least one pre-set image Contrasting, wherein, this image parameter can be color parameter, textural characteristics, shape facility etc..
In step B2, when the image of frame of video matches with the target image at least one pre-set image, by mesh People information corresponding to logo image is defined as the people information that frame of video is corresponding, wherein, this target image can be this at least one Any image in individual pre-set image.
When the image of frame of video matches with the target image at least one pre-set image, the figure of this frame of video is described As high with the similarity of this target image, substantially may determine that the personage in the image of this frame of video and the people in target image Thing is same personage, therefore, it is possible to automatically determine people information corresponding for this target image the most marked out into this The people information that frame of video is corresponding.
As shown in Figure 4, in one embodiment, step S102 in above-mentioned Fig. 1 can be performed as:
In step C1, obtain the audio parameter of audio frame;
Wherein, this audio parameter includes but not limited to: at least one in loudness, tone, tone color.
In step C2, the audio parameter of audio frame is mated with the audio parameter of at least one preset sound;
When the audio parameter of this audio frame is mated with the audio parameter of at least one preset sound, can will be somebody's turn to do The audio parameter of audio frame contrasts with the audio parameter of at least one preset sound.
At least one preset sound can include being pre-stored in the sound of this locality, or network side sound, and arrive In a few preset sound, the people information of each sound is the most identified.
And this preset sound can come from specific sound storehouse, for example, it is possible to from voice-over actor's voice bank.
It addition, in one embodiment, in order to be quickly found the target sound of coupling, by the audio parameter of audio frame with When the audio parameter of at least one preset sound mates, can first determine this audio frame according to the audio parameter of audio frame Personage's attribute informations such as corresponding the range of age of personage, sex, and then, according to these character attribute information from this at least one Preset sound is determined the sound to be selected with this character attribute information match, the most again by the audio parameter of this audio frame Mate with the audio parameter of each sound in this sound to be selected, thus determine final mesh from these sound to be selected Mark sound.
In step C3, when the audio parameter of the target sound in the audio parameter and at least one preset sound of audio frame When matching, the people information of target sound is defined as the people information that audio frame is corresponding.
When the audio parameter of audio frame matches with the audio parameter of the target sound at least one preset sound, say The audio parameter of this audio frame bright is high with the similarity of the audio parameter of target sound, substantially may determine that this audio frame is corresponding The personage personage corresponding with target sound be same person, therefore, it is possible to the people information of target sound is defined as audio frequency The people information that frame is corresponding.
In one embodiment, audio parameter includes: at least one in loudness, tone, tone color.
In one embodiment, the people information that audio frame is corresponding includes: the address of the personage that audio frame is corresponding, position, At least one information in contact method.
The people information that this audio frame is corresponding be this audio frame dub personnel, the address of the personage that this audio frame is corresponding can To be the name of the personnel that dub of this audio frame, the pet name etc.;
The people information that audio frame is corresponding includes but not limited at least one information above-mentioned, for example, it is also possible to include this sound Frequently the information such as address of the personage that frame is corresponding.
The above-mentioned information cuing method that corresponding disclosure embodiment provides, disclosure embodiment also provides for a kind of information alert Device, as it is shown in figure 5, this device includes:
First acquisition module 501, is configured to, when playing video, obtain the audio frame in video;
Second acquisition module 502, is configured to be analyzed audio frame, it is thus achieved that the people information that audio frame is corresponding;
Reminding module 503, is configured to point out people information corresponding for audio frame.
As shown in Figure 6, in one embodiment, device also includes:
3rd acquisition module 601, is configured to, when playing video, obtain frame of video corresponding with audio frame in video;
4th acquisition module 602, is configured to be identified frame of video, it is thus achieved that the people information that frame of video is corresponding;
Reminding module 503 may include that
First prompting submodule 5031, is configured as personage's letter that people information corresponding to frame of video is corresponding with audio frame When breath does not mates, people information corresponding for audio frame is pointed out.
As it is shown in fig. 7, in one embodiment, the 4th acquisition module 602 may include that
First matched sub-block 6021, is configured to mate the image of frame of video with at least one pre-set image;
First determines submodule 6022, is configured as image and the target figure at least one pre-set image of frame of video During as matching, people information corresponding for target image is defined as the people information that frame of video is corresponding.
As shown in Figure 8, in one embodiment, the second acquisition module 502 may include that
Obtain submodule 5021, be configured to obtain the audio parameter of audio frame;
Second matched sub-block 5022, is configured to the sound of the audio parameter of audio frame Yu at least one preset sound Parameter is mated;
Second determines submodule 5023, is configured as audio parameter and the mesh at least one preset sound of audio frame When the audio parameter of mark sound matches, the people information of described target sound is defined as personage's letter that described audio frame is corresponding Breath.
In one embodiment, audio parameter includes: at least one in loudness, tone, tone color.
In one embodiment, the people information that audio frame is corresponding includes: the address of the personage that audio frame is corresponding, position, At least one information in contact method.
The third aspect according to disclosure embodiment, it is provided that a kind of information presentation device, including:
Processor;
For storing the memorizer of processor executable;
Wherein, processor is configured to:
When playing video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
Above-mentioned processor is also configured to:
Described method also includes:
When playing described video, obtain frame of video corresponding with described audio frame in described video;
Described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding;
Described people information corresponding for described audio frame is pointed out, including:
When the people information corresponding with described audio frame when the people information that described frame of video is corresponding does not mates, by described sound Frequently the people information that frame is corresponding is pointed out.
Above-mentioned processor is also configured to:
Described described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding, including:
The image of described frame of video is mated with at least one pre-set image;
When the image of described frame of video matches with the target image at least one pre-set image described, by described mesh People information corresponding to logo image is defined as the people information that described frame of video is corresponding.
Above-mentioned processor is also configured to:
Described described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding, including:
Obtain the audio parameter of described audio frame;
The audio parameter of described audio frame is mated with the audio parameter of at least one preset sound;
When the audio parameter of described audio frame is joined with the sound of the target sound at least one preset sound parameter described When number matches, the people information of described target sound is defined as the people information that described audio frame is corresponding.
Above-mentioned processor is also configured to:
Described audio parameter includes: at least one in loudness, tone, tone color.
Above-mentioned processor is also configured to:
The people information that described audio frame is corresponding includes: the address of the personage that described audio frame is corresponding, position, contact method In at least one information.
Fig. 9 is according to a kind of block diagram for information presentation device 900 shown in an exemplary embodiment, and this device is suitable for In terminal unit.Such as, device 900 can be mobile phone, computer, digital broadcast terminal, messaging devices, game control Platform processed, tablet device, armarium, body-building equipment, individual number assistant etc..
With reference to Fig. 9, device 900 can include with next or at least two assembly: process assembly 902, memorizer 904, Power supply module 906, multimedia groupware 908, audio-frequency assembly 910, input/output (I/O) interface 912, sensor cluster 914, with And communications component 916.
Process assembly 902 and generally control the integrated operation of device 900, such as with display, call, data communication, phase The operation that machine operation and record operation are associated.Process assembly 902 and can include one or at least two processor 920 performs Instruction, to complete all or part of step of above-mentioned method.Additionally, process assembly 902 can include one or at least two Module, it is simple to process between assembly 902 and other assemblies is mutual.Such as, process assembly 902 and can include multi-media module, With facilitate multimedia groupware 908 and process between assembly 902 mutual.
Memorizer 904 is configured to store various types of data to support the operation at device 900.Showing of these data Example include on device 900 operation any storage object or the instruction of method, contact user data, telephone book data, Message, picture, video etc..Memorizer 904 can by any kind of volatibility or non-volatile memory device or they Combination realizes, such as static RAM (SRAM), Electrically Erasable Read Only Memory (EEPROM), erasable can Program read-only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory Reservoir, disk or CD.
The various assemblies that power supply module 906 is device 900 provide power supply.Power supply module 906 can include power management system System, one or at least two power supply, and other generate, manage and distribute, with for device 900, the assembly that power supply is associated.
The screen of one output interface of offer that multimedia groupware 908 is included between described device 900 and user.One In a little embodiments, screen can include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, screen Curtain may be implemented as touch screen, to receive the input signal from user.Touch panel includes one or at least two touch Sensor is with the gesture on sensing touch, slip and touch panel.Described touch sensor can not only sense touch or slide The border of action, but also detect the persistent period relevant to described touch or slide and pressure.In certain embodiments, Multimedia groupware 908 includes a front-facing camera and/or post-positioned pick-up head.When device 900 is in operator scheme, such as shooting mould When formula or video mode, front-facing camera and/or post-positioned pick-up head can receive the multi-medium data of outside.Each preposition shooting Head and post-positioned pick-up head can be a fixing optical lens system or have focal length and optical zoom ability.
Audio-frequency assembly 910 is configured to output and/or input audio signal.Such as, audio-frequency assembly 910 includes a Mike Wind (MIC), when device 900 is in operator scheme, during such as call model, logging mode and loudness recognition mode, mike is joined It is set to receive external audio signal.The audio signal received can be further stored at memorizer 904 or via communication set Part 916 sends.In certain embodiments, audio-frequency assembly 910 also includes a speaker, is used for exporting audio signal.
I/O interface 912 provides interface for processing between assembly 902 and peripheral interface module, above-mentioned peripheral interface module can To be keyboard, put striking wheel, button etc..These buttons may include but be not limited to: home button, volume button, start button and lock Set button.
Sensor cluster 914 includes one or at least two sensor, for providing the state of various aspects for device 900 Assessment.Such as, what sensor cluster 914 can detect device 900 opens/closed mode, the relative localization of assembly, such as institute Stating display and keypad that assembly is device 900, sensor cluster 914 can also detect device 900 or 900 1 groups of device The position change of part, the presence or absence that user contacts with device 900, device 900 orientation or acceleration/deceleration and device 900 Variations in temperature.Sensor cluster 914 can include proximity transducer, is configured to when not having any physical contact examine The existence of object near surveying.Sensor cluster 914 can also include optical sensor, such as CMOS or ccd image sensor, is used for Imaging applications uses.In certain embodiments, this sensor cluster 914 can also include acceleration transducer, and gyroscope passes Sensor, Magnetic Sensor, pressure transducer or temperature sensor.
Communications component 916 is configured to facilitate the communication of wired or wireless mode between device 900 and other equipment.Device 900 can access wireless network based on communication standard, such as WiFi, 2G or 3G, or combinations thereof.An exemplary enforcement In example, communications component 916 receives the broadcast singal from external broadcasting management system or broadcast related information via broadcast channel. In one exemplary embodiment, described communications component 916 also includes near-field communication (NFC) module, to promote junction service.Example As, can be based on RF identification (RFID) technology in NFC module, Infrared Data Association (IrDA) technology, ultra broadband (UWB) technology, Bluetooth (BT) technology and other technologies realize.
In the exemplary embodiment, device 900 can be by one or at least two application specific integrated circuit (ASIC), number Word signal processor (DSP), digital signal processing appts (DSPD), PLD (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic building bricks realize, be used for performing said method.
In the exemplary embodiment, a kind of non-transitory computer-readable recording medium including instruction, example are additionally provided As included the memorizer 904 of instruction, above-mentioned instruction can have been performed said method by the processor 920 of device 900.Such as, Described non-transitory computer-readable recording medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk With optical data storage devices etc..
A kind of non-transitory computer-readable recording medium, when the instruction in described storage medium is by said apparatus 900 When processor performs so that said apparatus 900 is able to carry out a kind of information cuing method, including:
When playing video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
In one embodiment, described method also includes:
When playing described video, obtain frame of video corresponding with described audio frame in described video;
Described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding;
Described people information corresponding for described audio frame is pointed out, including:
When the people information corresponding with described audio frame when the people information that described frame of video is corresponding does not mates, by described sound Frequently the people information that frame is corresponding is pointed out.
In one embodiment, described described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding, Including:
The image of described frame of video is mated with at least one pre-set image;
When the image of described frame of video matches with the target image at least one pre-set image described, by described mesh People information corresponding to logo image is defined as the people information that described frame of video is corresponding.
In one embodiment, described described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding, Including:
Obtain the audio parameter of described audio frame;
The audio parameter of described audio frame is mated with the audio parameter of at least one preset sound;
Audio parameter phase when audio parameter and the target sound at least one preset sound described of described audio frame During coupling, the people information of described target sound is defined as the people information that described audio frame is corresponding.
In one embodiment, described audio parameter includes: at least one in loudness, tone, tone color.
In one embodiment, the people information that described audio frame is corresponding includes: the title of the personage that described audio frame is corresponding Exhale, at least one information in position, contact method.
Art technology user person, after considering description and putting into practice disclosure disclosed herein, will readily occur to the disclosure Other embodiment.The application is intended to any modification, purposes or the adaptations of the disclosure, these modification, purposes Or adaptations follow the disclosure general principle and include the disclosure undocumented in the art known often Know or conventional techniques means.Description and embodiments is considered only as exemplary, and the true scope of the disclosure and spirit are by following Claim point out.
It should be appreciated that the disclosure is not limited to precision architecture described above and illustrated in the accompanying drawings, and And various modifications and changes can carried out without departing from the scope.The scope of the present disclosure is only limited by appended claim.

Claims (13)

1. an information cuing method, it is characterised in that including:
When playing video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
Method the most according to claim 1, it is characterised in that described method also includes:
When playing described video, obtain frame of video corresponding with described audio frame in described video;
Described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding;
Described people information corresponding for described audio frame is pointed out, including:
When the people information corresponding with described audio frame when the people information that described frame of video is corresponding does not mates, by described audio frame Corresponding people information is pointed out.
Method the most according to claim 2, it is characterised in that
Described described frame of video is identified, it is thus achieved that the people information that described frame of video is corresponding, including:
The image of described frame of video is mated with at least one pre-set image;
When the image of described frame of video matches with the target image at least one pre-set image described, by described target figure As corresponding people information is defined as the people information that described frame of video is corresponding.
Method the most according to claim 1, it is characterised in that
Described described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding, including:
Obtain the audio parameter of described audio frame;
The audio parameter of described audio frame is mated with the audio parameter of at least one preset sound;
When the audio parameter of described audio frame matches with the audio parameter of the target sound at least one preset sound described Time, the people information of described target sound is defined as the people information that described audio frame is corresponding.
Method the most according to claim 4, it is characterised in that
Described audio parameter includes: at least one in loudness, tone, tone color.
Method the most according to any one of claim 1 to 5, it is characterised in that
The people information that described audio frame is corresponding includes: in the address of the personage that described audio frame is corresponding, position, contact method At least one information.
7. an information presentation device, it is characterised in that including:
First acquisition module, for when playing video, obtaining the audio frame in described video;
Second acquisition module, for being analyzed described audio frame, it is thus achieved that the people information that described audio frame is corresponding;
Reminding module, for pointing out people information corresponding for described audio frame.
Device the most according to claim 7, it is characterised in that described device also includes:
3rd acquisition module, for when playing described video, obtaining frame of video corresponding with described audio frame in described video;
4th acquisition module, for being identified described frame of video, it is thus achieved that the people information that described frame of video is corresponding;
Described reminding module includes:
First prompting submodule, for when the people information that described frame of video the is corresponding people information corresponding with described audio frame not During coupling, people information corresponding for described audio frame is pointed out.
Device the most according to claim 8, it is characterised in that
Described 4th acquisition module includes:
First matched sub-block, for mating the image of described frame of video with at least one pre-set image;
First determines submodule, for when the image of described frame of video and the target image phase at least one pre-set image described During coupling, people information corresponding for described target image is defined as the people information that described frame of video is corresponding.
Device the most according to claim 7, it is characterised in that
Described second acquisition module includes:
Obtain submodule, for obtaining the audio parameter of described audio frame;
Second matched sub-block, for carrying out the audio parameter of described audio frame with the audio parameter of at least one preset sound Coupling;
Second determines submodule, for when the audio parameter of described audio frame and the target sound at least one preset sound described When the audio parameter of sound matches, the people information of described target sound is defined as the people information that described audio frame is corresponding.
11. devices according to claim 10, it is characterised in that
Described audio parameter includes: at least one in loudness, tone, tone color.
12. according to the device according to any one of claim 7 to 11, it is characterised in that
The people information that described audio frame is corresponding includes: in the address of the personage that described audio frame is corresponding, position, contact method At least one information.
13. 1 kinds of information presentation devices, it is characterised in that including:
Processor;
For storing the memorizer of processor executable;
Wherein, described processor is configured to:
When playing described video, obtain the audio frame in described video;
Described audio frame is analyzed, it is thus achieved that the people information that described audio frame is corresponding;
People information corresponding for described audio frame is pointed out.
CN201610875775.2A 2016-09-30 2016-09-30 Information prompting method and device Active CN106254939B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610875775.2A CN106254939B (en) 2016-09-30 2016-09-30 Information prompting method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610875775.2A CN106254939B (en) 2016-09-30 2016-09-30 Information prompting method and device

Publications (2)

Publication Number Publication Date
CN106254939A true CN106254939A (en) 2016-12-21
CN106254939B CN106254939B (en) 2020-02-07

Family

ID=57612228

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610875775.2A Active CN106254939B (en) 2016-09-30 2016-09-30 Information prompting method and device

Country Status (1)

Country Link
CN (1) CN106254939B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107493442A (en) * 2017-07-21 2017-12-19 北京奇虎科技有限公司 A kind of method and apparatus for editing video
CN108322770A (en) * 2017-11-22 2018-07-24 腾讯科技(深圳)有限公司 Video frequency program recognition methods, relevant apparatus, equipment and system
CN109885701A (en) * 2019-01-08 2019-06-14 广东乐心医疗电子股份有限公司 Multimedia data processing method, system, terminal and computer readable storage medium
CN110337030A (en) * 2019-08-08 2019-10-15 腾讯科技(深圳)有限公司 Video broadcasting method, device, terminal and computer readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120045093A1 (en) * 2010-08-23 2012-02-23 Nokia Corporation Method and apparatus for recognizing objects in media content
CN104796751A (en) * 2015-04-23 2015-07-22 福州大学 Television signal recognition method and device
CN105430494A (en) * 2015-12-02 2016-03-23 百度在线网络技术(北京)有限公司 Method and device for identifying audio from video in video playback equipment
CN105657535A (en) * 2015-12-29 2016-06-08 北京搜狗科技发展有限公司 Audio recognition method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120045093A1 (en) * 2010-08-23 2012-02-23 Nokia Corporation Method and apparatus for recognizing objects in media content
CN103080951A (en) * 2010-08-23 2013-05-01 诺基亚公司 Method and apparatus for recognizing objects in media content
CN104796751A (en) * 2015-04-23 2015-07-22 福州大学 Television signal recognition method and device
CN105430494A (en) * 2015-12-02 2016-03-23 百度在线网络技术(北京)有限公司 Method and device for identifying audio from video in video playback equipment
CN105657535A (en) * 2015-12-29 2016-06-08 北京搜狗科技发展有限公司 Audio recognition method and device

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107493442A (en) * 2017-07-21 2017-12-19 北京奇虎科技有限公司 A kind of method and apparatus for editing video
CN108322770A (en) * 2017-11-22 2018-07-24 腾讯科技(深圳)有限公司 Video frequency program recognition methods, relevant apparatus, equipment and system
WO2019101099A1 (en) * 2017-11-22 2019-05-31 腾讯科技(深圳)有限公司 Video program identification method and device, terminal, system, and storage medium
CN109885701A (en) * 2019-01-08 2019-06-14 广东乐心医疗电子股份有限公司 Multimedia data processing method, system, terminal and computer readable storage medium
CN109885701B (en) * 2019-01-08 2020-01-10 中山乐心电子有限公司 Multimedia data processing method, system, terminal and computer readable storage medium
CN110337030A (en) * 2019-08-08 2019-10-15 腾讯科技(深圳)有限公司 Video broadcasting method, device, terminal and computer readable storage medium
CN110337030B (en) * 2019-08-08 2020-08-11 腾讯科技(深圳)有限公司 Video playing method, device, terminal and computer readable storage medium

Also Published As

Publication number Publication date
CN106254939B (en) 2020-02-07

Similar Documents

Publication Publication Date Title
CN104469437A (en) Advertisement pushing method and device
CN105120191A (en) Video recording method and device
CN106254939A (en) Information cuing method and device
CN104320708A (en) User right handling method and device of smart television
CN106331761A (en) Live broadcast list display method and apparatuses
CN106210757A (en) Live broadcasting method, live broadcast device and live broadcast system
CN106231378A (en) The display packing of direct broadcasting room, Apparatus and system
CN104391711B (en) A kind of method and device that screen protection is set
CN105068976A (en) Ticket information exhibition method and device
US20170090684A1 (en) Method and apparatus for processing information
CN104980662A (en) Method for adjusting imaging style in shooting process, device thereof and imaging device
CN105549300A (en) Automatic focusing method and device
CN105120155A (en) Panoramic photograph generation method and device
CN104243829A (en) Self-shooting method and self-shooting device
CN106453528A (en) Method and device for pushing message
CN105208284A (en) Photographing reminding method and device
CN105704322A (en) Weather information obtaining method and apparatus
CN105657325A (en) Method, apparatus and system for video communication
CN105357449A (en) Shooting method and device, and image processing method and apparatus
CN106559712B (en) Video playing processing method and device and terminal equipment
CN105488829A (en) Method and device for generating head portrait
CN104702988A (en) AV (Audio/Video) playing method and device
CN106791535A (en) Video recording method and device
CN105095213B (en) Information correlation method and device
CN108259974A (en) Video matching method and device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant