CN106503127A - The music data processing method recognized based on facial action and system - Google Patents

The music data processing method recognized based on facial action and system Download PDF

Info

Publication number
CN106503127A
CN106503127A CN201610912440.3A CN201610912440A CN106503127A CN 106503127 A CN106503127 A CN 106503127A CN 201610912440 A CN201610912440 A CN 201610912440A CN 106503127 A CN106503127 A CN 106503127A
Authority
CN
China
Prior art keywords
data
music
foreground
trifle
facial action
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610912440.3A
Other languages
Chinese (zh)
Other versions
CN106503127B (en
Inventor
简仁贤
何芳琳
赵伟翔
于庭婕
黄品瑞
廖健宏
陈智凯
孙廷伟
杨闵淳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intelligent Technology (shanghai) Co Ltd
Original Assignee
Intelligent Technology (shanghai) Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intelligent Technology (shanghai) Co Ltd filed Critical Intelligent Technology (shanghai) Co Ltd
Priority to CN201610912440.3A priority Critical patent/CN106503127B/en
Publication of CN106503127A publication Critical patent/CN106503127A/en
Application granted granted Critical
Publication of CN106503127B publication Critical patent/CN106503127B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality

Abstract

The invention provides a kind of music data processing method recognized based on facial action and system, method is:Background music data and foreground music data are obtained, foreground music data are divided into multistage trifle by beat, include multiple bats per trifle;The face action of people is detected within the foreground music time, obtains multiple facial action data, the foreground music data of the corresponding trifle time span of each facial action;Background music data is persistently played, and the foreground music data of a trifle time span corresponding for each facial action data are mated, and combined with background music, generated new music.Music data processing method and system that the present invention is recognized based on facial action, employ based on face's key point technology of identification, the mode that music data and real image are combined, realize interacting for people and corresponding scene, need not be aided in by outside setting, implementation is simple, improves Consumer's Experience.

Description

The music data processing method recognized based on facial action and system
Technical field
The present invention relates to data processing field, more particularly to the music data process recognized based on facial action.
Background technology
In prior art, it is (such as Wii, dancing machine using mouse, keyboard, joystick, Touch Screen, external sensor mostly Pedal etc.), the mode such as human posture (such as Kinect) operate, realize interacting for people and corresponding scene.Wherein, human posture's operation Mode, needs the information such as deep space, it is necessary to using specific device;It is thus desirable to detecting whole body, needs larger space ability Game.In addition there is the problems such as accuracy rate is low, side-play amount is high, cause the interesting reduction in interaction;In addition such shape Scene interactivity is most to detect (example for human action:Raise one's hand, skirt), less combined with real image.
The application of face's key point identification at present is mainly combined with image composing technique, and face is become other animals, worn Upper different ornaments simultaneously combine facial action generation animation etc..
Thinning face is provided with the application during facial movements such as face's rehabilitation:Traditional facial movement through word flow process description or Reach the effect of facial movement through the face motion demonstration in study film, but such facial movement process not with true Image is combined, and Experience Degree is low.
Therefore, defect of the prior art is, for interacting for people and corresponding scene, needs real by external equipment auxiliary Existing, implementation is single, and Consumer's Experience is low.Face's key point technology of identification and real image cannot be combined.
Content of the invention
For above-mentioned technical problem, the present invention is provided a kind of music data processing method recognized based on facial action and is System, employ based on face's key point technology of identification, the mode that music data and real image are combined, realize people with corresponding The interaction of scene, it is not necessary to aided in by outside setting, implementation is simple, improves Consumer's Experience.
For solving above-mentioned technical problem, the technical scheme that the present invention is provided is:
In a first aspect, the present invention provides a kind of music data processing method recognized based on facial action, including:
Step S1, obtains background music data and foreground music data, the background music data and the foreground music Data are respectively the music of a hop count second to several minutes length;
The foreground music data are divided into multistage trifle by beat by step S2, include multiple bats per trifle;
Step S3, detects the face action of people within the foreground music time, and acquisition is more within the foreground music time Individual facial action data, the foreground music data of the corresponding trifle time span of each facial action;
Step S4, the background music data is persistently played, and will be right with which for each facial action data described The foreground music data of the trifle time span that answers are mated, and are combined with the background music, generate new music.
The technical scheme is that:Background music data and foreground music data, the background music data are first obtained The music of a hop count second to several minutes length is respectively with the foreground music data;Then by the foreground music data by section Bat is divided into multistage trifle, includes multiple bats per trifle;
Then the face action of people, acquisition multiple faces during foreground music time are detected within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;Finally by the background music number According to persistently being played, and the foreground music by the trifle time span corresponding for each facial action data described Data are mated, and are combined with the background music, generate new music.
The music data processing method that the present invention is recognized based on facial action, employs based on face's key point identification skill Art, the mode that music data and real image are combined realize interacting for people and corresponding scene, it is not necessary to by outside setting Aided in, implementation is simple, improves Consumer's Experience.
Further, after step S2, also include:
The background music data is persistently played, target face action data, target face action data is obtained Corresponding unique trifle foreground music;
According to the target face action data, obtain and move in the face started per trifle in before and after's bit time Make data;
The facial action data are carried out mating judgement with the target face action data, a trifle foreground voice is carried out The happy selection that plays:
When the facial action data are mated with the target face action data, the trifle foreground music is played, The one trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data are mismatched with the target face action data, the facial action data are not played Corresponding trifle foreground music.
Further, after step S2, also include:
The background music data is persistently played, obtain virtual scene data, the virtual scene data be to The virtual scene data of each position mobile object in human face;
According to the virtual scene data of each position mobile object in human face, corresponding facial action number is obtained According to the acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data are carried out mating judgement with the virtual scene data, the virtual scene data are carried out In corresponding mobile object process:
When the facial action data and the virtual scene Data Matching, will correspondingly move in the virtual scene data Object is removed;
When the facial action data are mismatched with the virtual scene data, to correspondingly moving in the virtual scene data Animal body does not deal with;
After the rear bit time started per trifle, without corresponding facial action and target face action data Match somebody with somebody, corresponding mobile object in the virtual scene data is removed.
Further, will after in the virtual scene data, corresponding mobile object is removed, including:
Obtain the effect data that corresponding mobile object is removed in the virtual scene data;
According to the effect data that corresponding mobile object in the virtual scene data is removed, to the virtual scene data The facial action data of coupling are evaluated, and obtain evaluation result.
Further, the identification of facial action is carried out by the identification of face key point and fuzzy control theory.
A kind of music data processing method recognized based on facial action of the present invention, based on face's key point technology of identification, The mode that music data and real image are combined, realizes interacting for people and corresponding scene, i.e., the knowledge being failure to actuate by face Not, mated with the music data of corresponding scene, realized the creation of music, corresponding virtual objects in the broadcasting of music and scene Elimination, be presented in user in animated way at the moment, it is not necessary to aided in by outside setting, implementation is simple, carries High Consumer's Experience.
Further, also include:
The background music data is persistently played, target face action data, the target face action is obtained The corresponding trifle foreground music of data, the foreground music is divided into the first foreground music and the second foreground music, before described first Scape music mates broadcasting with the background music, and second foreground music is mismatched with the background music and played;
According to the target face action data, obtain and move in the face started per trifle in before and after's bit time Make data;
The facial action data are carried out mating judgement with the target face action data, the target face is carried out The broadcasting of the corresponding trifle foreground music of action data is chosen:
When the facial action data are mated with the target face action data, the target face action data is played A corresponding trifle foreground music, and the trifle foreground music is the first prospect corresponding with the target face action data Music;
When the facial action data are mismatched with the target face action data, the target face action number is played According to a corresponding trifle foreground music, and the trifle foreground music be with before the target face action data corresponding second Scape music.
After facial action data mate target face action data successful match with subject performance data, corresponding broadcasting should The corresponding trifle foreground music of target face action, also, the first foreground music and background music match the broadcasting of harmony, , whereas if the facial action data that user makes are mismatched with target face action data, then correspondingly play target face and move Make corresponding second foreground music, the second foreground music does not match that harmonious broadcasting with background music, so by playing The difference of music can judge facial action data that user makes, and whether the match is successful with target face action data, improves Consumer's Experience.
Second aspect, the invention provides a kind of music data processing system recognized based on facial action, including:
Music data acquisition module, for obtaining background music data and foreground music data, the background music data The music of a hop count second to several minutes length is respectively with the foreground music data;
Music data processing module, for the foreground music data are divided into multistage trifle by beat, includes per trifle Multiple bats;
Facial action acquisition module, for detecting the face action of people within the foreground music time, obtains in the prospect Multiple facial action data in musical time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module, for persistently being played the background music data, and by each facial action described The foreground music data of the corresponding trifle time span of data are mated, and are combined with the background music, raw The music of Cheng Xin.
The technical scheme is that:Music data acquisition module is first passed through, background music data and foreground music is obtained Data, the background music data and the foreground music data are respectively the music of a hop count second to several minutes length;Then By music data processing module, the foreground music data are divided into multistage trifle by beat, include multiple bats per trifle;
Facial action acquisition module is then passed through, and the face action of people is detected within the foreground music time, is obtained described Multiple facial action data in the foreground music time, the foreground music number of the corresponding trifle time span of each facial action According to;Finally by musical composition module, the background music data is persistently played, and by each facial action number described Mated according to the foreground music data of the corresponding trifle time span, combined with the background music, generated New music.
A kind of music data processing system recognized based on facial action of the present invention, is employed based on face's key point identification Technology, the mode that music data and real image are combined realize interacting for people and corresponding scene, it is not necessary to set by outside Put and aided in, implementation is simple, improves Consumer's Experience.
Further, after the music data processing module, also include the selection of music playing module, be used for:
The background music data is persistently played, target face action data, target face action data is obtained Corresponding unique trifle foreground music;
According to the target face action data, obtain and move in the face started per trifle in before and after's bit time Make data;
The facial action data are carried out mating judgement with the target face action data, a trifle foreground voice is carried out The happy selection that plays:
When the facial action data are mated with the target face action data, the trifle foreground music is played, The one trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data are mismatched with the target face action data, the facial action data are not played Corresponding trifle foreground music.
Further, after the music data processing module, also include music virtual scene module, be used for:
The background music data is persistently played, obtain virtual scene data, the virtual scene data be to The virtual scene data of each position mobile object in human face;
According to the virtual scene data of each position mobile object in human face, corresponding facial action number is obtained According to the acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data are carried out mating judgement with the virtual scene data, the virtual scene data are carried out In corresponding mobile object process:
When the facial action data and the virtual scene Data Matching, will correspondingly move in the virtual scene data Object is removed;
When the facial action data are mismatched with the virtual scene data, to correspondingly moving in the virtual scene data Animal body does not deal with;
After the rear bit time started per trifle, without corresponding facial action and target face action data Match somebody with somebody, corresponding mobile object in the virtual scene data is removed.
Further, in the music virtual scene module, including effect assessment submodule, by the virtual scene number After according in, corresponding mobile object is removed, the effect assessment submodule is used for:
Obtain the effect data that corresponding mobile object is removed in the virtual scene data;
According to the effect data that corresponding mobile object in the virtual scene data is removed, to the virtual scene data The facial action data of coupling are evaluated, and obtain evaluation result.
Description of the drawings
In order to be illustrated more clearly that the specific embodiment of the invention or technical scheme of the prior art, below will be to concrete Needed for embodiment or description of the prior art, accompanying drawing to be used is briefly described.
Fig. 1 shows a kind of music data process side recognized based on facial action that first embodiment of the invention is provided The flow chart of method;
Fig. 2 shows a kind of music data process side recognized based on facial action that first embodiment of the invention is provided First schematic diagram of time shafts in method;
Fig. 3 shows a kind of music data process side recognized based on facial action that first embodiment of the invention is provided Second schematic diagram of time shafts in method;
Fig. 4 shows a kind of music data processing system recognized based on facial action that second embodiment of the invention is provided The schematic diagram of system.
Specific embodiment
The embodiment of technical solution of the present invention is described in detail below in conjunction with accompanying drawing.Following examples are only used for Technical scheme is described clearly, is therefore intended only as example, and the protection of the present invention can not be limited with this Scope.
Embodiment one
Fig. 1 shows a kind of music data process side recognized based on facial action that first embodiment of the invention is provided The flow chart of method;As shown in figure 1, the embodiment of the present invention one provides a kind of music data process side recognized based on facial action Method, including:
Step S1, obtains background music data and foreground music data, background music data and foreground music data difference Music for a hop count second to several minutes length;
Foreground music data are divided into multistage trifle by beat by step S2, include multiple bats per trifle;
Step S3, detects the face action of people, multiple faces of the acquisition within the foreground music time within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;
Step S4, background music data is persistently played, and little by corresponding for each facial action data one The foreground music data of section time span are mated, and are combined with background music, generate new music.
The technical scheme is that:Background music data and foreground music data are first obtained, background music data and front Scape music data is respectively the music of a hop count second to several minutes length;Then foreground music data are divided into multistage by beat little Section, includes multiple bats per trifle;
Then the face action of people is detected within the foreground music time, and multiple faces of the acquisition within the foreground music time are moved Make data, the foreground music data of the corresponding trifle time span of each facial action;Finally background music data is held Continued broadcasting is put, and the foreground music data of a trifle time span corresponding for each facial action data are mated, with Background music is combined, and generates new music.
The music data processing method that the present invention is recognized based on facial action, can arrange different scene informations, different Scene information have different demands, need to complete according to different demands in scene information at the appointed time, if different Scene information mate from different facial actions, then can realize different actions, and then realize the interaction of various people and scene.
The music data processing method that the present invention is recognized based on facial action, employs based on face's key point identification skill Art, the mode that music data and real image are combined realize interacting for people and corresponding scene, it is not necessary to by outside setting Aided in, implementation is simple, improves Consumer's Experience.
Specifically, after step S2, also include:
Background music data is persistently played, target face action data is obtained, target face action data is corresponding Unique trifle foreground music;
According to target face action data, obtain and start the facial action data in before and after's bit time in every trifle;
Facial action data and target face action data are carried out mating judgement, trifle foreground music broadcasting is carried out Choose:
When facial action data are mated with target face action data, a trifle foreground music, a trifle foreground voice is played Pleasure is and the unique corresponding foreground music of target face action data;
When facial action data are mismatched with target face action data, the corresponding trifle of facial action data is not played Foreground music.
Specifically, the foreground music of many trifles is set, and per trifle foreground music, corresponding only one facial action, corresponding After the match is successful in first facial action and first aim facial action, next facial action is carried out with next target The coupling of facial action, so can realize the continuous broadcasting of foreground music according to different target face actions.
Specifically, after step S2, also include:
Background music data is persistently played, obtain virtual scene data, virtual scene data be to human face in The virtual scene data of each position mobile object;
According to the virtual scene data of each position mobile object in human face, corresponding facial action data are obtained, The acquisition of corresponding facial action data is before the position that mobile object reaches in human face;
Facial action data and virtual scene data are carried out mating judgement, corresponding motive objects in virtual scene data are carried out The process of body:
When facial action data and virtual scene Data Matching, corresponding mobile object in virtual scene data is removed;
When facial action data are mismatched with virtual scene data, corresponding mobile object in virtual scene data is not made to locate Reason;
After the rear bit time that every trifle starts, mate with target face action data without corresponding facial action, Corresponding mobile object in virtual scene data is removed.
In conjunction with AR technology, in real scene, corresponding face facial positions arrange different virtual objects and move to face, Before and after the time point that virtual objects are moved in face facial positions, bit time is event horizon, correspondingly makes each virtual The corresponding elimination action of article, the facial action done by people and elimination action are then done next action, are increased after the match is successful Interest, while the match is successful or unsuccessfully has different audios and animation.Through after a period of time, match time shortens, Coupling eliminates speed of action and can accelerate.
Specifically, will after in virtual scene data, corresponding mobile object is removed, including:
Obtain the effect data that corresponding mobile object is removed in virtual scene data;
According to the effect data that corresponding mobile object in virtual scene data is removed, to the face with virtual scene Data Matching Portion's action data is evaluated, and obtains evaluation result.
As shown in Fig. 2 transverse axis is time shafts, time, the right represent later time earlier for left side representative.Relatively thick length The longitudinal axis is trifle separation, and the separation of point is clapped in shorter representative.In this legend, it is before every trifle starts to drive scope Two bats are all correct driving, and 0.5 bat before and after starting previous bat with every trifle is that perfection drives, in such as Fig. 2, the part that e shows, and e Before and after 0.5 clap as common driving, the part that such as f shows.Perfect, common driving time and scope all freely can be replaced.With The method is doing the evaluation criterion to virtual scene Data Matching facial action data.
Specifically, the identification of facial action is carried out by the identification of face key point and fuzzy control theory.
In the present invention, processed based on above-mentioned music data, be based on the identification of face key point, set up a reliability Face system, finally accurately recognizes facial action with reference to fuzzy control theory:Blink, cross-eye, choose eyebrow, frown, the nose that wrinkles, Loll, mouth of beeping, open one's mouth, wapperijaw, lick lip, close lightly lip, nod, head left rotation and right rotation, head rotate etc. up and down.Wherein face is crucial Technology of identification is put for known prior art, here does not do excessive narration.
Specifically, also include:
Background music data is persistently played, target face action data is obtained, target face action data is corresponding One trifle foreground music, foreground music are divided into the first foreground music and the second foreground music, the first foreground music and background music Coupling is played, and the second foreground music is mismatched with background music and played;
According to target face action data, obtain and start the facial action data in before and after's bit time in every trifle;
Facial action data and target face action data are carried out mating judgement, target face action data is carried out corresponding The broadcasting of one trifle foreground music is chosen:
When facial action data are mated with target face action data, before playing the corresponding trifle of target face action data Scape music, and a trifle foreground music is the first foreground music corresponding with target face action data;
When facial action data are mismatched with target face action data, the corresponding trifle of target face action data is played Foreground music, and a trifle foreground music is the second foreground music corresponding with target face action data.
After facial action data mate target face action data successful match with subject performance data, corresponding broadcasting should The corresponding trifle foreground music of target face action, also, the first foreground music and background music match the broadcasting of harmony, , whereas if the facial action data that user makes are mismatched with target face action data, then correspondingly play target face and move Make corresponding second foreground music, the second foreground music does not match that harmonious broadcasting with background music, so by playing The difference of music can judge facial action data that user makes, and whether the match is successful with target face action data, improves Consumer's Experience.
Specifically, the first foreground music may be configured as representing that successful music, the second foreground music can be set to represent and lose The foreground music for losing, so makes music more having any different property.
As shown in figure 3, transverse axis is time shafts, time, the right represent later time earlier for left side representative.Relatively thick length The longitudinal axis is trifle separation, and the separation of point is clapped in shorter representative.Example in Fig. 3, the scope of c is a trifle, the model of d Enclose for a bat.A for detecting time point, represent in Fig. 3 every trifle start before and after clap for detect time point, any at this moment Between put the facial action that does and can be detected.B is sphere of action, generally in units of a trifle, represents here in this figure little Do in the range of a bat before and after section and express one's feelings, can all drive this trifle to make feedback.Per a few bats of trifle, detect time point and effect model Enclose and all freely can replace.
A kind of music data processing method recognized based on facial action of the present invention, based on face's key point technology of identification, The mode that music data and real image are combined, realizes interacting for people and corresponding scene, i.e., the knowledge being failure to actuate by face Not, mated with the music data of corresponding scene, realized the creation of music, corresponding virtual objects in the broadcasting of music and scene Elimination, be presented in user in animated way at the moment, it is not necessary to aided in by outside setting, implementation is simple, carries High Consumer's Experience.
Embodiment two
Fig. 4 shows a kind of music data processing system recognized based on facial action that second embodiment of the invention is provided The schematic diagram of system;As shown in figure 4, the embodiment of the present invention two provides a kind of music data processing system recognized based on facial action System 10, including:
Music data acquisition module 101, for obtaining background music data and foreground music data, background music data and Foreground music data are respectively the music of a hop count second to several minutes length;
Music data processing module 102, for foreground music data are divided into multistage trifle by beat, per trifle comprising many Individual bat;
Facial action acquisition module 103, for detecting the face action of people within the foreground music time, obtains in foreground voice Multiple facial action data in the happy time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module 104, for background music data is persistently played, and by each facial action data with The foreground music data of its corresponding trifle time span are mated, and are combined with background music, generate new music.
The technical scheme is that:Music data acquisition module 101 is first passed through, background music data and foreground voice is obtained Happy data, background music data and foreground music data are respectively the music of a hop count second to several minutes length;Then pass through sound Foreground music data are divided into multistage trifle by beat by happy data processing module 102, include multiple bats per trifle;
Facial action acquisition module 103 is then passed through, and the face action of people is detected within the foreground music time, is obtained front Multiple facial action data in scape musical time, the foreground music data of the corresponding trifle time span of each facial action; Finally by musical composition module 104, background music data is persistently played, and will be right with which for each facial action data The foreground music data of the trifle time span that answers are mated, and are combined with background music, generate new music.
A kind of music data processing system 10 recognized based on facial action of the present invention, is employed and is known based on face's key point Other technology, the mode that music data and real image are combined realize interacting for people and corresponding scene, it is not necessary to by outside Setting is aided in, and implementation is simple, improves Consumer's Experience.
Specifically, after music data processing module 102, also include the selection of music playing module, be used for:
Background music data is persistently played, target face action data is obtained, target face action data is corresponding Unique trifle foreground music;
According to target face action data, obtain and start the facial action data in before and after's bit time in every trifle;
Facial action data and target face action data are carried out mating judgement, trifle foreground music broadcasting is carried out Choose:
When facial action data are mated with target face action data, a trifle foreground music, a trifle foreground voice is played Pleasure is and the unique corresponding foreground music of target face action data;
When facial action data are mismatched with target face action data, the corresponding trifle of facial action data is not played Foreground music.
Specifically, after music data processing module 102, also include music virtual scene module, be used for:
Background music data is persistently played, obtain virtual scene data, virtual scene data be to human face in The virtual scene data of each position mobile object;
According to the virtual scene data of each position mobile object in human face, corresponding facial action data are obtained, The acquisition of corresponding facial action data is before the position that mobile object reaches in human face;
Facial action data and virtual scene data are carried out mating judgement, corresponding motive objects in virtual scene data are carried out The process of body:
When facial action data and virtual scene Data Matching, corresponding mobile object in virtual scene data is removed;
When facial action data are mismatched with virtual scene data, corresponding mobile object in virtual scene data is not made to locate Reason;
After the rear bit time that every trifle starts, mate with target face action data without corresponding facial action, Corresponding mobile object in virtual scene data is removed.
Specifically, in music virtual scene module 104, including effect assessment submodule, will be right in virtual scene data After answering mobile object to remove, effect assessment submodule is used for:
Obtain the effect data that corresponding mobile object is removed in virtual scene data;
According to the effect data that corresponding mobile object in virtual scene data is removed, to the face with virtual scene Data Matching Portion's action data is evaluated, and obtains evaluation result.
Specifically, also include facial action identification module 100, for recognizing by face key point and fuzzy control theory Carry out the identification of facial action.
Specifically, also include the selection of music playing module, be used for:
Background music data is persistently played, target face action data is obtained, target face action data is corresponding One trifle foreground music, foreground music are divided into the first foreground music and the second foreground music, the first foreground music and background music Coupling is played, and the second foreground music is mismatched with background music and played;
According to target face action data, obtain and start the facial action data in before and after's bit time in every trifle;
Facial action data and target face action data are carried out mating judgement, target face action data is carried out corresponding The broadcasting of one trifle foreground music is chosen:
When facial action data are mated with target face action data, before playing the corresponding trifle of target face action data Scape music, and a trifle foreground music is the first foreground music corresponding with target face action data;
When facial action data are mismatched with target face action data, the corresponding trifle of target face action data is played Foreground music, and a trifle foreground music is the second foreground music corresponding with target face action data.
After facial action data mate target face action data successful match with subject performance data, corresponding broadcasting should The corresponding trifle foreground music of target face action, also, the first foreground music and background music match the broadcasting of harmony, , whereas if the facial action data that user makes are mismatched with target face action data, then correspondingly play target face and move Make corresponding second foreground music, the second foreground music does not match that harmonious broadcasting with background music, so by playing The difference of music can judge facial action data that user makes, and whether the match is successful with target face action data, improves Consumer's Experience.
Specifically, the first foreground music may be configured as representing that successful music, the second foreground music can be set to represent and lose The foreground music for losing, so makes music more having any different property.
A kind of music data processing system recognized based on facial action of the present invention, based on face's key point technology of identification, The mode that music data and real image are combined, realizes interacting for people and corresponding scene, i.e., by the knowledge of face action Not, mated with the music data of corresponding scene, realized the creation of music, corresponding virtual objects in the broadcasting of music and scene Elimination, be presented in user in animated way at the moment, it is not necessary to aided in by outside setting, implementation is simple, carries High Consumer's Experience.
Embodiment three
A kind of music data processing method recognized based on facial action in conjunction with the embodiment of the present invention one, and embodiment A kind of music data processing system recognized based on facial action in two, is illustrated in conjunction with specific scene of game.
Scene one
Melody is created:Before background music is persistently played, and each facial action corresponds to one section of one trifle time span Scape music, in the facial action that every trifle starts to detect in before and after's bit time, can all drive the trifle to play corresponding Foreground music.A series of facial actions are such as set, blink, choose eyebrow, frown, the nose that wrinkles etc., then recognize the facial action of people, root According to the difference of the facial action of identification, the different music of corresponding broadcasting, the trifle music that plays every time constitute different music, User is made to do different music according to the mood of oneself wound.
Scene two
Music game:Background music and foreground music are the music of a hop count minute, and background music is persistently played, play according to So in units of trifle, specified facial action is randomly generated by game per little festival-gathering, player must be in driving time (per trifle Bit time before and after beginning) complete to specify facial action, the foreground music of the trifle persistently can play;If there is more than one prospect Music, also may done correct expression more multiple after by music be superimposed up.Motivational music is combined with human face action, trip is made Play is challenging and interesting.
Scene three
AR Rhythmic games:Background music is persistently played, and having different articles on picture has past each portion on the face of rhythm Displacement is moved, and player must do corresponding expression and remove the article when article moves to the position.For example:Mosquito flies toward eyes Past, it is necessary to which it is killed by blink when mosquito reaches eyes.Success (be divided into perfection, common) unsuccessfully has different sounds Effect, animation and score.Through after a period of time, rhythm tempo can be accelerated, increase game degree of difficulty.
The image for directly being obtained using photographic head is made face's action recognition to operate game, and can in real time with true face's shadow Interesting animation effect is produced as combining.This mode of operation is not required to other purchase of equipment, you can differentiates face in real time, exactly and moves Make;And do not need extremity mode of operation, can also benefit handicapped people, allow them also enjoy the enjoyment of game.
Finally it should be noted that:Various embodiments above only in order to technical scheme to be described, rather than a limitation;To the greatest extent Pipe has been described in detail to the present invention with reference to foregoing embodiments, it will be understood by those within the art that:Its according to So the technical scheme described in foregoing embodiments can be modified, or which part or all technical characteristic are entered Row equivalent;And these modifications or replacement, do not make the essence of appropriate technical solution depart from various embodiments of the present invention technology The scope of scheme, its all should cover in the middle of the scope of the claim and description of the present invention.

Claims (10)

1. the music data processing method for being recognized based on facial action, it is characterised in that include:
Step S1, obtains background music data and foreground music data, the background music data and the foreground music data The music of a respectively hop count second to several minutes length;
The foreground music data are divided into multistage trifle by beat by step S2, include multiple bats per trifle;
Step S3, detects the face action of people, multiple faces of the acquisition within the foreground music time within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;
Step S4, the background music data is persistently played, and will be corresponding for each facial action data described The foreground music data of the one trifle time span are mated, and are combined with the background music, generate new music.
2. the music data processing method for being recognized based on facial action according to claim 1, it is characterised in that the step After S2, also include:
The background music data is persistently played, target face action data is obtained, target face action data is corresponding Unique trifle foreground music;
According to the target face action data, obtain and start the facial action number in before and after's bit time described per trifle According to;
The facial action data are carried out mating judgement with the target face action data, a trifle foreground music is carried out and is broadcast The selection that puts:
When the facial action data are mated with the target face action data, the trifle foreground music is played, described One trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data are mismatched with the target face action data, the facial action data are not played corresponding A trifle foreground music.
3. the music data processing method for being recognized based on facial action according to claim 1, it is characterised in that
After step S2, also include:
The background music data is persistently played, virtual scene data are obtained, the virtual scene data are to people face The virtual scene data of each position mobile object in portion;
According to the virtual scene data of each position mobile object in human face, corresponding facial action data are obtained, The acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data are carried out mating judgement with the virtual scene data, it is right in the virtual scene data to carry out Answer the process of mobile object:
When the facial action data and the virtual scene Data Matching, by corresponding mobile object in the virtual scene data Remove;
When the facial action data are mismatched with the virtual scene data, to corresponding motive objects in the virtual scene data Body does not deal with;
After the rear bit time started per trifle, mate with target face action data without corresponding facial action, Corresponding mobile object in the virtual scene data is removed.
4. the music data processing method for being recognized based on facial action according to claim 3, it is characterised in that
After in the virtual scene data, corresponding mobile object is removed, including:
Obtain the effect data that corresponding mobile object is removed in the virtual scene data;
According to the effect data that corresponding mobile object in the virtual scene data is removed, to the virtual scene Data Matching The facial action data evaluated, obtain evaluation result.
5. the music data processing method for being recognized based on facial action according to claim 1, it is characterised in that
The identification that facial action is carried out by the identification of face key point and fuzzy control theory.
6. according to claim 1 or claim 2 based on facial action recognize music data processing method, it is characterised in that
Also include:
The background music data is persistently played, target face action data, the target face action data is obtained A corresponding trifle foreground music, the foreground music are divided into the first foreground music and the second foreground music, first foreground voice Pleasure mates broadcasting with the background music, and second foreground music is mismatched with the background music and played;
According to the target face action data, obtain and start the facial action number in before and after's bit time described per trifle According to;
The facial action data are carried out mating judgement with the target face action data, the target face action is carried out The broadcasting of the corresponding trifle foreground music of data is chosen:
When the facial action data are mated with the target face action data, the target face action data is played corresponding One trifle foreground music, and the trifle foreground music is the first foreground voice corresponding with the target face action data Happy;
When the facial action data are mismatched with the target face action data, the target face action data pair is played A trifle foreground music is answered, and the trifle foreground music is the second foreground voice corresponding with the target face action data Happy.
7. the music data processing system for being recognized based on facial action, it is characterised in that include:
Music data acquisition module, for obtaining background music data and foreground music data, the background music data and institute State the music that foreground music data are respectively a hop count second to several minutes length;
Music data processing module, for the foreground music data are divided into multistage trifle by beat, per trifle comprising multiple Bat;
Facial action acquisition module, for detecting the face action of people within the foreground music time, obtains in the foreground music Multiple facial action data in time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module, for persistently being played the background music data, and by each facial action data described The foreground music data of the corresponding trifle time span are mated, and are combined with the background music, generate new Music.
8. the music data processing system for being recognized based on facial action according to claim 7, it is characterised in that
After the music data processing module, also include the selection of music playing module, be used for:
The background music data is persistently played, target face action data is obtained, target face action data is corresponding Unique trifle foreground music;
According to the target face action data, obtain and start the facial action number in before and after's bit time described per trifle According to;
The facial action data are carried out mating judgement with the target face action data, a trifle foreground music is carried out and is broadcast The selection that puts:
When the facial action data are mated with the target face action data, the trifle foreground music is played, described One trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data are mismatched with the target face action data, the facial action data are not played corresponding A trifle foreground music.
9. the music data processing system for being recognized based on facial action according to claim 7, it is characterised in that
After the music data processing module, also include music virtual scene module, be used for:
The background music data is persistently played, virtual scene data are obtained, the virtual scene data are to people face The virtual scene data of each position mobile object in portion;
According to the virtual scene data of each position mobile object in human face, corresponding facial action data are obtained, The acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data are carried out mating judgement with the virtual scene data, it is right in the virtual scene data to carry out Answer the process of mobile object:
When the facial action data and the virtual scene Data Matching, by corresponding mobile object in the virtual scene data Remove;
When the facial action data are mismatched with the virtual scene data, to corresponding motive objects in the virtual scene data Body does not deal with;
After the rear bit time started per trifle, mate with target face action data without corresponding facial action, Corresponding mobile object in the virtual scene data is removed.
10. the music data processing system for being recognized based on facial action according to claim 9, it is characterised in that
In the music virtual scene module, including effect assessment submodule, will correspondingly move in the virtual scene data After object is removed, the effect assessment submodule is used for:
Obtain the effect data that corresponding mobile object is removed in the virtual scene data;
According to the effect data that corresponding mobile object in the virtual scene data is removed, to the virtual scene Data Matching The facial action data evaluated, obtain evaluation result.
CN201610912440.3A 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification Active CN106503127B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610912440.3A CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610912440.3A CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Publications (2)

Publication Number Publication Date
CN106503127A true CN106503127A (en) 2017-03-15
CN106503127B CN106503127B (en) 2019-09-27

Family

ID=58294244

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610912440.3A Active CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Country Status (1)

Country Link
CN (1) CN106503127B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108905193A (en) * 2018-07-03 2018-11-30 百度在线网络技术(北京)有限公司 Game manipulates processing method, equipment and storage medium
CN109343770A (en) * 2018-09-27 2019-02-15 腾讯科技(深圳)有限公司 Interaction feedback method, equipment and recording medium
CN110047520A (en) * 2019-03-19 2019-07-23 北京字节跳动网络技术有限公司 Control method, device, electronic equipment and the computer readable storage medium that audio plays

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764940A (en) * 2003-03-31 2006-04-26 索尼株式会社 Tempo analysis device and tempo analysis method
CN101836219A (en) * 2007-11-01 2010-09-15 索尼爱立信移动通讯有限公司 Generating music playlist based on facial expression
CN102640149A (en) * 2009-12-04 2012-08-15 索尼计算机娱乐公司 Music recommendation system, information processing device, and information processing method
CN102880388A (en) * 2012-09-06 2013-01-16 北京天宇朗通通信设备股份有限公司 Music processing method, music processing device and mobile terminal
CN102929476A (en) * 2012-09-06 2013-02-13 北京天宇朗通通信设备股份有限公司 Method and device for controlling main menu of terminal
CN103383694A (en) * 2012-12-14 2013-11-06 李博文 Method and system for organizing, managing and marking music document
CN104851435A (en) * 2015-06-06 2015-08-19 孔霞 Music intelligent playing method based on network communication
CN105518783A (en) * 2013-08-19 2016-04-20 谷歌公司 Content-based video segmentation

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764940A (en) * 2003-03-31 2006-04-26 索尼株式会社 Tempo analysis device and tempo analysis method
CN101836219A (en) * 2007-11-01 2010-09-15 索尼爱立信移动通讯有限公司 Generating music playlist based on facial expression
CN102640149A (en) * 2009-12-04 2012-08-15 索尼计算机娱乐公司 Music recommendation system, information processing device, and information processing method
CN102880388A (en) * 2012-09-06 2013-01-16 北京天宇朗通通信设备股份有限公司 Music processing method, music processing device and mobile terminal
CN102929476A (en) * 2012-09-06 2013-02-13 北京天宇朗通通信设备股份有限公司 Method and device for controlling main menu of terminal
CN103383694A (en) * 2012-12-14 2013-11-06 李博文 Method and system for organizing, managing and marking music document
CN105518783A (en) * 2013-08-19 2016-04-20 谷歌公司 Content-based video segmentation
CN104851435A (en) * 2015-06-06 2015-08-19 孔霞 Music intelligent playing method based on network communication

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108905193A (en) * 2018-07-03 2018-11-30 百度在线网络技术(北京)有限公司 Game manipulates processing method, equipment and storage medium
CN108905193B (en) * 2018-07-03 2022-04-15 百度在线网络技术(北京)有限公司 Game manipulation processing method, device and storage medium
CN109343770A (en) * 2018-09-27 2019-02-15 腾讯科技(深圳)有限公司 Interaction feedback method, equipment and recording medium
CN110047520A (en) * 2019-03-19 2019-07-23 北京字节跳动网络技术有限公司 Control method, device, electronic equipment and the computer readable storage medium that audio plays

Also Published As

Publication number Publication date
CN106503127B (en) 2019-09-27

Similar Documents

Publication Publication Date Title
CN102473320B (en) Bringing a visual representation to life via learned input from the user
Collins Playing with sound: a theory of interacting with sound and music in video games
CN102129343B (en) Directed performance in motion capture system
US9519989B2 (en) Visual representation expression based on player expression
Johnson et al. Sympathetic interfaces: using a plush toy to direct synthetic characters
TWI377055B (en) Interactive rehabilitation method and system for upper and lower extremities
CN102822869B (en) Capture view and the motion of the performer performed in the scene for generating
JP2019532374A5 (en)
Bode Making believe: Screen performance and special effects in popular cinema
WO2009021124A2 (en) System and method for a motion sensing amusement device
CN106503127B (en) Music data processing method and system based on facial action identification
JP7355483B2 (en) Game systems and programs
JP2018075260A (en) Game system and program
JP2014023745A (en) Dance teaching device
JP6839730B2 (en) Information processing system, information processing method and computer program
Taheri et al. Exploratory design of a hands-free video game controller for a quadriplegic individual
CN114712862A (en) Virtual pet interaction method, electronic device and computer-readable storage medium
EP3964271A1 (en) User input method and apparatus
JP2003088681A (en) Game information, information storage medium, and game apparatus
JP2020130592A (en) Program, method, information processing device, and system
Barbulescu et al. Making movies from make-believe games
TWI824883B (en) A virtual reality interactive system that uses virtual reality to simulate expressions and emotions for training
Larsson Discerning emotion through movement: A study of body language in portraying emotion in animation
Junius Puppitor: Building an Acting Interface for Videogames
US20240042323A1 (en) Information processing system, method for processing information, and non-transitory computer-readable information storage medium

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant