CN106503127B - Music data processing method and system based on facial action identification - Google Patents

Music data processing method and system based on facial action identification Download PDF

Info

Publication number
CN106503127B
CN106503127B CN201610912440.3A CN201610912440A CN106503127B CN 106503127 B CN106503127 B CN 106503127B CN 201610912440 A CN201610912440 A CN 201610912440A CN 106503127 B CN106503127 B CN 106503127B
Authority
CN
China
Prior art keywords
data
music
foreground
trifle
facial action
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610912440.3A
Other languages
Chinese (zh)
Other versions
CN106503127A (en
Inventor
简仁贤
何芳琳
赵伟翔
于庭婕
黄品瑞
廖健宏
陈智凯
孙廷伟
杨闵淳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intelligent Technology (shanghai) Co Ltd
Original Assignee
Intelligent Technology (shanghai) Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intelligent Technology (shanghai) Co Ltd filed Critical Intelligent Technology (shanghai) Co Ltd
Priority to CN201610912440.3A priority Critical patent/CN106503127B/en
Publication of CN106503127A publication Critical patent/CN106503127A/en
Application granted granted Critical
Publication of CN106503127B publication Critical patent/CN106503127B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The present invention provides a kind of music data processing method and system based on facial action identification, methods are as follows: obtains background music data and foreground music data, foreground music data is divided into multistage trifle by beat, every trifle includes multiple bats;The face action that people is detected within the foreground music time obtains multiple facial action data, the foreground music data of the corresponding trifle time span of each facial action;Background music data is carried out to continue broadcasting, and the foreground music data of the corresponding trifle time span of each facial action data are matched, and in conjunction with background music, generate new music.The present invention is based on the music data processing methods and system of facial action identification, using based on face's key point identification technology, the mode that music data and real image are combined, realize the interaction of people and corresponding scene, it does not need to be assisted by outside setting, implementation is simple, improves user experience.

Description

Music data processing method and system based on facial action identification
Technical field
The present invention relates to data processing fields, more particularly to the music data processing based on facial action identification.
Background technique
It in the prior art, is mostly to utilize mouse, keyboard, joystick, Touch Screen, external sensor (such as Wii, dancing machine Pedal etc.), human posture (such as Kinect) mode operate, realize people and corresponding scene interaction.Wherein, human posture operates Mode needs the information such as deep space, it is necessary to use specific device;Therefore it needs to detect whole body, needs biggish space ability Game.In addition the problems such as there are also accuracy rate is low, offset is high, causes the interest in interactive process to reduce;In addition such shape Scene interactivity major part is less in conjunction with real image for human action detecting (example: raising one's hand, skirt).
Face mainly in conjunction with image composing technique, is become other animals, worn by the application of face's key point identification at present Upper difference ornaments simultaneously combine facial action to generate animation etc..
The application when facial movements such as thin face and face's rehabilitation is provided: traditional facial movement through text process description or Achieve the effect that facial movement through the face movement demonstration in study film, but such facial movement process there is no with it is true Image combines, and Experience Degree is low.
Therefore, defect in the prior art is the interaction for people and corresponding scene, needs to assist by external equipment real Existing, implementation is single, and user experience is low.Face's key point identification technology and real image can not be combined.
Summary of the invention
In view of the above technical problems, the present invention provide it is a kind of based on facial action identification music data processing method and be System uses and is based on face's key point identification technology, the mode that music data and real image are combined, and realizes people and corresponding The interaction of scene does not need to be assisted by outside setting, and implementation is simple, improves user experience.
In order to solve the above technical problems, present invention provide the technical scheme that
In a first aspect, the present invention provides a kind of music data processing method based on facial action identification, comprising:
Step S1 obtains background music data and foreground music data, the background music data and the foreground music Data are respectively music of the number of segment second to several minutes of length;
The foreground music data are divided into multistage trifle by beat by step S2, and every trifle includes multiple bats;
Step S3 detects the face action of people within the foreground music time, obtains more within the foreground music time A facial action data, the foreground music data of the corresponding trifle time span of each facial action;
Step S4 carries out the background music data to continue broadcasting, and each facial action data are right with it The foreground music data for the trifle time span answered are matched, and in conjunction with the background music, generate new music.
The technical scheme is that first obtaining background music data and foreground music data, the background music data It is respectively music of the number of segment second to several minutes of length with the foreground music data;Then by the foreground music data by section Bat is divided into multistage trifle, and every trifle includes multiple bats;
Then the face action that people is detected within the foreground music time, obtains multiple faces within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;Finally by the background music number According to carrying out continuing broadcasting, and by the foreground music of the corresponding trifle time span of each facial action data Data are matched, and in conjunction with the background music, generate new music.
The present invention is based on the music data processing methods of facial action identification, use and identify skill based on face's key point Art, the mode that music data and real image are combined realize the interaction of people and corresponding scene, do not need to be arranged by outside It is assisted, implementation is simple, improves user experience.
Further, after the step S2, further includes:
The background music data is carried out to continue broadcasting, obtains target face action data, target face action data Corresponding unique trifle foreground music;
According to the target face action data, acquisition starts the face in the bit time of front and back in every trifle and moves Make data;
The facial action data and the target face action data are subjected to matching judgement, carry out a trifle foreground voice The happy selection played:
When the facial action data are matched with the target face action data, the trifle foreground music is played, The one trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data and target face action data mismatch, the facial action data are not played Corresponding trifle foreground music.
Further, after the step S2, further includes:
By the background music data carry out continue broadcasting, obtain virtual scene data, the virtual scene data be to The virtual scene data of each position mobile object in human face;
According to the virtual scene data of the position mobile object each into human face, corresponding facial action number is obtained According to the acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data and the virtual scene data are subjected to matching judgement, carry out the virtual scene data The processing of middle corresponding mobile object:
When the facial action data and the virtual scene Data Matching, moved corresponding in the virtual scene data Object removes;
When the facial action data and the virtual scene data mismatch, to shifting corresponding in the virtual scene data Animal body is not dealt with;
After the rear bit time that every trifle starts, without corresponding facial action and target face action data Match, mobile object removal will be corresponded in the virtual scene data.
Further, will be corresponded in the virtual scene data mobile object remove after, comprising:
Obtain the effect data that mobile object removal is corresponded in the virtual scene data;
According to corresponded in the virtual scene data mobile object removal effect data, to the virtual scene data The matched facial action data are evaluated, and evaluation result is obtained.
Further, the identification of facial action is carried out by the identification of face key point and fuzzy control theory.
A kind of music data processing method based on facial action identification of the present invention, is based on face's key point identification technology, The mode that music data and real image are combined realizes the interaction of people and corresponding scene, i.e., the knowledge being failure to actuate by face Not, it is matched with the music data of corresponding scene, realizes the creation of music, correspond to virtual objects in the broadcasting and scene of music Elimination, be presented in user in animated way at the moment, do not need by outside setting assisted, implementation is simple, mentions High user experience.
Further, further includes:
The background music data is carried out to continue broadcasting, obtains target face action data, the target face movement Data correspond to a trifle foreground music, and the foreground music is divided into the first foreground music and the second foreground music, before described first Scape music matches broadcasting with the background music, and second foreground music and the background music, which mismatch, to be played;
According to the target face action data, acquisition starts the face in the bit time of front and back in every trifle and moves Make data;
The facial action data and the target face action data are subjected to matching judgement, carry out the target face The broadcasting of the corresponding trifle foreground music of action data is chosen:
When the facial action data are matched with the target face action data, the target face action data is played A corresponding trifle foreground music, and the trifle foreground music is the first prospect corresponding with the target face action data Music;
When the facial action data and target face action data mismatch, the target face movement number is played According to a corresponding trifle foreground music, and the trifle foreground music be with before the target face action data corresponding second Scape music.
After facial action data match target face action data successful match with subject performance data, corresponding play should Target face acts corresponding trifle foreground music, also, the first foreground music and background music match harmonious broadcasting, , whereas if facial action data and target face action data that user makes mismatch, then the corresponding target face that plays moves Make corresponding second foreground music, the second foreground music and background music do not match that harmonious broadcasting, pass through broadcasting in this way The difference of music can judge facial action data that user makes whether with target face action data successful match, improve User experience.
Second aspect, the present invention provides a kind of music data processing systems based on facial action identification, comprising:
Music data obtains module, for obtaining background music data and foreground music data, the background music data It is respectively music of the number of segment second to several minutes of length with the foreground music data;
Music data processing module, for the foreground music data to be divided into multistage trifle by beat, every trifle includes Multiple bats;
Facial action obtains module, for detecting the face action of people within the foreground music time, obtains in the prospect Multiple facial action data in musical time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module continues broadcasting for carrying out the background music data, and by each facial action The foreground music data of the corresponding trifle time span of data are matched, raw in conjunction with the background music The music of Cheng Xin.
The technical solution of the present invention is as follows: first passing through music data obtains module, background music data and foreground music are obtained Data, the background music data and the foreground music data are respectively music of the number of segment second to several minutes of length;Then By music data processing module, the foreground music data are divided into multistage trifle by beat, every trifle includes multiple bats;
Then module is obtained by facial action, the face action of people is detected within the foreground music time, obtained described Multiple facial action data in the foreground music time, the foreground music number of the corresponding trifle time span of each facial action According to;Finally by musical composition module, the background music data continue broadcasting, and by each facial action number It is matched according to the foreground music data of the corresponding trifle time span, in conjunction with the background music, is generated New music.
A kind of music data processing system based on facial action identification of the present invention, is used and is identified based on face's key point Technology, the mode that music data and real image are combined realize the interaction of people and corresponding scene, do not need to set by outside It sets and is assisted, implementation is simple, improves user experience.
Further, after the music data processing module, further include the selection of music playing module, be used for:
The background music data is carried out to continue broadcasting, obtains target face action data, target face action data Corresponding unique trifle foreground music;
According to the target face action data, acquisition starts the face in the bit time of front and back in every trifle and moves Make data;
The facial action data and the target face action data are subjected to matching judgement, carry out a trifle foreground voice The happy selection played:
When the facial action data are matched with the target face action data, the trifle foreground music is played, The one trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data and target face action data mismatch, the facial action data are not played Corresponding trifle foreground music.
Further, after the music data processing module, further include music virtual scene module, be used for:
By the background music data carry out continue broadcasting, obtain virtual scene data, the virtual scene data be to The virtual scene data of each position mobile object in human face;
According to the virtual scene data of the position mobile object each into human face, corresponding facial action number is obtained According to the acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data and the virtual scene data are subjected to matching judgement, carry out the virtual scene data The processing of middle corresponding mobile object:
When the facial action data and the virtual scene Data Matching, moved corresponding in the virtual scene data Object removes;
When the facial action data and the virtual scene data mismatch, to shifting corresponding in the virtual scene data Animal body is not dealt with;
After the rear bit time that every trifle starts, without corresponding facial action and target face action data Match, mobile object removal will be corresponded in the virtual scene data.
Further, in the music virtual scene module, including effect assessment submodule, by the virtual scene number After removing according to middle corresponding mobile object, the effect assessment submodule is used for:
Obtain the effect data that mobile object removal is corresponded in the virtual scene data;
According to corresponded in the virtual scene data mobile object removal effect data, to the virtual scene data The matched facial action data are evaluated, and evaluation result is obtained.
Detailed description of the invention
It, below will be to specific in order to illustrate more clearly of the specific embodiment of the invention or technical solution in the prior art Embodiment or attached drawing needed to be used in the description of the prior art are briefly described.
Fig. 1 shows a kind of music data processing side based on facial action identification provided by first embodiment of the invention The flow chart of method;
Fig. 2 shows a kind of music data processing sides based on facial action identification provided by first embodiment of the invention First schematic diagram of time shaft in method;
Fig. 3 shows a kind of music data processing side based on facial action identification provided by first embodiment of the invention Second schematic diagram of time shaft in method;
Fig. 4 shows a kind of music data processing system based on facial action identification provided by second embodiment of the invention The schematic diagram of system.
Specific embodiment
It is described in detail below in conjunction with embodiment of the attached drawing to technical solution of the present invention.Following embodiment is only used for Clearly illustrate technical solution of the present invention, therefore be intended only as example, and cannot be used as a limitation and limit protection of the invention Range.
Embodiment one
Fig. 1 shows a kind of music data processing side based on facial action identification provided by first embodiment of the invention The flow chart of method;As shown in Figure 1, the embodiment of the present invention one provides a kind of music data processing side based on facial action identification Method, comprising:
Step S1 obtains background music data and foreground music data, background music data and foreground music data difference For a number of segment second to several minutes of long music;
Foreground music data are divided into multistage trifle by beat by step S2, and every trifle includes multiple bats;
Step S3 detects the face action of people within the foreground music time, obtains multiple faces within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;
Background music data continue broadcasting by step S4, and by corresponding one small of each facial action data The foreground music data of section time span are matched, and in conjunction with background music, generate new music.
The technical scheme is that first obtaining background music data and foreground music data, background music data is with before Scape music data is respectively music of the number of segment second to several minutes of length;Then it is small foreground music data to be divided into multistage by beat Section, every trifle include multiple bats;
Then the face action that people is detected within the foreground music time, the multiple faces obtained within the foreground music time are dynamic Make data, the foreground music data of the corresponding trifle time span of each facial action;Finally background music data is held Continued broadcasting is put, and the foreground music data of the corresponding trifle time span of each facial action data are matched, with Background music combines, and generates new music.
The present invention is based on the music data processing methods of facial action identification, and different scene informations can be set, different Scene information have different needs, need to complete according to different demand in scene information at the appointed time, if different Scene information matched from different facial actions, then may be implemented different movements, and then realize interacting for various people and scene.
The present invention is based on the music data processing methods of facial action identification, use and identify skill based on face's key point Art, the mode that music data and real image are combined realize the interaction of people and corresponding scene, do not need to be arranged by outside It is assisted, implementation is simple, improves user experience.
Specifically, after step S2, further includes:
Background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding Unique trifle foreground music;
According to target face action data, obtains and start the facial action data in the bit time of front and back in every trifle;
Facial action data and target face action data are subjected to matching judgement, carry out what a trifle foreground music played It chooses:
When facial action data are matched with target face action data, a trifle foreground music, a trifle foreground voice are played It is happy for the unique corresponding foreground music of target face action data;
When facial action data and target face action data mismatch, the corresponding trifle of facial action data is not played Foreground music.
Specifically, the foreground music of more trifles is set, and every trifle foreground music corresponds to only one facial action, corresponding After first facial action and first aim facial action successful match, next facial action and next target are carried out The matching of facial action can act according to different target faces in this way, realize the continuous broadcasting of foreground music.
Specifically, after step S2, further includes:
Background music data is carried out to continue broadcasting, obtains virtual scene data, virtual scene data are into human face The virtual scene data of each position mobile object;
According to the virtual scene data of position each into human face mobile object, corresponding facial action data are obtained, The acquisition of corresponding facial action data is before the position that mobile object reaches in human face;
Facial action data and virtual scene data are subjected to matching judgement, carry out corresponding to motive objects in virtual scene data The processing of body:
When facial action data and virtual scene Data Matching, mobile object removal will be corresponded in virtual scene data;
When facial action data and virtual scene data mismatch, do not make to locate to mobile object is corresponded in virtual scene data Reason;
After the rear bit time that every trifle starts, matched without corresponding facial action with target face action data, Mobile object removal will be corresponded in virtual scene data.
In conjunction with AR technology, face facial positions are corresponded in real scene, and to be arranged different virtual objects mobile to face, Bit time is event horizon before and after virtual objects are moved to the time point in face facial positions, and correspondence is made each virtual Article corresponding elimination movement is then done next movement, is increased after the facial action and elimination that people is done act successful match Interest, while successful match or unsuccessfully having different an audio and animation.After a period of time, match time shortens, Movement speed is eliminated in matching to be accelerated.
Specifically, will be corresponded in virtual scene data mobile object remove after, comprising:
Obtain the effect data that mobile object removal is corresponded in virtual scene data;
According to the effect data for corresponding to mobile object removal in virtual scene data, to the face with virtual scene Data Matching Portion's action data is evaluated, and evaluation result is obtained.
As shown in Fig. 2, horizontal axis is time shaft, time, the right represent later time earlier for left side representative.More slightly grow The longitudinal axis is trifle separation, and the separation of point is clapped in relatively thin short representative.In this legend, driving range is before every trifle starts Two bats are all correct driving, and 0.5 bat drives before and after starting previous bat with every trifle to be perfect, in Fig. 2, the part that e is shown, and e Front and back 0.5 clap as common driving, the part shown such as f.Perfect, common driving time and range all can be replaced freely.With The method does the evaluation criterion to virtual scene Data Matching facial action data.
Specifically, the identification of facial action is carried out by the identification of face key point and fuzzy control theory.
In the present invention, handled based on above-mentioned music data, be based on the identification of face key point, establish one it is reliable it Face system, finally combine fuzzy control theory accurately identify facial action: blink, cross-eye, choose eyebrow, frown, the nose that wrinkles, It lolls, mouth of beeping, open one's mouth, wapperijaw, lick lip, close lightly lip, nod, head rotates left and right, head is rotated up and down etc..Wherein face is crucial Point identification technology is the well known prior art, does not do excessive narration herein.
Specifically, further includes:
Background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding One trifle foreground music, foreground music are divided into the first foreground music and the second foreground music, the first foreground music and background music Matching plays, and the second foreground music and background music are mismatched and played;
According to target face action data, obtains and start the facial action data in the bit time of front and back in every trifle;
Facial action data and target face action data are subjected to matching judgement, it is corresponding to carry out target face action data The broadcasting of one trifle foreground music is chosen:
When facial action data are matched with target face action data, before playing the corresponding trifle of target face action data Scape music, and a trifle foreground music is the first foreground music corresponding with target face action data;
When facial action data and target face action data mismatch, broadcasting target face action data corresponds to a trifle Foreground music, and a trifle foreground music is the second foreground music corresponding with target face action data.
After facial action data match target face action data successful match with subject performance data, corresponding play should Target face acts corresponding trifle foreground music, also, the first foreground music and background music match harmonious broadcasting, , whereas if facial action data and target face action data that user makes mismatch, then the corresponding target face that plays moves Make corresponding second foreground music, the second foreground music and background music do not match that harmonious broadcasting, pass through broadcasting in this way The difference of music can judge facial action data that user makes whether with target face action data successful match, improve User experience.
Specifically, the first foreground music may be configured as indicating that successful music, the second foreground music can be set to indicate to lose The foreground music lost makes music more having any different property in this way.
As shown in figure 3, horizontal axis is time shaft, time, the right represent later time earlier for left side representative.More slightly grow The longitudinal axis is trifle separation, and the separation of point is clapped in relatively thin short representative.With the example in Fig. 3, the range of c is a trifle, the model of d It encloses and is clapped for one.A is detecting time point, represents in Fig. 3 and starts front and back one in every trifle and clap as detecting time point, it is any at this moment Between put the facial action done and can be detected.B is sphere of action, usually as unit of a trifle, is represented in this figure small herein The front and back one of section claps in range and does expression, all this trifle can be driven to make feedback.Every trifle is several to clap, detects time point and effect model Enclosing all can freely replace.
A kind of music data processing method based on facial action identification of the present invention, is based on face's key point identification technology, The mode that music data and real image are combined realizes the interaction of people and corresponding scene, i.e., the knowledge being failure to actuate by face Not, it is matched with the music data of corresponding scene, realizes the creation of music, correspond to virtual objects in the broadcasting and scene of music Elimination, be presented in user in animated way at the moment, do not need by outside setting assisted, implementation is simple, mentions High user experience.
Embodiment two
Fig. 4 shows a kind of music data processing system based on facial action identification provided by second embodiment of the invention The schematic diagram of system;As shown in figure 4, second embodiment of the present invention provides a kind of music data processing systems based on facial action identification System 10, comprising:
Music data obtains module 101, for obtaining background music data and foreground music data, background music data and Foreground music data are respectively music of the number of segment second to several minutes of length;
Music data processing module 102, for foreground music data to be divided into multistage trifle by beat, every trifle includes more A bat;
Facial action obtains module 103, for detecting the face action of people within the foreground music time, obtains in foreground voice Multiple facial action data in the happy time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module 104, for by background music data carry out continue broadcasting, and by each facial action data with The foreground music data of its corresponding trifle time span are matched, and in conjunction with background music, generate new music.
The technical solution of the present invention is as follows: first passing through music data obtains module 101, background music data and foreground voice are obtained Happy data, background music data and foreground music data are respectively music of the number of segment second to several minutes of length;Then pass through sound Foreground music data are divided into multistage trifle by beat by happy data processing module 102, and every trifle includes multiple bats;
Then module 103 is obtained by facial action, the face action of people is detected within the foreground music time, obtained preceding Multiple facial action data in scape musical time, the foreground music data of the corresponding trifle time span of each facial action; Finally by musical composition module 104, background music data is carried out to continue broadcasting, and each facial action data are right with it The foreground music data for the trifle time span answered are matched, and in conjunction with background music, generate new music.
A kind of music data processing system 10 based on facial action identification of the present invention, is used and is known based on face's key point Other technology, the mode that music data and real image are combined are realized the interaction of people and corresponding scene, are not needed by outside Setting is assisted, and implementation is simple, improves user experience.
Specifically, after music data processing module 102, further include the selection of music playing module, be used for:
Background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding Unique trifle foreground music;
According to target face action data, obtains and start the facial action data in the bit time of front and back in every trifle;
Facial action data and target face action data are subjected to matching judgement, carry out what a trifle foreground music played It chooses:
When facial action data are matched with target face action data, a trifle foreground music, a trifle foreground voice are played It is happy for the unique corresponding foreground music of target face action data;
When facial action data and target face action data mismatch, the corresponding trifle of facial action data is not played Foreground music.
Specifically, after music data processing module 102, further include music virtual scene module, be used for:
Background music data is carried out to continue broadcasting, obtains virtual scene data, virtual scene data are into human face The virtual scene data of each position mobile object;
According to the virtual scene data of position each into human face mobile object, corresponding facial action data are obtained, The acquisition of corresponding facial action data is before the position that mobile object reaches in human face;
Facial action data and virtual scene data are subjected to matching judgement, carry out corresponding to motive objects in virtual scene data The processing of body:
When facial action data and virtual scene Data Matching, mobile object removal will be corresponded in virtual scene data;
When facial action data and virtual scene data mismatch, do not make to locate to mobile object is corresponded in virtual scene data Reason;
After the rear bit time that every trifle starts, matched without corresponding facial action with target face action data, Mobile object removal will be corresponded in virtual scene data.
Specifically, in music virtual scene module 104, including effect assessment submodule, will be right in virtual scene data After answering mobile object to remove, effect assessment submodule is used for:
Obtain the effect data that mobile object removal is corresponded in virtual scene data;
According to the effect data for corresponding to mobile object removal in virtual scene data, to the face with virtual scene Data Matching Portion's action data is evaluated, and evaluation result is obtained.
It specifically, further include facial action identification module 100, for being identified by face key point and fuzzy control theory Carry out the identification of facial action.
Specifically, further include the selection of music playing module, be used for:
Background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding One trifle foreground music, foreground music are divided into the first foreground music and the second foreground music, the first foreground music and background music Matching plays, and the second foreground music and background music are mismatched and played;
According to target face action data, obtains and start the facial action data in the bit time of front and back in every trifle;
Facial action data and target face action data are subjected to matching judgement, it is corresponding to carry out target face action data The broadcasting of one trifle foreground music is chosen:
When facial action data are matched with target face action data, before playing the corresponding trifle of target face action data Scape music, and a trifle foreground music is the first foreground music corresponding with target face action data;
When facial action data and target face action data mismatch, broadcasting target face action data corresponds to a trifle Foreground music, and a trifle foreground music is the second foreground music corresponding with target face action data.
After facial action data match target face action data successful match with subject performance data, corresponding play should Target face acts corresponding trifle foreground music, also, the first foreground music and background music match harmonious broadcasting, , whereas if facial action data and target face action data that user makes mismatch, then the corresponding target face that plays moves Make corresponding second foreground music, the second foreground music and background music do not match that harmonious broadcasting, pass through broadcasting in this way The difference of music can judge facial action data that user makes whether with target face action data successful match, improve User experience.
Specifically, the first foreground music may be configured as indicating that successful music, the second foreground music can be set to indicate to lose The foreground music lost makes music more having any different property in this way.
A kind of music data processing system based on facial action identification of the present invention, is based on face's key point identification technology, The mode that music data and real image are combined realizes the interaction of people and corresponding scene, i.e., the knowledge acted by face Not, it is matched with the music data of corresponding scene, realizes the creation of music, correspond to virtual objects in the broadcasting and scene of music Elimination, be presented in user in animated way at the moment, do not need by outside setting assisted, implementation is simple, mentions High user experience.
Embodiment three
The music data processing method and embodiment identified in conjunction with one of embodiment of the present invention one based on facial action One of two music data processing systems based on facial action identification, are illustrated in conjunction with specific scene of game.
Scene one
Melody creation: before background music persistently plays, and each facial action corresponds to one section of one trifle time span Scape music starts the facial action detected in the bit time of front and back in every trifle, can all drive the trifle to play corresponding Foreground music.For example a series of facial actions are set, it blinks, chooses eyebrow, frowns, the nose that wrinkles etc., then identifying the facial action of people, root Corresponding to play different music according to the difference of the facial action of identification, the trifle music played every time forms different music, User is set to do different music according to the mood of oneself wound.
Scene two
Music game: background music and foreground music are the music of a number of segment minute, and background music persistently plays, game according to So as unit of trifle, specified facial action is randomly generated by game per small festival-gathering, player must in driving time (every trifle Start front and back bit time) specified facial action is completed, the foreground music of the trifle can continue to play;If there is more than one prospect Music, also may done correct expression it is more multiple after by music superposition up.By motivational music in conjunction with human face action, make to swim It plays more challenging and interest.
Scene three
AR Rhythmic game: background music persistently plays, and different articles are had on picture rhythmic past portion each on the face Displacement is dynamic, and player must do corresponding expression and remove the article when article is moved to the position.Such as: mosquito flies toward eyes Past, it is necessary to which blink kills it when mosquito reaches eyes.Success (being divided into perfect, common) unsuccessfully has different sounds Effect, animation and score.After a period of time, rhythm tempo can be accelerated, and increase game degree of difficulty.
Directly make face's action recognition using the image of camera acquirement to operate game, and can in real time with true face's shadow Interesting animation effect is generated as combining.This mode of operation is not required to other purchase of equipment, and it is dynamic can in real time, accurately to differentiate face Make;And the mode of operation of four limbs is not needed, handicapped people can be also benefited, allows them that can also enjoy the enjoyment of game.
Finally, it should be noted that the above embodiments are only used to illustrate the technical solution of the present invention., rather than its limitations;To the greatest extent Pipe present invention has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that: its according to So be possible to modify the technical solutions described in the foregoing embodiments, or to some or all of the technical features into Row equivalent replacement;And these are modified or replaceed, various embodiments of the present invention technology that it does not separate the essence of the corresponding technical solution The range of scheme should all cover within the scope of the claims and the description of the invention.

Claims (8)

1. the music data processing method based on facial action identification characterized by comprising
Step S1 obtains background music data and foreground music data, the background music data and the foreground music data The music that a respectively number of segment second is grown to several minutes;
The foreground music data are divided into multistage trifle by beat by step S2, and every trifle includes multiple bats;
Step S3 detects the face action of people within the foreground music time, obtains multiple faces within the foreground music time Portion's action data, the foreground music data of the corresponding trifle time span of each facial action;
Step S4 carries out the background music data to continue broadcasting, and each facial action data are corresponding The foreground music data of the one trifle time span are matched, and in conjunction with the background music, generate new music;
After the step S2, further includes:
The background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding Unique trifle foreground music;
According to the target face action data, obtains and start the facial action number in the bit time of front and back in every trifle According to;
The facial action data and the target face action data are subjected to matching judgement, a trifle foreground music is carried out and broadcasts The selection put:
When the facial action data are matched with the target face action data, the broadcasting one trifle foreground music is described One trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data and target face action data mismatch, it is corresponding that the facial action data are not played A trifle foreground music.
2. the music data processing method according to claim 1 based on facial action identification, which is characterized in that
After the step S2, further includes:
The background music data is carried out to continue broadcasting, obtains virtual scene data, the virtual scene data are to people face The virtual scene data of each position mobile object in portion;
According to the virtual scene data of the position mobile object each into human face, corresponding facial action data are obtained, The acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data and the virtual scene data are subjected to matching judgement, it is right in the virtual scene data to carry out Answer the processing of mobile object:
When the facial action data and the virtual scene Data Matching, mobile object will be corresponded in the virtual scene data It removes;
When the facial action data and the virtual scene data mismatch, to corresponding to motive objects in the virtual scene data Body is not dealt with;
After the rear bit time that every trifle starts, matched without corresponding facial action with target face action data, Mobile object removal will be corresponded in the virtual scene data.
3. the music data processing method according to claim 2 based on facial action identification, which is characterized in that
To be corresponded in the virtual scene data mobile object remove after, comprising:
Obtain the effect data that mobile object removal is corresponded in the virtual scene data;
According to corresponded in the virtual scene data mobile object removal effect data, to the virtual scene Data Matching The facial action data evaluated, obtain evaluation result.
4. the music data processing method according to claim 1 based on facial action identification, which is characterized in that
The identification of facial action is carried out by the identification of face key point and fuzzy control theory.
5. the music data processing method according to claim 1 based on facial action identification, which is characterized in that
Further include:
The background music data is carried out to continue broadcasting, obtains target face action data, the target face action data A corresponding trifle foreground music, the foreground music are divided into the first foreground music and the second foreground music, first foreground voice Happy that broadcasting is matched with the background music, second foreground music and the background music, which mismatch, to be played;
According to the target face action data, obtains and start the facial action number in the bit time of front and back in every trifle According to;
The facial action data and the target face action data are subjected to matching judgement, carry out the target face movement The broadcasting of the corresponding trifle foreground music of data is chosen:
When the facial action data are matched with the target face action data, it is corresponding to play the target face action data One trifle foreground music, and the trifle foreground music is the first foreground voice corresponding with the target face action data It is happy;
When the facial action data and target face action data mismatch, the target face action data pair is played A trifle foreground music is answered, and the trifle foreground music is the second foreground voice corresponding with the target face action data It is happy.
6. the music data processing system based on facial action identification characterized by comprising
Music data obtains module, for obtaining background music data and foreground music data, the background music data and institute Stating foreground music data is respectively music of the number of segment second to several minutes of length;
Music data processing module, for the foreground music data to be divided into multistage trifle by beat, every trifle includes multiple Bat;
Facial action obtains module, for detecting the face action of people within the foreground music time, obtains in the foreground music Multiple facial action data in time, the foreground music data of the corresponding trifle time span of each facial action;
Musical composition module continues broadcasting for carrying out the background music data, and by each facial action data The foreground music data of the corresponding trifle time span are matched, and in conjunction with the background music, are generated new Music;
After the music data processing module, further includes the selection of music playing module, is used for:
The background music data is carried out to continue broadcasting, obtains target face action data, target face action data is corresponding Unique trifle foreground music;
According to the target face action data, obtains and start the facial action number in the bit time of front and back in every trifle According to;
The facial action data and the target face action data are subjected to matching judgement, a trifle foreground music is carried out and broadcasts The selection put:
When the facial action data are matched with the target face action data, the broadcasting one trifle foreground music is described One trifle foreground music is and the unique corresponding foreground music of the target face action data;
When the facial action data and target face action data mismatch, it is corresponding that the facial action data are not played A trifle foreground music.
7. the music data processing system according to claim 6 based on facial action identification, which is characterized in that
After the music data processing module, further includes music virtual scene module, is used for:
The background music data is carried out to continue broadcasting, obtains virtual scene data, the virtual scene data are to people face The virtual scene data of each position mobile object in portion;
According to the virtual scene data of the position mobile object each into human face, corresponding facial action data are obtained, The acquisition of the corresponding facial action data is before the position that the mobile object reaches in the human face;
The facial action data and the virtual scene data are subjected to matching judgement, it is right in the virtual scene data to carry out Answer the processing of mobile object:
When the facial action data and the virtual scene Data Matching, mobile object will be corresponded in the virtual scene data It removes;
When the facial action data and the virtual scene data mismatch, to corresponding to motive objects in the virtual scene data Body is not dealt with;
After the rear bit time that every trifle starts, matched without corresponding facial action with target face action data, Mobile object removal will be corresponded in the virtual scene data.
8. the music data processing system according to claim 7 based on facial action identification, which is characterized in that
In the music virtual scene module, including effect assessment submodule, it is moved by corresponding in the virtual scene data After object removes, the effect assessment submodule is used for:
Obtain the effect data that mobile object removal is corresponded in the virtual scene data;
According to corresponded in the virtual scene data mobile object removal effect data, to the virtual scene Data Matching The facial action data evaluated, obtain evaluation result.
CN201610912440.3A 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification Active CN106503127B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610912440.3A CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610912440.3A CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Publications (2)

Publication Number Publication Date
CN106503127A CN106503127A (en) 2017-03-15
CN106503127B true CN106503127B (en) 2019-09-27

Family

ID=58294244

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610912440.3A Active CN106503127B (en) 2016-10-19 2016-10-19 Music data processing method and system based on facial action identification

Country Status (1)

Country Link
CN (1) CN106503127B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108905193B (en) * 2018-07-03 2022-04-15 百度在线网络技术(北京)有限公司 Game manipulation processing method, device and storage medium
CN109343770B (en) * 2018-09-27 2021-07-20 腾讯科技(深圳)有限公司 Interactive feedback method, apparatus and recording medium
CN110047520B (en) * 2019-03-19 2021-09-17 北京字节跳动网络技术有限公司 Audio playing control method and device, electronic equipment and computer readable storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764940A (en) * 2003-03-31 2006-04-26 索尼株式会社 Tempo analysis device and tempo analysis method
CN101836219A (en) * 2007-11-01 2010-09-15 索尼爱立信移动通讯有限公司 Generating music playlist based on facial expression
CN102640149A (en) * 2009-12-04 2012-08-15 索尼计算机娱乐公司 Music recommendation system, information processing device, and information processing method
CN102880388A (en) * 2012-09-06 2013-01-16 北京天宇朗通通信设备股份有限公司 Music processing method, music processing device and mobile terminal
CN102929476A (en) * 2012-09-06 2013-02-13 北京天宇朗通通信设备股份有限公司 Method and device for controlling main menu of terminal
CN103383694A (en) * 2012-12-14 2013-11-06 李博文 Method and system for organizing, managing and marking music document
CN104851435A (en) * 2015-06-06 2015-08-19 孔霞 Music intelligent playing method based on network communication
CN105518783A (en) * 2013-08-19 2016-04-20 谷歌公司 Content-based video segmentation

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1764940A (en) * 2003-03-31 2006-04-26 索尼株式会社 Tempo analysis device and tempo analysis method
CN101836219A (en) * 2007-11-01 2010-09-15 索尼爱立信移动通讯有限公司 Generating music playlist based on facial expression
CN102640149A (en) * 2009-12-04 2012-08-15 索尼计算机娱乐公司 Music recommendation system, information processing device, and information processing method
CN102880388A (en) * 2012-09-06 2013-01-16 北京天宇朗通通信设备股份有限公司 Music processing method, music processing device and mobile terminal
CN102929476A (en) * 2012-09-06 2013-02-13 北京天宇朗通通信设备股份有限公司 Method and device for controlling main menu of terminal
CN103383694A (en) * 2012-12-14 2013-11-06 李博文 Method and system for organizing, managing and marking music document
CN105518783A (en) * 2013-08-19 2016-04-20 谷歌公司 Content-based video segmentation
CN104851435A (en) * 2015-06-06 2015-08-19 孔霞 Music intelligent playing method based on network communication

Also Published As

Publication number Publication date
CN106503127A (en) 2017-03-15

Similar Documents

Publication Publication Date Title
JP7041763B2 (en) Technology for controlling a virtual image generation system using the user's emotional state
CN102822869B (en) Capture view and the motion of the performer performed in the scene for generating
Collins Playing with sound: a theory of interacting with sound and music in video games
CN102473320B (en) Bringing a visual representation to life via learned input from the user
Miller Playable bodies: dance games and intimate media
JP2019532374A5 (en)
TWI377055B (en) Interactive rehabilitation method and system for upper and lower extremities
CN106503127B (en) Music data processing method and system based on facial action identification
CN102129343A (en) Directed performance in motion capture system
JP2018011201A (en) Information processing apparatus, information processing method, and program
WO2009021124A2 (en) System and method for a motion sensing amusement device
Augello et al. Creative robot dance with variational encoder
Bishko Animation principles and Laban movement analysis: movement frameworks for creating empathic character performances
Guzman-Sanchez Underground Dance Masters: final history of a forgotten era
JP5399966B2 (en) GAME DEVICE, GAME DEVICE CONTROL METHOD, AND PROGRAM
JP2014023745A (en) Dance teaching device
Dower et al. Performing for motion capture: A guide for practitioners
CN114253393A (en) Information processing apparatus, terminal, method, and computer-readable recording medium
Miura et al. GoalBaural: A training application for goalball-related aural sense
Hugill et al. Audio only computer games–Papa Sangre
Tinwell et al. Survival horror games-an uncanny modality
Flash et al. The adventures of Grandmaster Flash: My life, my beats
TW201112186A (en) Wherein each of basic poses corresponds to a musical note and is stored in a music mapping device for creating various music performances according to the motion of the kid, thereby helping the kid to create his/her own favorite music
Winchell Ventriloquism for fun and profit
Larsson Discerning emotion through movement: A study of body language in portraying emotion in animation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant