CN106096720A - A kind of method that dance movement is automatically synthesized - Google Patents

A kind of method that dance movement is automatically synthesized Download PDF

Info

Publication number
CN106096720A
CN106096720A CN201610410516.2A CN201610410516A CN106096720A CN 106096720 A CN106096720 A CN 106096720A CN 201610410516 A CN201610410516 A CN 201610410516A CN 106096720 A CN106096720 A CN 106096720A
Authority
CN
China
Prior art keywords
action
fragment
music
dance movement
action fragment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610410516.2A
Other languages
Chinese (zh)
Inventor
王晓磊
邓际锋
雷涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Rulei Technology Co Ltd
Original Assignee
Hangzhou Rulei Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Rulei Technology Co Ltd filed Critical Hangzhou Rulei Technology Co Ltd
Priority to CN201610410516.2A priority Critical patent/CN106096720A/en
Publication of CN106096720A publication Critical patent/CN106096720A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/008Artificial life, i.e. computing arrangements simulating life based on physical entities controlled by simulated intelligence so as to replicate intelligent life forms, e.g. based on robots replicating pets or humans in their appearance or behaviour

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Robotics (AREA)
  • Toys (AREA)

Abstract

The invention discloses a kind of method that dance movement is automatically synthesized, comprise the steps: S1: identify the speed in music and tempo characteristic;S2: choose the action fragment meeting musical features from maneuver library, through conversion process, connects and forms continuous print dance movement.The present invention, by identifying the feature of music, chooses action fragment according to musical features, action fragment is made conversion process, and the dance movement synthesized can meet music-tempo, style;Dance movement links up, does not repeats, and has good sight.Achieve intelligent toy simply and be automatically synthesized the demand of dance movement.

Description

A kind of method that dance movement is automatically synthesized
Technical field
The present invention relates to a kind of method that dance movement is automatically synthesized, belong to intelligent toy technical field.
Background technology
In computer graphics and the field such as robot, toy, there is the demand being automatically synthesized dance movement.Common application Scene is such as: virtual figure image is danced with music beat on computers;Robot or toy accompanied by music beat are danced.
Common prior art such as the personage's dance movement in computer game.Common practice is for whole song, Write dance movement.When playing melody, play the action preset.This method does not have motility, it is impossible to be automatically synthesized dance Step action, can only pre-set action.The dance movement that the most this method makes can only be arranged in pairs or groups specific melody, applied field Scape is single.
The additionally prior art as disclosed by Chinese patent 200910101046, the method using machine learning, selection Join the dance movement of musical features, to be automatically synthesized dance movement.The advantage of this invention is, selected dance movement can be preferably Ground meets music style and speed.Its shortcoming is, the technical scheme used is complicated, and operand is big, is not suitable for small machines The equipment such as people, toy, is not suitable for the use scene high to requirement of real-time yet.
Summary of the invention
The present invention is directed to deficiencies of the prior art, it is provided that a kind of method that dance movement is automatically synthesized, simply Realize intelligent toy and be automatically synthesized the demand of dance movement, and dance movement links up, does not repeats, sight is good.
To this end, the present invention adopts the following technical scheme that
A kind of method that dance movement is automatically synthesized, comprises the steps:
S1: identify the speed in music and tempo characteristic;
S2: choose the action fragment meeting musical features from maneuver library, through conversion process, connects formation continuous print dancing and moves Make;
Described action fragment is a bit of action performed, and defines the body posture variation pattern of performing artist, wherein, action The starting and ending attitude of fragment is a few reference attitude, and action fragment comprises Speed attribute, describes this section of action symbol Closing the music of what speed, definition mode is with the velocity characteristic of music;Action fragment also comprises beat attribute, describes this section and moves Making to be suitable for the music of what beat type, definition mode is with the tempo characteristic of music;Whether action fragment also comprises can mirror image Attribute;Described reference attitude refers to, certain several body posture is defined as benchmark, the starting and ending appearance of everything fragment State is all the one in these several reference attitudes.
Further, in step s 2, the step choosing action fragment from maneuver library is:
S2-1: according to the velocity characteristic identified in music, filter out the action fragment in the range of Speed attribute value is positioned at certain;
S2-2: according to the tempo characteristic identified in music, then filter out some beat attributes action movie equal to music beat Section;
S2-3: randomly select one in the action fragment filtered out, and the beginning attitude of this action fragment is equal to performing artist The end attitude of previous action.
Further, in step s 2, the conversion process of described action fragment is: make new advances choosing after step S2-3 After action fragment, according to the velocity characteristic T2 identified in its Speed attribute value T1, with music, this action fragment is done speed change Processing, speed change ratio is T1/T2;
Further, after selecting new element fragment after step S2-3, if this action fragment can mirror image, then Mirror image processing is done in this action, as next action fragment, after being spliced to the action fragment of transformed process.
Further, the variable-speed processing of described action fragment refers to, the overall time of action fragment is accelerated in proportion or Slack-off.
Further, the mirror image processing of described action fragment refers to, by action fragment along about actor body axis Mirror image, becomes a new action fragment.
Further, in the method for attachment of said two action fragment, after the end attitude of previous action is equal to, one moves The beginning attitude made.
The present invention is by identifying the feature of music, and chooses action fragment according to musical features, convert action fragment Processing, the dance movement synthesized can meet music-tempo, style;And dance movement links up, do not repeats, have and well view and admire Property.By defining several reference actions, make the splicing between action fragment smooth, do not have factitious attitude to suddenly change, the most raw Hard action switching;By choosing action fragment from maneuver library, and splice through speed change, mirror image etc., make the dancing of synthesis move Make abundanter, be not simple repetition.Compared to prior art, the present invention also has the advantage that operand is little, it is adaptable to minicomputer Application on the low-speed devices such as device people, toy.Achieve the demand being automatically synthesized dance movement simply.
Accompanying drawing explanation
Fig. 1 is the workflow diagram of the present invention;
Fig. 2 is the schematic description of the feature in maneuver library of the present invention to action fragment;
Fig. 3 is the schematic description of the feature of reference actions in the present invention.
Detailed description of the invention
In order to make those skilled in the art be better understood from the present invention program, below in conjunction with in the embodiment of the present invention Accompanying drawing, carries out clear, complete description to the technical scheme in the embodiment of the present invention.
As it is shown in figure 1, the method that the dance movement of the present invention is automatically synthesized, comprise the steps:
S1: identify the speed in music and tempo characteristic: wherein, the metric form of speed is " clap/minute ", and beat refers to 2/4 The beat types such as bat, 3/4 bat;Identification to musical features can use the prior art of association area, is not repeated herein;By knowing Other musical features, corresponds to the dance movement being automatically synthesized on the speed of music, beat, it is ensured that dance movement is in accordance with music;
S2: choose the action fragment meeting musical features from maneuver library, through conversion process, connects formation continuous print dancing and moves Make;
Maneuver library contains some default action fragments, and wherein, action fragment is a bit of action performed, and defines table The body posture variation pattern of person's of drilling (virtual image on robot, toy or computer), wherein, action fragment initial and Terminating attitude is a few reference attitude, and action fragment comprises Speed attribute, describes what speed this section of action meets Music, definition mode is with the velocity characteristic of music;Action fragment also comprises beat attribute, describes what this section of action be suitable for The music of beat type, definition mode is with the tempo characteristic of music;Whether action fragment also comprises can mirror image attribute;Described benchmark Attitude refers to, certain several body posture is defined as benchmark, and the starting and ending attitude of everything fragment is all that these are several One in reference attitude.
The action fragment describing mode to actor body attitude, it is possible to use the existing method in this area, such as, at 3D In modeling software, parts each to actor body define a series of key frames, and each key frame defines duration, joint rotates Direction etc., by these modes, can be with the attitudes vibration of health in a series of definition a period of times.
In step s 2, the concrete steps choosing action fragment from maneuver library can be:
S2-1: according to the velocity characteristic identified in music, filter out the action fragment in the range of Speed attribute value is positioned at certain;
S2-2: according to the tempo characteristic identified in music, then filter out some beat attributes action movie equal to music beat Section;
S2-3: randomly select one in the action fragment filtered out, and the beginning attitude of this action fragment is equal to performing artist The end attitude of previous action.
In step s 2, the conversion process of described action fragment is: selecting new element fragment after step S2-3 Afterwards, according to the velocity characteristic T2 identified in its Speed attribute value T1, with music, this action fragment is done variable-speed processing, become Speed ratio example is T1/T2, is formed;
After selecting new element fragment after step S2-3, if this action fragment can mirror image, then by this action Do mirror image processing, as next action fragment, after being spliced to the action fragment of transformed process.
The variable-speed processing of described action fragment refers to, is accelerated in proportion or slack-off the overall time of action fragment.
The mirror image processing of described action fragment refers to, by action fragment along left and right, actor body axis mirror image, becomes One new action fragment.
In the method for attachment of said two action fragment, the end attitude of previous action is equal to the beginning appearance of latter action State.Two such action energy smooth registration,
As Figure 1-3, below as a example by a robot automatically can danced with music, detailed dance movement is automatically synthesized Method:
In the present embodiment, robot uses single-chip microcomputer (MCU) to control, and the health of robot has three mobilizable joints, The range of movement in each joint is by a definition, and the scope of numeral is 0 to 100;The current action of robot is to compile in Fig. 2 Number it is the action fragment of 1, needs the next action of synthesis.
" music " in Fig. 1 sound outside the mike of robot interior picks up, form is PCM;" musical features is known Not " step use prior art, identify music-tempo be 126 bats/minute, beat is 2/4 bat.
The step of " action is chosen " is:
First according to the speed 126 of music clap/minute, filter out from the maneuver library shown in Fig. 2 speed 90-150 clap/minute In the range of action fragment, the action fragment of numbered 1 and 2 meets this condition;
Subsequently, it is 2/4 bat according to the beat of music, filters out the action equal to 2/4 bat of the action beat, the numbered action of 1,2 Fragment meets this condition;
Then, according to the end attitude " attitude B " of robot current action, filter out the action that reference attitude is attitude B, numbering Be 2 action fragment meet this condition;
Due to the selection result only one of which action, so choosing action fragment 2 as next action.
The step of " action conversion process " is:
First according to the speed 126 of music clap/minute, and the speed of action fragment 2 be 120 bats/minute, by action fragment 2 Do quickening to process, make the time span of action fragment 2 become 120/126 ≈ 0.95 times;Action after process is " action fragment 2a”;
Then because action fragment 2 can mirror image, the attitude description of action fragment 2a is done according to robot health axis Left and right mirror image processing, becomes " action fragment 2b ";
The step of " action splicing " is: action fragment 2a, action fragment 2b is spliced successively after current action, becomes coherent Dance movement.
In the whole step of embodiment, robot by accepting external voice, use the present invention method, be automatically synthesized Going out to meet the dance movement of music, then corresponding motion is made in the joint of driven machine people.Thus allow robot with music jump Dance.
Obviously, described embodiment is only a part of embodiment of the present invention rather than whole embodiments.Based on Embodiment in the present invention, it is every other that those of ordinary skill in the art are obtained under not making creative work premise Embodiment, all should belong to the scope of protection of the invention.

Claims (7)

1. the method that dance movement is automatically synthesized, comprises the steps:
S1: identify the speed in music and tempo characteristic;
S2: choose the action fragment meeting musical features from maneuver library, through conversion process, connects formation continuous print dancing and moves Make;
Described action fragment is a bit of action performed, and defines the body posture variation pattern of performing artist, wherein, action The starting and ending attitude of fragment is a few reference attitude, and action fragment comprises Speed attribute, describes this section of action symbol Closing the music of what speed, definition mode is with the velocity characteristic of music;Action fragment also comprises beat attribute, describes this section and moves Making to be suitable for the music of what beat type, definition mode is with the tempo characteristic of music;Whether action fragment also comprises can mirror image Attribute;Described reference attitude refers to, certain several body posture is defined as benchmark, the starting and ending appearance of everything fragment State is all the one in these several reference attitudes.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: in step s 2, from action The step choosing action fragment in storehouse is:
S2-1: according to the velocity characteristic identified in music, filter out the action fragment in the range of Speed attribute value is positioned at certain;
S2-2: according to the tempo characteristic identified in music, then filter out some beat attributes action movie equal to music beat Section;
S2-3: randomly select one in the action fragment filtered out, and the beginning attitude of this action fragment is equal to performing artist The end attitude of previous action.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: in step s 2, described dynamic The conversion process making fragment is: after selecting new element fragment after step S2-3, according to its Speed attribute value T1, with The velocity characteristic T2 identified in music, does variable-speed processing to this action fragment, and speed change ratio is T1/T2.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: selecting after step S2-3 Take out after new element fragment, if this action fragment can mirror image, then mirror image processing is done in this action, moves as the next one Make fragment, after being spliced to the action fragment of transformed process.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: the speed change of described action fragment Process refers to, is accelerated in proportion or slack-off the overall time of action fragment.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: the mirror image of described action fragment Process refers to, by action fragment along left and right, actor body axis mirror image, becomes a new action fragment.
The method that dance movement the most according to claim 1 is automatically synthesized, it is characterised in that: said two action fragment In method of attachment, the end attitude of previous action is equal to the beginning attitude of latter action.
CN201610410516.2A 2016-06-12 2016-06-12 A kind of method that dance movement is automatically synthesized Pending CN106096720A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610410516.2A CN106096720A (en) 2016-06-12 2016-06-12 A kind of method that dance movement is automatically synthesized

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610410516.2A CN106096720A (en) 2016-06-12 2016-06-12 A kind of method that dance movement is automatically synthesized

Publications (1)

Publication Number Publication Date
CN106096720A true CN106096720A (en) 2016-11-09

Family

ID=57229172

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610410516.2A Pending CN106096720A (en) 2016-06-12 2016-06-12 A kind of method that dance movement is automatically synthesized

Country Status (1)

Country Link
CN (1) CN106096720A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107594739A (en) * 2017-09-03 2018-01-19 泉州迪特工业产品设计有限公司 A kind of intelligence dazzles cruel footwear and its implementation
CN107682654A (en) * 2017-09-30 2018-02-09 北京金山安全软件有限公司 Video recording method, shooting device, electronic equipment and medium
CN108052250A (en) * 2017-12-12 2018-05-18 北京光年无限科技有限公司 Virtual idol deductive data processing method and system based on multi-modal interaction
CN108527376A (en) * 2018-02-27 2018-09-14 深圳狗尾草智能科技有限公司 Control method, apparatus, equipment and the medium of robot dance movement
CN110992449A (en) * 2019-11-29 2020-04-10 网易(杭州)网络有限公司 Dance action synthesis method, device, equipment and storage medium
CN112365568A (en) * 2020-11-06 2021-02-12 广州小鹏汽车科技有限公司 Audio processing method and device, electronic equipment and storage medium
WO2021134942A1 (en) * 2019-12-31 2021-07-08 网易(杭州)网络有限公司 Dance animation processing method and apparatus, electronic device, and storage medium
CN113590872A (en) * 2021-07-28 2021-11-02 广州艾美网络科技有限公司 Method, device and equipment for generating dance spectral plane
CN113643442A (en) * 2021-08-06 2021-11-12 福建凯米网络科技有限公司 Virtual object dance generation method and computer storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101524594A (en) * 2009-04-17 2009-09-09 中国地质大学(武汉) Anthropomorphic robot autonomously dancing along with rhythm
CN101615302A (en) * 2009-07-30 2009-12-30 浙江大学 The dance movement generation method that music data drives based on machine learning
CN101693371A (en) * 2009-09-30 2010-04-14 深圳先进技术研究院 Robot capable of dancing by following music beats

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101524594A (en) * 2009-04-17 2009-09-09 中国地质大学(武汉) Anthropomorphic robot autonomously dancing along with rhythm
CN101615302A (en) * 2009-07-30 2009-12-30 浙江大学 The dance movement generation method that music data drives based on machine learning
CN101693371A (en) * 2009-09-30 2010-04-14 深圳先进技术研究院 Robot capable of dancing by following music beats

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107594739A (en) * 2017-09-03 2018-01-19 泉州迪特工业产品设计有限公司 A kind of intelligence dazzles cruel footwear and its implementation
CN107682654A (en) * 2017-09-30 2018-02-09 北京金山安全软件有限公司 Video recording method, shooting device, electronic equipment and medium
CN107682654B (en) * 2017-09-30 2019-11-26 北京金山安全软件有限公司 Video recording method, shooting device, electronic equipment and medium
CN108052250A (en) * 2017-12-12 2018-05-18 北京光年无限科技有限公司 Virtual idol deductive data processing method and system based on multi-modal interaction
CN108527376A (en) * 2018-02-27 2018-09-14 深圳狗尾草智能科技有限公司 Control method, apparatus, equipment and the medium of robot dance movement
CN110992449B (en) * 2019-11-29 2023-04-18 网易(杭州)网络有限公司 Dance action synthesis method, device, equipment and storage medium
CN110992449A (en) * 2019-11-29 2020-04-10 网易(杭州)网络有限公司 Dance action synthesis method, device, equipment and storage medium
WO2021134942A1 (en) * 2019-12-31 2021-07-08 网易(杭州)网络有限公司 Dance animation processing method and apparatus, electronic device, and storage medium
CN112365568A (en) * 2020-11-06 2021-02-12 广州小鹏汽车科技有限公司 Audio processing method and device, electronic equipment and storage medium
CN113590872A (en) * 2021-07-28 2021-11-02 广州艾美网络科技有限公司 Method, device and equipment for generating dance spectral plane
CN113590872B (en) * 2021-07-28 2023-11-28 广州艾美网络科技有限公司 Method, device and equipment for generating dancing spectrum surface
CN113643442A (en) * 2021-08-06 2021-11-12 福建凯米网络科技有限公司 Virtual object dance generation method and computer storage medium
CN113643442B (en) * 2021-08-06 2024-05-14 福建凯米网络科技有限公司 Virtual object dance generation method and computer storage medium

Similar Documents

Publication Publication Date Title
CN106096720A (en) A kind of method that dance movement is automatically synthesized
CN108202334B (en) Dance robot capable of identifying music beats and styles
Jordà Sonigraphical Instruments: From FMOL to the reacTable*.
Hoffman et al. Shimon: an interactive improvisational robotic marimba player
Schöllig et al. A platform for dance performances with multiple quadrocopters
Hamilton et al. Gesture-based collaborative virtual reality performance in carillon
Caspersen It starts from any point: bill and the Frankfurt ballet
CN105080774A (en) Expandable fountain controlling system and method for synchronizing music and water type
CN105243682B (en) A kind of method of limb member model, role and two-dimension cartoon making
Landry et al. Participatory design research methodologies: A case study in dancer sonification
Willett et al. Reasons for rhythm: Multimodal perspectives on musical play
CN108829253A (en) A kind of analog music commander's playback method and device
CN102799749B (en) Automatic generating method and generating system for distributed music lamplight performance scheme
Copeland Cunningham, Collage, and the computer
Oh et al. Crafting mechatronic percussion with everyday materials
Kumra et al. Dual flexible 7 dof arm robot learns like a child to dance using q-learning
JP6752465B1 (en) Computer programs and game systems
Assayag Improvising in creative symbolic interaction
Demian Disdance-new artistic environment
Hamilton Perceptually coherent mapping schemata for virtual space and musical method
Saltz Plays are games, movies are pictures: Ludic vs. pictorial representation
Jung CHOREOGRAPHIC SOUND COMPOSITION: Towards a poetics of restriction
Morris Snake in the Labyrinth: Scenes from the Machine’s Deep Q-Learning Experience
Birringer et al. Sound and Wearables
Montgomery Becoming burlesque: Performer training in contemporary burlesque

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20161109