WO2019184634A1 - 人机交互系统、方法、计算机可读存储介质及交互装置 - Google Patents

人机交互系统、方法、计算机可读存储介质及交互装置 Download PDF

Info

Publication number
WO2019184634A1
WO2019184634A1 PCT/CN2019/075928 CN2019075928W WO2019184634A1 WO 2019184634 A1 WO2019184634 A1 WO 2019184634A1 CN 2019075928 W CN2019075928 W CN 2019075928W WO 2019184634 A1 WO2019184634 A1 WO 2019184634A1
Authority
WO
WIPO (PCT)
Prior art keywords
human
image
gesture
display unit
instruction
Prior art date
Application number
PCT/CN2019/075928
Other languages
English (en)
French (fr)
Inventor
唐堂
张勃
蒋志鹏
文林福
Original Assignee
北京字节跳动网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字节跳动网络技术有限公司 filed Critical 北京字节跳动网络技术有限公司
Publication of WO2019184634A1 publication Critical patent/WO2019184634A1/zh
Priority to US17/020,561 priority Critical patent/US20200409471A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Definitions

  • the present disclosure relates to the field of artificial intelligence, and in particular to a human-computer interaction system, a human-computer interaction method, a computer-readable storage medium, and a human-machine interaction device.
  • motion capture technology has become a key technology in the study of human motion posture, playing an increasingly important role. People realize that it is very necessary to realize the interaction between human motion and information equipment by recognizing human motion posture. .
  • motion capture technology has been widely used in large-scale entertainment equipment, animation production, gait analysis, biomechanics, ergonomics, etc., and with the popularization of mobile devices such as mobile phones and tablet computers, mobile devices such as mobile phones and tablet computers. It is a must-have item for people to be entertained with simplicity, convenience, and no restrictions on time and place. Therefore, applying motion capture technology to mobile phones along with mobile devices such as mobile phones and tablet computers, so that people can experience good entertainment effects. The problem to be treated.
  • An embodiment of the first aspect of the present disclosure provides a human-computer interaction method, including:
  • displaying one or more gesture images in the gesture template group on the display unit, before collecting the motion image of the person further comprising: extracting a gesture template group corresponding to the instruction according to the instruction.
  • one or more gesture images in the gesture template group are displayed on the display unit, and the motion image of the collection person is further included before:
  • the audio is played before the action image of the person is matched with the currently displayed gesture image and the matching result is displayed on the display unit.
  • matching the motion image of the person with the currently displayed gesture image, and displaying the matching result on the display unit includes:
  • Corresponding animations and/or scores are displayed on the display unit based on the matching results.
  • extracting the gesture template group corresponding to the instruction according to the instruction includes:
  • Extracting one or more of the pre-stored plurality of gesture images constitutes a gesture template group corresponding to the instruction.
  • the human-computer interaction method further includes:
  • the total score is matched to a preset score rating and the level at which the score is located is displayed on the display unit.
  • the method before displaying the gesture image, the method further includes:
  • the distance between the human and the aircraft is detected, and when the distance between the human and the machine is within the preset range, preparation is started to display the posture image on the display unit.
  • An embodiment of the second aspect of the present disclosure provides a human-computer interaction system, including: an interaction module, configured to display one or more gesture images in the gesture template group on a display unit, and collect an action image of a person; and And a comparison module, configured to match the motion image of the person with the currently displayed gesture image, and display a matching result on the display unit.
  • the human-computer interaction system further includes: an extraction module, configured to extract a gesture template group corresponding to the instruction according to the instruction.
  • the human-computer interaction system further includes: an extraction module, configured to extract audio corresponding to the instruction according to the instruction; and the interaction module controls the audio playback.
  • the comparison module includes: a processing unit, configured to extract a plurality of single-frame images from the motion image of the person; a matching unit, configured to match the single-frame image with the gesture image, and Generating a matching result; and executing unit for displaying a corresponding animation and/or score on the display unit according to the matching result.
  • the gesture template group is composed of one or more gesture images selected from a plurality of pre-stored gesture images.
  • the human-computer interaction system further includes: a summary module, configured to: when the matching result includes a score, all the scores displayed after the gesture image display ends are summarized to obtain a total score; and a rating module, configured to: The total score is matched to a preset score rating and the level at which the score is located is displayed on the display unit.
  • a summary module configured to: when the matching result includes a score, all the scores displayed after the gesture image display ends are summarized to obtain a total score
  • a rating module configured to: The total score is matched to a preset score rating and the level at which the score is located is displayed on the display unit.
  • the human-computer interaction system further includes: an identification module, configured to detect a distance between the human-machines, and when the distance between the human-machines is within a preset range, prepare to start displaying the gesture on the display unit. image.
  • an identification module configured to detect a distance between the human-machines, and when the distance between the human-machines is within a preset range, prepare to start displaying the gesture on the display unit. image.
  • An embodiment of the third aspect of the present disclosure provides a computer readable storage medium having stored thereon a computer program, the program being executed by a processor to implement the steps of the human-computer interaction method of any of the above.
  • An embodiment of the fourth aspect of the present disclosure provides a human-machine interaction apparatus including a memory, a processor, and a program stored on the memory and executable on the processor, the processor executing the program to implement any of the above The steps of the human-computer interaction method described in the item.
  • a gesture image is displayed on the display unit (for example, multiple stickmans, animations, animal images, and the like in different postures), and the user performs the same limb motions as the gesture images, so that the user forms a dancing motion, and simultaneously , collecting the image of the user, matching the motion image of the person with the posture image, and displaying the matching result (such as a score and/or an animation effect) on the display unit according to the degree of matching between the motion of the person and the posture image,
  • the user has a guiding role, enabling the user to perform standard dance moves, improving the entertainment effect, thereby improving the user experience.
  • FIG. 1 is a schematic structural diagram of hardware of a terminal device according to an embodiment of the present disclosure
  • FIG. 2 is a structural block diagram of a first embodiment of a human-machine interaction system according to the present disclosure
  • FIG. 3 is a structural block diagram of a second embodiment of a human-machine interaction system according to the present disclosure.
  • FIG. 4 is a structural block diagram of a third embodiment of a human-machine interaction system according to the present disclosure.
  • FIG. 5 is a structural block diagram of a fourth embodiment of a human-machine interaction system according to the present disclosure.
  • FIG. 6 is a structural block diagram of a fifth embodiment of the human-machine interaction system according to the present disclosure.
  • FIG. 7 is a block diagram showing a flow structure of a sixth embodiment of the human-machine interaction system according to the present disclosure.
  • FIG. 8 is a block diagram showing a flow structure of a seventh embodiment of the human-computer interaction system according to the present disclosure.
  • FIG. 9 is a schematic flowchart of an interaction method of a first embodiment of a human-computer interaction method according to the present disclosure.
  • FIG. 10 is a schematic flowchart diagram of an interaction method of a second embodiment of the human-computer interaction method according to the present disclosure.
  • FIG. 11 is a schematic flowchart diagram of an interaction method of a third embodiment of the human-computer interaction method according to the present disclosure.
  • FIG. 12 is a schematic flowchart diagram of an interaction method of a fourth embodiment of the human-computer interaction method according to the present disclosure.
  • FIG. 13 is a schematic flowchart diagram of an interaction method of a fifth embodiment of the human-computer interaction method according to the present disclosure.
  • FIG. 14 is a schematic flowchart diagram of an interaction method of a sixth embodiment of the human-computer interaction method according to the present disclosure.
  • FIG. 15 is a schematic flowchart diagram of an interaction method of a seventh embodiment of the human-computer interaction method according to the present disclosure.
  • 16 is a schematic diagram of a computer readable storage medium of an embodiment of the present disclosure.
  • FIG. 17 is a schematic structural diagram of a human-machine interaction apparatus according to an embodiment of the present disclosure.
  • 100 human-computer interaction system 101 extraction module, 1011 image unit, 1012 audio unit, 102 interaction module, 103 comparison module, 1031 processing unit, 1032 matching unit, 1033 execution unit, 104 summary module, 105 rating module, 106 identification module, 1 wireless communication unit, 2 input unit, 3 user input unit, 4 sensing unit, 5 output unit, 6 memory, 7 interface unit, 8 controller, 9 power unit, 80 human machine interaction device, 801 memory, 802 processor 900 computer readable storage medium, 901 non-transitory computer readable instructions.
  • the human-machine interaction device that is, the terminal device
  • the terminal device in the present disclosure may include, but is not limited to, such as a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a PDA (personal Mobile assistants such as digital assistants, PADs (tablets), PMPs (portable multimedia players), navigation devices, in-vehicle terminal devices, in-vehicle display terminals, in-vehicle electronic rearview mirrors, and the like, and such as digital TVs, desktop computers, and the like Fixed terminal equipment.
  • PDA personal Mobile assistants such as digital assistants, PADs (tablets), PMPs (portable multimedia players)
  • navigation devices in-vehicle terminal devices, in-vehicle display terminals, in-vehicle electronic rearview mirrors, and the like, and such as digital TVs, desktop computers, and the like Fixed terminal equipment.
  • the terminal device may include a wireless communication unit 1, an A/V (audio/video) input unit 2, a user input unit 3, a sensing unit 4, an output unit 5, a memory 6, and an interface unit 7. , controller 8 and power supply unit 9 and the like.
  • the A/V (audio/video) input unit 2 includes, but is not limited to, a camera, a front camera, a rear camera, and various audio and video input devices. It should be understood by those skilled in the art that the components included in the terminal devices listed in the above embodiments may include fewer or more components than the types described above.
  • the various embodiments described herein can be implemented in a computer readable medium using, for example, computer software, hardware, or any combination thereof.
  • the embodiments described herein may be through the use of application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays ( An FPGA, a processor, a controller, a microcontroller, a microprocessor, at least one of the electronic units designed to perform the functions described herein, in some cases, such an embodiment may be in the controller Implementation.
  • implementations such as procedures or functions may be implemented with separate software modules that permit the execution of at least one function or operation.
  • the software code can be implemented by a software application (or program) written in any suitable programming language, which can be stored in memory and executed by the controller.
  • the human-machine interaction system 100 provided by the embodiment of the first aspect of the present disclosure includes: an interaction module 102 and a comparison module 103.
  • the interaction module 102 is configured to display one or more gesture images in the gesture template group on the display unit, and collect a motion image of the person;
  • the comparison module 103 is configured to use the motion image of the human and the currently displayed gesture image. Match and display the matching result on the display unit.
  • the display unit (the display unit can be a display screen or the like) displays a gesture image (eg, multiple sticks, animations, animal images, etc. in different postures), and the gesture image displays a human hand, The position, angle, etc. of the boom, the arm, the thigh, the calf, the torso and the head at different times, the user performs the same limb movements as the posture images, so that the user forms a dancing motion, and at the same time, the interaction module collects the image of the user.
  • a gesture image eg, multiple sticks, animations, animal images, etc. in different postures
  • the comparison module matches the human motion image with the gesture image, and displays matching results (such as scores and/or animation effects) on the display unit according to the degree of matching between the human motion and the posture image, and guides the user who does not dance.
  • the function enables the user to make standard dance movements and improve the entertainment effect, thereby improving the user experience.
  • the human-machine interaction system 100 provided by the embodiment of the first aspect of the present disclosure includes: an extraction module 101, an interaction module 102, and a comparison module 103.
  • the extraction module 101 is configured to extract a gesture template group corresponding to the instruction according to the instruction;
  • the interaction module 102 is configured to display one or more gesture images in the gesture template group on the display unit, and collect an action image of the human;
  • the comparison module 103 is configured to match the motion image of the person with the currently displayed posture image, and display the matching result on the display unit.
  • the display unit (the display unit can be a display screen or the like) displays a gesture image (eg, multiple sticks, animations, animal images, etc. in different postures), and the gesture image displays a human hand, The position, angle, etc. of the boom, the arm, the thigh, the calf, the torso and the head at different times, the user performs the same limb movements as the posture images, so that the user forms a dancing motion, and at the same time, the interaction module collects the image of the user.
  • a gesture image eg, multiple sticks, animations, animal images, etc. in different postures
  • the comparison module matches the human motion image with the gesture image, and displays matching results (such as scores and/or animation effects) on the display unit according to the degree of matching between the human motion and the posture image, and guides the user who does not dance.
  • the function enables the user to make standard dance movements and improve the entertainment effect, thereby improving the user experience.
  • the extraction module 101 is configured to extract a gesture template group and audio corresponding to the instruction according to the instruction;
  • the interaction module 102 is configured to control audio playback, and display multiple gestures in the gesture template group on the display unit.
  • the image captures a motion image of the person;
  • the comparison module 103 is configured to match the motion image of the person with the currently displayed gesture image, and display the matching result on the display unit.
  • the human-computer interaction system 100 displays a posture image (such as a plurality of stickman, an animation, an animal image, etc. in different postures) along with a music playback display unit (the display unit may be a display screen or the like), in the posture image
  • a posture image such as a plurality of stickman, an animation, an animal image, etc. in different postures
  • the display unit may be a display screen or the like
  • the display shows the position, angle, etc. of the human hand, the big arm, the lower arm, the thigh, the calf, the torso and the head at different times.
  • the interaction module collects the image of the user, and the comparison module matches the motion image of the person with the gesture image, and displays the matching result (such as a score and/or an animation effect) on the display unit according to the degree of matching between the motion of the person and the posture image.
  • the matching result such as a score and/or an animation effect
  • the comparison module 103 includes a processing unit 1031, a matching unit 1032, and an execution unit 1033.
  • the processing unit 1031 is configured to extract a plurality of single-frame images from a motion image of a person; the matching unit 1032 is configured to match the single-frame image with the gesture image, and generate a matching result; the execution unit 1033 is configured to match The result shows the corresponding animation and/or score on the display unit.
  • the processing unit 1031 extracts a plurality of single-frame images from the motion image of the collector, such as taking one hundred frames per unit time, and the matching unit 1032 determines one by matching the one-frame image with the gesture image.
  • the coincidence rate of the hundred frames and the attitude image, the detection method can achieve accurate detection, improve the detection precision of the product, thereby improving the experience of the product; displaying the score and/or animation on the display unit, the animation can be perfect, good , great, miss and other digital animation, or display the next heart rain, the stars and rain and other special effects from the display unit.
  • the extraction module 101 includes an image unit 1011 and an audio unit 1012.
  • the image unit 1011 is configured to extract one or more of the plurality of gesture images to form a gesture template group corresponding to the instruction;
  • the audio unit 1012 is configured to call up the audio corresponding to the instruction.
  • a plurality of posture templates are pre-stored in the image unit 1011, and the image unit 1011 extracts a plurality of posture templates according to different instructions selected by the user and sorts the posture template groups; in one embodiment of the present disclosure, one hundred is included.
  • a gesture template which extracts the first, third, fifth, 20th, 66th, 78th, 82th, and 96th posture templates according to the first instruction to form a posture template group; and extracts the second, the second, the second, the second, the second, the second, the second, the The posture templates of 55, 69, 73, 86, and 96 constitute a posture template group; and the gesture templates of the seventh, 13, 29, 35, 38, 46, 52, 68, 71, 86, and 91 posture templates are formed according to the third instruction,
  • the audio unit 1012 extracts corresponding music according to different instructions selected by the user.
  • a plurality of gesture templates are pre-stored in the extraction module 101, and the gesture template group is composed of one or more gesture images selected from a plurality of pre-stored gesture images, and the extraction module 101 selects different according to the user.
  • the instruction extracts a plurality of pose templates and sorts the pose template groups; in one embodiment of the present disclosure, includes one hundred pose templates, and extracts the first, third, fifth, twenty, sixteenth, eighty, eight, five, sixteenth, sixteenth
  • the posture template constitutes a posture template group; the second, the 12th, 22nd, 25th, 37th, 47th, 55th, 69th, 86th, 86th, and 96th posture templates are extracted according to the second instruction to form the posture template group; and the seventh instruction is extracted according to the third instruction.
  • the pose templates of 29, 35, 38, 46, 52, 68, 71, 86, and 91 constitute a pose template group.
  • the human-machine interaction system 100 further includes: a summary module 104 and a rating module 105.
  • the summary module 104 is configured to summarize all the scores displayed after the end of the audio or after the end of the gesture image display to obtain a total score when the matching result includes the score; the rating module 105 is configured to rate the total score with the preset score Matches and displays the level at which the score is on the display unit.
  • the user can know the scores and grades of the dance through the summary module 104 and the rating module 105.
  • the user can rank with other users' scores and grades according to the scores and ranks, thereby increasing the product interaction.
  • users can share videos with scores and grades to friends, so that friends can have a direct judgment on their own dance.
  • the human-machine interaction system 100 further includes: an identification module 106 for detecting a distance between the human and the machine, when the distance between the human and the machine When within the preset range, it is ready to start playing audio and/or ready to start displaying the gesture image on the display unit.
  • the setting of the identification module 106 can ensure that the user is completely in the display unit, and can make the dance action performed by the user better match the posture image on the display unit, that is, the user's limb is avoided.
  • Exceeding the display unit causes the mismatch to be sent, thereby improving the comfort of the product, thereby increasing the market competitiveness of the product.
  • the distance between the user and the mobile phone is within a reasonable range, thereby enabling the user
  • the content displayed on the display unit can be clearly seen, thereby increasing the comfort of use of the product, thereby increasing the market competitiveness of the product.
  • the human-computer interaction method provided by the embodiment of the second aspect of the present disclosure includes the following steps:
  • Step 30 Display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • step 40 the motion image of the person is matched with the currently displayed gesture image, and the matching result is displayed on the display unit.
  • the display unit (the display unit can be a display screen, etc.) displays a gesture image (such as multiple stickman, animation, animal image, etc. in different postures), and the gesture image displays a human hand, The position, angle, etc.
  • a gesture image such as multiple stickman, animation, animal image, etc. in different postures
  • the user performs the same limb movements as the posture images, so that the user forms a dancing motion, and at the same time, the user's image is collected, and the person is
  • the motion image matches the gesture image, and according to the degree of matching between the motion of the person and the gesture image, the matching result (such as a score and/or an animation effect) is displayed on the display unit, and the user who is less likely to dance has a guiding effect, so that the user The ability to do standard dance moves enhances the entertainment effect and thus enhances the user experience.
  • the human-computer interaction method in this embodiment includes:
  • Step 10 extract a gesture template group and audio corresponding to the instruction according to the instruction
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • step 40 the motion image of the person is matched with the currently displayed gesture image, and the matching result is displayed on the display unit.
  • the method for human-computer interaction displays a posture image (such as a plurality of stickman, animation, animal image, etc. in different postures) along with the music display unit (the display unit can be a display screen, etc.), in the posture image
  • the display shows the position, angle, etc. of the human hand, the big arm, the lower arm, the thigh, the calf, the torso and the head at different times.
  • the user performs the same limb movement with the posture image with the music, so that the user forms a dancing action, and at the same time , collecting the image of the user, matching the motion image of the person with the posture image, and displaying the matching result (such as a score and/or an animation effect) on the display unit according to the degree of matching between the motion of the person and the posture image,
  • the user has a guiding role, enabling the user to perform standard dance moves, improving the entertainment effect, thereby improving the user experience.
  • step 40 includes:
  • Step 41 Extract a plurality of single-frame images from a motion image of a person
  • Step 42 Match the single frame image with the attitude image, and generate a matching result
  • Step 43 Display corresponding animations and/or scores on the display unit according to the matching result.
  • Step 10 extract a gesture template group and audio corresponding to the instruction according to the instruction
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • Step 41 Extract a plurality of single-frame images from a motion image of a person
  • Step 42 Match the single frame image with the attitude image, and generate a matching result
  • Step 43 Display corresponding animations and/or scores on the display unit according to the matching result.
  • a plurality of single-frame images are extracted from the motion image of the collector, such as taking one hundred frames per unit time, and the matching unit determines one hundred frames and gestures by matching one hundred frames of images with the attitude image.
  • the coincidence rate of the image, the detection method can achieve accurate detection, improve the detection precision of the product, thereby improving the experience of the product; displaying the score and/or animation on the display unit, the animation can be perfect, good, great, miss Wait for a digital animation, or display special effects such as rain, star rain, etc. from the display unit.
  • step 10 includes:
  • Step 11 Extract one or more of the plurality of posture images to form a posture template group corresponding to the instruction
  • Step 12 calling up the audio corresponding to the instruction
  • Step 11 Extract one or more of the pre-stored plurality of posture images to form a posture template group corresponding to the instruction
  • Step 12 calling up the audio corresponding to the instruction
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • step 40 the motion image of the person is matched with the currently displayed gesture image, and the matching result is displayed on the display unit.
  • a plurality of posture templates are pre-stored, and the posture template group is composed of one or more posture images selected from a plurality of pre-stored posture images, and a plurality of posture templates are extracted according to different instructions of the user and the postures are sorted.
  • a template group in one embodiment of the present disclosure, comprising one hundred gesture templates, and extracting the first, third, fifth, 20th, 66th, 78th, 82th, and 96th posture templates according to the first instruction to form a posture template group;
  • the instruction extracts the 2, 12, 22, 25, 37, 47, 55, 69, 73, 86, 96 pose templates to form the pose template group; and extracts the 7, 13, 29, 35, 38, 46, 52 according to the third instruction.
  • the 68, 71, 86, and 91 pose templates form a pose template group, and the corresponding music is extracted according to different instructions selected by the user.
  • the human-computer interaction method further includes:
  • Step 50 when the matching result includes a score, all the scores displayed after the end of the audio or the end of the gesture image display are summed to obtain a total score;
  • step 60 the total score is matched with the preset score rating, and the level at which the score is located is displayed on the display unit.
  • Step 10 extract a gesture template group and audio corresponding to the instruction according to the instruction
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • Step 40 Matching the motion image of the person with the currently displayed posture image, and displaying the matching result on the display unit;
  • Step 50 when the matching result includes a score, all the scores displayed after the end of the audio or the end of the gesture image display are summed to obtain a total score;
  • step 60 the total score is matched with the preset score rating, and the level at which the score is located is displayed on the display unit.
  • the user can know the scores and grades of the dance.
  • the user can rank according to the scores and grades of other users according to the scores and grades, thereby increasing the interactivity and interest of the products.
  • Users can share videos with scores and grades to friends, so that friends can have a direct judgment on how well they dance.
  • the method further includes:
  • Step 20 Detecting the distance between the human and the machine, when the distance between the human and the machine is within the preset range, preparing to start playing the audio and/or preparing to start displaying the gesture image on the display unit.
  • the human-computer interaction method in this embodiment includes:
  • Step 10 extract a gesture template group and audio corresponding to the instruction according to the instruction
  • Step 20 detecting a distance between the human and the machine, when the distance between the human and the machine is within a preset range, preparing to start playing audio and/or preparing to start displaying the gesture image on the display unit;
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • step 40 the motion image of the person is matched with the currently displayed gesture image, and the matching result is displayed on the display unit.
  • Step 10 extract a gesture template group and audio corresponding to the instruction according to the instruction
  • Step 20 detecting a distance between the human and the machine, when the distance between the human and the machine is within a preset range, preparing to start playing audio and/or preparing to start displaying the gesture image on the display unit;
  • Step 30 Play audio, display one or more gesture images in the gesture template group on the display unit, and collect an action image of the person;
  • Step 40 Matching the motion image of the person with the currently displayed posture image, and displaying the matching result on the display unit;
  • Step 50 When the matching result includes a score, all the scores displayed after the end of the audio or after the end of the gesture image display are summarized to obtain a total score;
  • step 60 the total score is matched with the preset score rating, and the level at which the score is located is displayed on the display unit.
  • the setting of the identification step can ensure that the user is completely in the display area of the display unit, and can better match the dance action performed by the user with the posture image on the display unit, that is, the user is avoided. If the limb exceeds the display unit, the mismatch occurs, thereby improving the comfort of the product, thereby increasing the market competitiveness of the product, and on the other hand, making the distance between the user and the mobile phone within a reasonable range, thereby The user can clearly see the content displayed on the display unit, thereby increasing the comfort of the product, thereby increasing the market competitiveness of the product.
  • an identification frame is displayed on the display unit, and when the image of the person is in the identification frame, the audio is ready to start playing or is ready to start displaying the gesture image on the display unit.
  • the identification frame is a humanoid frame.
  • the humanoid frame is a human body shape, and the user performs the same action as the humanoid frame shape according to the humanoid frame, and when the user is all located in the humanoid frame, is ready to start playing audio or is ready to start displaying the posture image on the display unit, and appears on the display unit.
  • Countdown, countdown ends playing audio.
  • the identification frame is a humanoid frame
  • the humanoid frame is a human upper body shape (may also be a lower body), and the user performs the same action as the humanoid frame shape according to the humanoid frame, and the user's upper body is located in the humanoid shape.
  • the frame is ready to start playing audio or ready to start displaying the posture image on the display unit, the countdown appears on the display unit, and the countdown ends the playback of the audio.
  • the role of the recognition frame is to identify the target to ensure smooth progress of the interaction, and therefore, an identification frame that satisfies the recognition function should be within the protection scope of the present disclosure.
  • an embodiment of a third aspect of the present disclosure provides a computer readable storage medium having stored thereon a computer program, the program being executed by a processor to implement the steps of any one of the above-described human-computer interaction methods.
  • the computer readable storage medium may include, but is not limited to, any type of disk, including a flash memory, a hard disk, a multimedia card, a card type memory (eg, SD or DX memory, etc.), a static random access memory (SRAM), and an electrically erasable.
  • EEPROM programmable read-only memory
  • PROM programmable read-only memory
  • magnetic memory floppy disk, optical disk, DVD, CD-ROM, micro-driver and magneto-optical disk
  • ROM programmable read-only memory
  • RAM programmable read-only memory
  • EPROM programmable read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • DRAM magnetic RAM
  • VRAM flash memory
  • computer readable storage medium 900 has stored thereon non-transitory computer readable instructions 901.
  • a human body interaction method based on a human body dynamic gesture according to an embodiment of the present disclosure described above is performed.
  • the human-machine interaction device provided by the embodiment of the fourth aspect of the present disclosure includes a memory, a processor, and a program stored on the memory and operable on the processor, and the processor implements any one of the above-mentioned human-computer interaction method methods when executing the program. A step of.
  • a memory is used to store non-transitory computer readable instructions.
  • the memory may include one or more computer program products, which may include various forms of computer readable storage media, such as volatile memory and/or nonvolatile memory.
  • Volatile memory can include, for example, random access memory (RAM) and/or caches and the like.
  • the non-volatile memory may include, for example, a read only memory (ROM), a hard disk, a flash memory, or the like.
  • the processor may be a central processing unit (CPU) or other form of processing unit having data processing capabilities and/or instruction execution capabilities, and may control other components in the human-machine interaction device to perform The desired function.
  • the processor is configured to execute computer readable instructions stored in the memory such that the human interaction device performs the interactive method described above.
  • the human-machine interaction device 80 includes a memory 801 and a processor 802.
  • the components in the human-machine interaction device 80 are interconnected by a bus system and/or other form of connection mechanism (not shown).
  • Memory 801 is for storing non-transitory computer readable instructions.
  • memory 801 may include one or more computer program products, which may include various forms of computer readable storage media, such as volatile memory and/or nonvolatile memory.
  • Volatile memory can include, for example, random access memory (RAM) and/or caches and the like.
  • the non-volatile memory may include, for example, a read only memory (ROM), a hard disk, a flash memory, or the like.
  • Processor 802 can be a central processing unit (CPU) or other form of processing unit with data processing capabilities and/or instruction execution capabilities, and can control other components in human-machine interaction device 80 to perform desired functions.
  • the processor 802 is configured to execute computer readable instructions stored in the memory 801 such that the human interaction device 80 performs the human body dynamic gesture based human interaction method described above.
  • the human-machine interaction device is the same as the embodiment described above based on the human body dynamic attitude-based human-computer interaction method, and a repetitive description thereof will be omitted herein.
  • the human-machine interaction device is a mobile device
  • the camera of the mobile device collects an image of the user
  • the song and gesture template group corresponding to the instruction is downloaded by the mobile device, and after the song and gesture template group are downloaded,
  • An identification box appears on the display unit of the mobile device (the identification frame can be a humanoid frame).
  • the identification frame can be a humanoid frame.
  • the image of the user is placed in the identification frame, the mobile device starts playing music, and multiple displays are displayed on the display unit.
  • Attitude images such as multiple stickman, animation, animal images, etc.
  • the user starts to make dance moves to match their body movements with these gesture images, according to the degree of matching between the user's movements and the posture image.
  • the animation can be digital animations such as perfect, good, great, miss, etc., or display special effects such as rain, star rain, etc. from the display unit).
  • the display unit of the mobile device is displayed.
  • Display scores and grades users can download their own dance videos or share them or enter the row List, mobile devices for mobile phones, tablets and so on.
  • the term “plurality” means two or more, unless specifically defined otherwise.
  • the terms “installation”, “connected”, “connected”, “fixed” and the like should be understood broadly.
  • “connecting” may be a fixed connection, a detachable connection, or an integral connection; “connected” may They are directly connected or indirectly connected through an intermediary.
  • the specific meanings of the above terms in the present disclosure can be understood by those skilled in the art on a case-by-case basis.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本公开提供了一种人机交互系统、方法、计算机可读存储介质及交互装置,其中,人机交互系统包括:交互模块和对比模块,交互模块用于在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;对比模块用于将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。

Description

人机交互系统、方法、计算机可读存储介质及交互装置
本申请要求于2018年3月29日提交中国专利局、申请号为201810273850.7、申请名称为“人机交互系统、方法、计算机可读存储介质及交互装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本公开涉及人工智能领域,具体而言,涉及一种人机交互系统、人机交互方法、计算机可读存储介质及人机交互装置。
背景技术
本公开对于背景技术的描述属于与本公开相关的相关技术,仅仅是用于说明和便于理解本公开的公开内容,不应理解为申请人明确认为或推定申请人认为是本公开在首次提出申请的申请日的现有技术。
近年来,动作捕捉技术已经成为人体运动姿态研究中的一项关键技术,发挥着越来越重要的作用,人们意识到非常有必要通过识别人体运动姿态实现人体动作和信息设备之间的交互功能。然而已有动作捕捉技术一般应用于大型娱乐设备、动画制作、步态分析、生物力学、人机工程等领域,而随着手机、平板电脑等移动设备的使用普及,手机、平板电脑等移动设备以简单、方便、不受时间和地点限制等特点成为人们娱乐消遣必备品,因此,将动作捕捉技术应用到手机随着手机、平板电脑等移动设备中,以使人体验良好的娱乐效果是亟待的问题。
发明内容
本公开第一方面的实施例提供了一种人机交互方法,包括:
在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果。
可选地,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像之前还包括:根据指令提取与所述指令相对应的姿态模板组。
可选地,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像之前还包括:
根据指令提取与所述指令相对应的音频;
在将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果之前播放所述音频。
可选地,将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果包括:
从所述人的动作图像中提取多个单帧图像;
将所述单帧图像与所述姿态图像进行匹配,并生成匹配结果;
根据所述匹配结果在所述显示单元上显示相应的动画和/或分数。
可选地,根据指令提取与所述指令相对应的姿态模板组包括:
提取预存的多个姿态图像中的一个或多个姿态图像构成与所述指令相对应的姿态模板组。
可选地,所述人机交互方法还包括:
在所述匹配结果包括分数时,所述姿态图像显示结束后将显示的所有分数汇总得到总分;
将总分与预设的分数评级匹配,并在所述显示单元上显示分数所处的等级。
可选地,在显示所述姿态图像之前还包括:
检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始在显示单元上显示所述姿态图像。
本公开第二方面的实施例提供了一种人机交互系统,包括:交互模块,用于在显示单元上显示所述姿态模板组中的一个或多个姿态图像,采集人的动作图像;和对比模块,用于将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果。
可选地,人机交互系统还包括:提取模块,用于根据指令提取与所述指令相对应的姿态模板组。
可选地,人机交互系统还包括:提取模块,用于根据指令提取与所述指令相对应的音频;所述交互模块控制所述音频播放。
可选地,所述对比模块包括:处理单元,用于从所述人的动作图像中提取多个单帧图像;匹配单元,用于将所述单帧图像与所述姿态图像进行匹配,并生成匹配结果;和执行单元,用于根据所述匹配结果在所述显示单元上显示相应的动画和/或分数。
可选地,所述姿态模板组由选自多个预存的姿态图像中的一个或多个姿态图像构成。
可选地,所述人机交互系统还包括:汇总模块,用于在所述匹配结果包括分数时,所述姿态图像显示结束后将显示的所有分数汇总得到总分;和评级模块,用于将总分与预设的分数评级匹配,并在所述显示单元上显示分数所处的等级。
可选地,所述人机交互系统还包括:识别模块,用于检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始在显示单元上显示所述姿态图像。
本公开第三方面的实施例提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述任一项所述人机交互方法的步骤。
本公开第四方面的实施例提供了一种人机交互装置,包括存储器、处理器及存储在存储器上并可在处理器上运行的程序,所述处理器执行所述程序时实现上述任一项所述人机交互方法的步骤。
本公开提供的技术方案,显示单元上显示姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),用户去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,采集用户的图像,将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
本公开的附加方面和优点将在下面的描述部分中变得明显,或通过本公开的实践了解到。
要理解的是,前面的一般描述和下面的详细描述两者都是示例性的,并且意图在于提供要求保护的技术的进一步说明。
附图说明
本公开的上述和/或附加的方面和优点从结合下面附图对实施例的描述中将变得明显和容易理解,其中:
图1为本公开实施例的终端设备的硬件结构示意图;
图2是本公开所述人机交互系统第一种实施例的结构框图;
图3是本公开所述人机交互系统第二种实施例的结构框图;
图4是本公开所述人机交互系统第三种实施例的结构框图;
图5是本公开所述人机交互系统第四种实施例的结构框图;
图6是本公开所述人机交互系统第五种实施例的结构框图;
图7是本公开所述人机交互系统第六实施例的流结构框图;
图8是本公开所述人机交互系统第七实施例的流结构框图;
图9是本公开所述人机交互方法第一种实施例交互方法的流程示意图;
图10是本公开所述人机交互方法第二种实施例交互方法的流程示意图;
图11是本公开所述人机交互方法第三种实施例交互方法的流程示意图;
图12是本公开所述人机交互方法第四种实施例交互方法的流程示意图;
图13是本公开所述人机交互方法第五种实施例交互方法的流程示意图;
图14是本公开所述人机交互方法第六种实施例交互方法的流程示意图;
图15是本公开所述人机交互方法第七种实施例交互方法的流程示意图;
图16是本公开实施例的计算机可读存储介质的示意图;
图17是本公开实施例的人机交互装置的结构示意图。
其中,图1至图8、图16和图17中附图标记与部件名称之间的对应关系为:
100人机交互系统,101提取模块,1011图像单元,1012音频单元,102交互模块,103对比模块,1031处理单元,1032匹配单元,1033执行单元,104汇总模块,105评级模块,106识别模块,1无线通信单元,2输入单元,3用户输入单元,4感测单元,5输出单元,6存储器,7接口单元,8控制器,9电源单元,80人机交互装置,801存储器,802处理器,900计算机可读存储介质,901非暂时性计算机可读指令。
具体实施方式
为了能够更清楚地理解本公开的上述目的、特征和优点,下面结合附图和具体实施方式对本公开进行进一步的详细描述。需要说明的是,在不冲突的情况下,本申请的实施例及实施例中的特征可以相互组合。
在下面的描述中阐述了很多具体细节以便于充分理解本公开,但是,本公开还可以采用其他不同于在此描述的其他方式来实施,因此,本公开的保护范围并不受下面公开的具体实施例的限制。
下述讨论提供了本公开的多个实施例。虽然每个实施例代表了发明的单一组合,但是本公开不同实施例可以替换,或者合并组合,因此本公开也可认为包含所记载的相同和/或不同实施例的所有可能组合。因而,如果一个实施例包含A、B、C,另一个实施例包含B和D的组合,那么本公开也应视为包括含有A、B、C、D的一个或多个所有其他可能的组合的实施例,尽管该实施例可能并未在以下内容中有明确的文字记载。
如图1所示,人机交互装置即终端设备可以以各种形式来实施,本公开中的终端设备可以包括但不限于诸如移动电话、智能电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携 式多媒体播放器)、导航装置、车载终端设备、车载显示终端、车载电子后视镜等等的移动终端设备以及诸如数字TV、台式计算机等等的固定终端设备。
在本公开的一个实施例中,终端设备可以包括无线通信单元1、A/V(音频/视频)输入单元2、用户输入单元3、感测单元4、输出单元5、存储器6、接口单元7、控制器8和电源单元9等等。其中,A/V(音频/视频)输入单元2包括但不限于,摄像头、前置摄像头,后置摄像头,各类音视频输入设备。本领域的技术人员应该理解,上述实施例列出的终端设备所包括的组件,不止上述所述的种类,可以包括更少或者更多的组件。
本领域的技术人员应该理解,这里描述的各种实施方式可以以使用例如计算机软件、硬件或其任何组合的计算机可读介质来实施。对于硬件实施,这里描述的实施方式可以通过使用特定用途集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理装置(DSPD)、可编程逻辑装置(PLD)、现场可编程门阵列(FPGA)、处理器、控制器、微控制器、微处理器、被设计为执行这里描述的功能的电子单元中的至少一种来实施,在一些情况下,这样的实施方式可以在控制器中实施。对于软件实施,诸如过程或功能的实施方式可以与允许执行至少一种功能或操作的单独的软件模块来实施。软件代码可以由以任何适当的编程语言编写的软件应用程序(或程序)来实施,软件代码可以存储在存储器中并且由控制器执行。
如图2所示,本公开第一方面的实施例提供的人机交互系统100,包括:交互模块102和对比模块103。
根据示例实施例,交互模块102用于在显示单元上显示姿态模板组中的一个或多个姿态图像,并采集人的动作图像;对比模块103用于将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
本公开提供的人机交互系统100,显示单元(显示单元可为显示屏等)上显示姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),在姿态图像显示的是人手、大臂、小臂、大腿、小腿、躯干及头等在不同时刻的位置、角度等,用户去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,交互模块采集用户的图像,对比模块将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
如图3所示,本公开第一方面的实施例提供的人机交互系统100,包括:提取模块101、交互模块102和对比模块103。
根据示例实施例,提取模块101用于根据指令提取与指令相对应的姿态模板组;交互模块102用于在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;对比模块103用于将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
本公开提供的人机交互系统100,显示单元(显示单元可为显示屏等)上显示姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),在姿态图像显示的是人手、大臂、小臂、大腿、小腿、躯干及头等在不同时刻的位置、角度等,用户去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,交互模块采集用户的图像,对比模块将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
在本公开的一个实施例中,提取模块101用于根据指令提取与指令相对应的姿态模板组及音频;交互模块102用于控制音频播放,在显示单元上显示姿态模板组中的多个姿态图像,采集人的动作图像;对比模块103用于将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
本公开提供的人机交互系统100,随着音乐的播放显示单元(显示单元可为显示屏等)上显示姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),在姿态图像显示的是人手、大臂、小臂、大腿、小腿、躯干及头等在不同时刻的位置、角度等,用户随着音乐去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,交互模块采集用户的图像,对比模块将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
在本公开的一个实施例中,如图4所示,对比模块103包括:处理单元1031、匹配单元1032和执行单元1033。
根据示例实施例,处理单元1031用于从人的动作图像中提取多个单帧图像;匹配单元1032用于将单帧图像与姿态图像进行匹配,并生成匹配结果;执行单元1033用于根据匹配结果在显示单元上显示相应的动画和/或分数。
在该实施例中,处理单元1031从采集人的动作图像转中提取多个单帧图像,如单位时间内取一百帧图像,匹配单元1032通过将一百帧图像与姿态图像匹配,判断一百帧与姿态图像的重合率,该种检测方式能够实现精准的检测,提高了产品的检测精度,从而提高了产品的体验效果;显示单 元上显示分数和/或动画,动画可为perfect、good、great、miss等数字动画,或从显示单元上显示下心雨、下星星雨等特效。
在本公开的一个实施例中,如图5所示,提取模块101包括:图像单元1011和音频单元1012。
根据示例实施例,图像单元1011用于提取多个姿态图像中的一个或多个姿态图像构成与指令相对应的姿态模板组;音频单元1012用于调出与指令相对应的音频。
在该实施例中,图像单元1011中预先存储有很多姿态模板,图像单元1011根据用户选择不同的指令提取多个姿态模板并排序构成姿态模板组;在本公开的一个实施例中,包括一百个姿态模板,根据第一指令提取第1、3、5、20、66、78、82、96姿态模板构成姿态模板组;根据第二指令提取第2、12、22、25、37、47、55、69、73、86、96姿态模板构成姿态模板组;根据第三指令提取第7、13、29、35、38、46、52、68、71、86、91姿态模板构成姿态模板组,音频单元1012根据用户选择不同的指令提取相应的音乐。
在本公开的一个实施例中,提取模块101中预先存储有很多姿态模板,姿态模板组由选自多个预存的姿态图像中的一个或多个姿态图像构成,提取模块101根据用户选择不同的指令提取多个姿态模板并排序构成姿态模板组;在本公开的一个实施例中,包括一百个姿态模板,根据第一指令提取第1、3、5、20、66、78、82、96姿态模板构成姿态模板组;根据第二指令提取第2、12、22、25、37、47、55、69、73、86、96姿态模板构成姿态模板组;根据第三指令提取第7、13、29、35、38、46、52、68、71、86、91姿态模板构成姿态模板组。
在本公开的一个实施例中,如图6所示,人机交互系统100还包括:汇总模块104和评级模块105。
根据示例实施例,汇总模块104用于在匹配结果包括分数时,音频结束后或姿态图像显示结束后将显示的所有分数汇总得到总分;评级模块105用于将总分与预设的分数评级匹配,并在显示单元上显示分数所处的等级。
在该实施例中,通过汇总模块104和评级模块105用户可以知道这段跳舞的分数和等级,一方面,用户可以根据分数和等级与其他用户的分数和等级进行排名,增加了产品的互动性和趣味性,另一方面,用户可将带有分数和等级的视频分享给朋友,使朋友能够对自己跳舞好坏有个直接的判断。
在本公开的一个实施例中,如图7和图8所示,人机交互系统100还包括:识别模块106,识别模块106用于检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始播放音频和/或准备开始在显示单元上显示姿态图像。
在该实施例中,识别模块106的设置,一方面,能够保证用户完全处在显示单元内,能够使用户做的舞蹈动作与显示单元上的姿态图像更好的匹配,即避免了用户的肢体超出显示单元导致匹配不准的情况发送,从而提高了产品的使用舒适度,进而提高了产品的市场竞争力,另一方面,使用户与手机之间的距离处于合理的范围内,从而使用户能够清楚的看到显示单元上显示的内容,进而增加了产品的使用舒适,进而增加了产品的市场竞争力。
实施例一
如图9所示,本公开第二方面的实施例提供的人机交互方法,包括如下步骤:
步骤30,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
本公开提供的人机交互的方法,显示单元(显示单元可为显示屏等)上显示姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),在姿态图像显示的是人手、大臂、小臂、大腿、小腿、躯干及头等在不同时刻的位置、角度等,用户去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,采集用户的图像,将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
实施例二
在本公开的一个实施例中,如图10所示,本实施例中人机交互方法包括:
步骤10,根据指令提取与指令相对应的姿态模板组及音频;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
本公开提供的人机交互的方法,随着音乐的播放显示单元(显示单元可为显示屏等)上显示姿态图像(如呈不同姿态多个火柴人、动画、动物 图像等),在姿态图像显示的是人手、大臂、小臂、大腿、小腿、躯干及头等在不同时刻的位置、角度等,用户随着音乐去做跟这些姿态图像相同的肢体动作,使用户形成跳舞的动作,同时,采集用户的图像,将人的动作图像与姿态图像匹配,并根据人的动作与姿态图像的匹配程度,在显示单元上显示匹配结果(如分数和/或动画特效),对不太会跳舞的用户有引导作用,使用户能够做标准的舞蹈动作,提高了娱乐效果,从而提高了用户的体验效果。
实施例三
在本公开的一个实施例中,如图11所示,步骤40包括:
步骤41,从人的动作图像中提取多个单帧图像;
步骤42,将单帧图像与姿态图像进行匹配,并生成匹配结果;
步骤43,根据匹配结果在显示单元上显示相应的动画和/或分数。
在本实施例中人机交互方法包括:
步骤10,根据指令提取与指令相对应的姿态模板组及音频;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤41,从人的动作图像中提取多个单帧图像;
步骤42,将单帧图像与姿态图像进行匹配,并生成匹配结果;
步骤43,根据匹配结果在显示单元上显示相应的动画和/或分数。
在该实施例中,从采集人的动作图像转中提取多个单帧图像,如单位时间内取一百帧图像,匹配单元通过将一百帧图像与姿态图像匹配,判断一百帧与姿态图像的重合率,该种检测方式能够实现精准的检测,提高了产品的检测精度,从而提高了产品的体验效果;显示单元上显示分数和/或 动画,动画可为perfect、good、great、miss等数字动画,或从显示单元上显示下心雨、下星星雨等特效。
实施例四
在本公开的一个实施例中,如图12所示,步骤10包括:
步骤11,提取多个姿态图像中的一个或多个姿态图像构成与指令相对应的姿态模板组;
步骤12,调出与指令相对应的音频;
在本实施例中人机交互方法包括:
步骤11,提取预存的多个姿态图像中的一个或多个姿态图像构成与指令相对应的姿态模板组;
步骤12,调出与指令相对应的音频;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
在该实施例中,预先存储有很多姿态模板,姿态模板组由选自多个预存的姿态图像中的一个或多个姿态图像构成,根据用户选择不同的指令提取多个姿态模板并排序构成姿态模板组;在本公开的一个实施例中,包括一百个姿态模板,根据第一指令提取第1、3、5、20、66、78、82、96姿态模板构成姿态模板组;根据第二指令提取第2、12、22、25、37、47、55、69、73、86、96姿态模板构成姿态模板组;根据第三指令提取第7、13、29、35、38、46、52、68、71、86、91姿态模板构成姿态模板组,根据用户选择不同的指令提取相应的音乐。
实施例五
在本公开的一个实施例中,如图13所示,人机交互方法还包括:
步骤50,在匹配结果包括分数时,音频结束或姿态图像显示结束后将显示的所有分数汇总得到总分;
步骤60,将总分与预设的分数评级匹配,并在显示单元上显示分数所处的等级。
在本实施例中人机交互方法包括:
步骤10,根据指令提取与指令相对应的姿态模板组及音频;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果;
步骤50,在匹配结果包括分数时,音频结束或姿态图像显示结束后将显示的所有分数汇总得到总分;
步骤60,将总分与预设的分数评级匹配,并在显示单元上显示分数所处的等级。
在该实施例中,用户可以知道这段跳舞的分数和等级,一方面,用户可以根据分数和等级与其他用户的分数和等级进行排名,增加了产品的互动性和趣味性,另一方面,用户可将带有分数和等级的视频分享给朋友,使朋友能够对自己跳舞好坏有个直接的判断。
实施例六
在本公开的一个实施例中,如图14和图15所示,在显示姿态图像之前还包括:
步骤20,检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始播放音频和/或准备开始在显示单元上显示姿态图像。
如图14所示,在本实施例中人机交互方法包括:
步骤10,根据指令提取与指令相对应的姿态模板组及音频;
步骤20,检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始播放音频和/或准备开始在显示单元上显示姿态图像;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果。
或者,如图15所示,实施例七
步骤10,根据指令提取与指令相对应的姿态模板组及音频;
步骤20,检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始播放音频和/或准备开始在显示单元上显示姿态图像;
步骤30,播放音频,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
步骤40,将人的动作图像与当前显示的姿态图像匹配,并在显示单元上显示匹配结果;
步骤50,在匹配结果包括分数时,音频结束后或姿态图像显示结束后将显示的所有分数汇总得到总分;
步骤60,将总分与预设的分数评级匹配,并在显示单元上显示分数所处的等级。
在该实施例中,识别步骤的设置,一方面,能够保证用户完全处在显示单元的显示区域内,能够使用户做的舞蹈动作与显示单元上的姿态图像更好的匹配,即避免了用户的肢体超出显示单元导致匹配不准的情况发生,从而提高了产品的使用舒适度,进而提高了产品的市场竞争力,另一 方面,使用户与手机之间的距离处于合理的范围内,从而使用户能够清楚的看到显示单元上显示的内容,进而增加了产品的使用舒适,进而增加了产品的市场竞争力。
在本公开的一个实施例中,显示单元上显示识别框,人的图像处于识别框内时,准备开始播放音频或准备开始在显示单元上显示姿态图像,可选地,识别框为人形框,该人形框为人的全身形状,用户根据人形框做与人形框形状相同的动作,且用户的全部位于人形框内时,准备开始播放音频或准备开始在显示单元上显示姿态图像,显示单元上出现倒计时,倒计时结束播放音频。在本公开的另一具体实施例中,识别框为人形框,该人形框为人的上半身形状(也可为下半身),用户根据人形框做与人形框形状相同的动作,且用户的上半身位于人形框内时,准备开始播放音频或准备开始在显示单元上显示姿态图像,显示单元上出现倒计时,倒计时结束播放音频。本领域的技术的人员应该理解,识别框的作用是识别目标,以确保交互的顺利的进行,因此,只要能够满足识别作用的识别框都应该在本公开的保护范围内。
如图16所示,本公开第三方面的实施例提供的计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述任一项人机交互方法的步骤。其中,计算机可读存储介质可以包括但不限于任何类型的盘,包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等等)、静态随机访问存储器(SRAM)、电可擦除可编程只读存储器(EEPROM)、可编程只读存储器(PROM)、磁性存储器、软盘、光盘、DVD、CD-ROM、微型驱动器以及磁光盘、ROM、RAM、EPROM、EEPROM、DRAM、VRAM、闪速存储器设备、磁卡或光卡、纳米系统(包括分子存储器IC),或适合于存储指令和/或数据的任何类型的媒介或 设备。在本公开的一个实施例中,计算机可读存储介质900其上存储有非暂时性计算机可读指令901。当所述非暂时性计算机可读指令901由处理器运行时,执行参照上述描述的根据本公开实施例的基于人体动态姿态的人机交互方法
本公开第四方面的实施例提供的人机交互装置,包括存储器、处理器及存储在存储器上并可在处理器上运行的程序,处理器执行程序时实现上述任一项人机交互方法方法的步骤。
在本公开的一个实施例中,存储器用于存储非暂时性计算机可读指令。根据示例实施例,存储器可以包括一个或多个计算机程序产品,计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。易失性存储器例如可以包括随机存取存储器(RAM)和/或高速缓冲存储器(cache)等。非易失性存储器例如可以包括只读存储器(ROM)、硬盘、闪存等。在本公开的一个实施例中,处理器可以是中央处理单元(CPU)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元,并且可以控制人机交互装置中的其它组件以执行期望的功能。在本公开的一个实施例中,处理器用于运行存储器中存储的计算机可读指令,使得人机交互装置执行上述交互方法。
在本公开的一个实施例中,如图17所示,人机交互装置80包括存储器801和处理器802。人机交互装置80中的各组件通过总线系统和/或其它形式的连接机构(未示出)互连。
存储器801用于存储非暂时性计算机可读指令。根据示例实施例,存储器801可以包括一个或多个计算机程序产品,计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。易失性存储器例如可以包括随机存取存储器(RAM)和/或高速缓冲存 储器(cache)等。非易失性存储器例如可以包括只读存储器(ROM)、硬盘、闪存等。
处理器802可以是中央处理单元(CPU)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元,并且可以控制人机交互装置80中的其它组件以执行期望的功能。在本公开的一个实施例中,所述处理器802用于运行存储器801中存储的计算机可读指令,使得人机交互装置80执行上述基于人体动态姿态的人机交互方法。人机交互装置与上述基于人体动态姿态的人机交互方法描述的实施例相同,在此将省略其重复描述。
在本公开的一个实施例中,人机交互装置为移动设备,移动设备的摄像头采集用户的图像,通过移动设备下载与指令相对应的歌曲和姿态模板组,歌曲和姿态模板组下载好后,移动设备的显示单元上出现识别框(该识框可为人形框),通过调整用户与移动设备的距离,将用户的图像处于识别框内,移动设备开始播放音乐,同时显示单元上显示多个姿态图像(如呈不同姿态多个火柴人、动画、动物图像等),用户开始做舞蹈动作,以使自己的肢体动作与这些姿态图像匹配,根据用户的动作与姿态图像的匹配程度,在显示单元上显示分数和/或动画(动画可为perfect、good、great、miss等数字动画,或者从显示单元上显示下心雨、下星星雨等特效),音乐播放完成后,移动设备的显示单元上显示分数和等级,用户可将自己的跳舞视频下载下来或者分享出去或者进入排行榜,移动设备可为手机、平板电脑等。
在本公开中,术语“多个”则指两个或两个以上,除非另有明确的限定。术语“安装”、“相连”、“连接”、“固定”等术语均应做广义理解,例如,“连接”可以是固定连接,也可以是可拆卸连接,或一体地连接;“相连”可以 是直接相连,也可以通过中间媒介间接相连。对于本领域的普通技术人员而言,可以根据具体情况理解上述术语在本公开中的具体含义。
在本说明书的描述中,术语“一个实施例”、“一些实施例”、“具体实施例”等的描述意指结合该实施例或示例描述的具体特征、结构、材料或特点包含于本公开的至少一个实施例或示例中。在本说明书中,对上述术语的示意性表述不一定指的是相同的实施例或实例。而且,描述的具体特征、结构、材料或特点可以在任何的一个或多个实施例或示例中以合适的方式结合。
以上所述仅为本公开的可选实施例而已,并不用于限制本公开,对于本领域的技术人员来说,本公开可以有各种更改和变化。凡在本公开的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本公开的保护范围之内。

Claims (16)

  1. 一种人机交互方法,包括:
    在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;
    将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果。
  2. 根据权利要求1所述的人机交互方法,其中,在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像之前还包括:
    根据指令提取与所述指令相对应的姿态模板组。
  3. 根据权利要求1所述的人机交互方法,其中,在显示单元上显示姿态模板组中的一个或多个姿态图像,并采集人的动作图像之前还包括:
    根据指令提取与所述指令相对应的音频;
    在将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果之前播放所述音频。
  4. 根据权利要求1所述的人机交互方法,其中,将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果包括:
    从所述人的动作图像中提取多个单帧图像;
    将所述单帧图像与所述姿态图像进行匹配,并生成匹配结果;
    根据所述匹配结果在所述显示单元上显示相应的动画和/或分数。
  5. 根据权利要求2所述的人机交互方法,其中,根据指令提取与所述指令相对应的姿态模板组包括:
    提取预存的多个姿态图像中的一个或多个姿态图像构成与所述指令相对应的姿态模板组。
  6. 根据权利要求4所述的人机交互方法,其中,还包括:
    在所述匹配结果包括分数时,所述姿态图像显示结束后将显示的所有分数汇总得到总分;
    将总分与预设的分数评级匹配,并在所述显示单元上显示分数所处的等级。
  7. 根据权利要求1至6中任一项所述的人机交互方法,其中,
    在显示所述姿态图像之前还包括:
    检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始在显示单元上显示所述姿态图像。
  8. 一种人机交互系统,包括:
    交互模块,用于在显示单元上显示姿态模板组中的一个或多个姿态图像,采集人的动作图像;和
    对比模块,用于将所述人的动作图像与当前显示的所述姿态图像匹配,并在所述显示单元上显示匹配结果。
  9. 根据权利要求8所述的人机交互系统,其中,还包括:
    提取模块,用于根据指令提取与所述指令相对应的姿态模板组。
  10. 根据权利要求8所述的人机交互系统,其中,还包括:
    提取模块,用于根据指令提取与所述指令相对应的音频;
    所述交互模块控制所述音频播放。
  11. 根据权利要求8所述的人机交互系统,其中,所述对比模块包括:
    处理单元,用于从所述人的动作图像中提取多个单帧图像;
    匹配单元,用于将所述单帧图像与所述姿态图像进行匹配,并生成匹配结果;和
    执行单元,用于根据所述匹配结果在所述显示单元上显示相应的动画和/或分数。
  12. 根据权利要求8所述的人机交互系统,其中,
    所述姿态模板组由预存的多个姿态图像中的一个或多个姿态图像构成。
  13. 根据权利要求11所述的人机交互系统,其中,还包括:
    汇总模块,用于在所述匹配结果包括分数时,所述姿态图像显示结束后将显示的所有分数汇总得到总分;和
    评级模块,用于将总分与预设的分数评级匹配,并在所述显示单元上显示分数所处的等级。
  14. 根据权利要求8至13中任一项所述的人机交互系统,其中,还包括:
    识别模块,用于检测人机之间的距离,当人机之间的距离在预设范围内时,准备开始在显示单元上显示所述姿态图像。
  15. 一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现所述权利要求1-7中任一项所述人机交互方法的步骤。
  16. 一种人机交互装置,包括存储器、处理器及存储在存储器上并可在处理器上运行的程序,所述处理器执行所述程序时实现所述权利要求1-7中任一项所述人机交互方法的步骤。
PCT/CN2019/075928 2018-03-29 2019-02-22 人机交互系统、方法、计算机可读存储介质及交互装置 WO2019184634A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/020,561 US20200409471A1 (en) 2018-03-29 2020-09-14 Human-machine interaction system, method, computer readable storage medium and interaction device

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810273850.7A CN108536293B (zh) 2018-03-29 2018-03-29 人机交互系统、方法、计算机可读存储介质及交互装置
CN201810273850.7 2018-03-29

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/020,561 Continuation US20200409471A1 (en) 2018-03-29 2020-09-14 Human-machine interaction system, method, computer readable storage medium and interaction device

Publications (1)

Publication Number Publication Date
WO2019184634A1 true WO2019184634A1 (zh) 2019-10-03

Family

ID=63481632

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/075928 WO2019184634A1 (zh) 2018-03-29 2019-02-22 人机交互系统、方法、计算机可读存储介质及交互装置

Country Status (3)

Country Link
US (1) US20200409471A1 (zh)
CN (1) CN108536293B (zh)
WO (1) WO2019184634A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4167067A4 (en) * 2020-07-24 2023-12-13 Beijing Bytedance Network Technology Co., Ltd. INTERACTION METHOD AND APPARATUS, DEVICE AND READABLE MEDIUM

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108536293B (zh) * 2018-03-29 2020-06-30 北京字节跳动网络技术有限公司 人机交互系统、方法、计算机可读存储介质及交互装置
CN109669601A (zh) * 2018-09-26 2019-04-23 深圳壹账通智能科技有限公司 信息展示方法、信息展示装置、信息展示设备及存储介质
CN109376669A (zh) * 2018-10-30 2019-02-22 南昌努比亚技术有限公司 智能助手的控制方法、移动终端及计算机可读存储介质
CN109985380A (zh) * 2019-04-09 2019-07-09 北京马尔马拉科技有限公司 互联网游戏人机交互方法及系统
CN113596353B (zh) * 2021-08-10 2024-06-14 广州艾美网络科技有限公司 体感互动数据处理方法、装置及体感互动设备

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120198359A1 (en) * 2011-01-28 2012-08-02 VLoungers, LLC Computer implemented system and method of virtual interaction between users of a virtual social environment
CN102724449A (zh) * 2011-03-31 2012-10-10 青岛海信电器股份有限公司 互动式电视机和利用显示装置实现与用户互动的方法
CN108536293A (zh) * 2018-03-29 2018-09-14 北京字节跳动网络技术有限公司 人机交互系统、方法、计算机可读存储介质及交互装置

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9358456B1 (en) * 2010-06-11 2016-06-07 Harmonix Music Systems, Inc. Dance competition game
CN102419819B (zh) * 2010-10-25 2014-10-08 深圳市中控生物识别技术有限公司 人脸图像识别方法和系统
CN102500094B (zh) * 2011-10-28 2013-10-30 北京航空航天大学 一种基于kinect的动作训练方法
CN106446569A (zh) * 2016-09-29 2017-02-22 宇龙计算机通信科技(深圳)有限公司 一种运动指导方法和终端
CN106448279A (zh) * 2016-10-27 2017-02-22 重庆淘亿科技有限公司 一种舞蹈教学互动体验方法及系统
CN107038455B (zh) * 2017-03-22 2019-06-28 腾讯科技(深圳)有限公司 一种图像处理方法及装置

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120198359A1 (en) * 2011-01-28 2012-08-02 VLoungers, LLC Computer implemented system and method of virtual interaction between users of a virtual social environment
CN102724449A (zh) * 2011-03-31 2012-10-10 青岛海信电器股份有限公司 互动式电视机和利用显示装置实现与用户互动的方法
CN108536293A (zh) * 2018-03-29 2018-09-14 北京字节跳动网络技术有限公司 人机交互系统、方法、计算机可读存储介质及交互装置

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4167067A4 (en) * 2020-07-24 2023-12-13 Beijing Bytedance Network Technology Co., Ltd. INTERACTION METHOD AND APPARATUS, DEVICE AND READABLE MEDIUM

Also Published As

Publication number Publication date
US20200409471A1 (en) 2020-12-31
CN108536293B (zh) 2020-06-30
CN108536293A (zh) 2018-09-14

Similar Documents

Publication Publication Date Title
WO2019184634A1 (zh) 人机交互系统、方法、计算机可读存储介质及交互装置
WO2019184633A1 (zh) 人机交互系统、方法、计算机可读存储介质及交互装置
CN111726536B (zh) 视频生成方法、装置、存储介质及计算机设备
CN108289180B (zh) 根据肢体动作处理视频的方法、介质和终端装置
WO2022166243A1 (zh) 捏合手势检测识别方法、装置及系统
TW201113743A (en) Method, electronic apparatus and computer program product for creating biologic feature data
US20140285517A1 (en) Display device and method to display action video
CN106571072A (zh) 基于增强现实幼教卡片的实现方法
JPWO2017047182A1 (ja) 情報処理装置、情報処理方法、及びプログラム
CN109034063A (zh) 人脸特效的多人脸跟踪方法、装置和电子设备
WO2022174594A1 (zh) 基于多相机的裸手追踪显示方法、装置及系统
US20210279935A1 (en) Motion Tracking and Image Recognition of Hand Gestures to Animate a Digital Puppet, Synchronized with Recorded Audio
CN108096833B (zh) 基于级联神经网络的体感游戏控制方法及装置、计算设备
CN112396679A (zh) 虚拟对象显示方法及装置、电子设备、介质
WO2023108842A1 (zh) 基于健身教学训练的动作评价方法及系统
CN108815845B (zh) 人机交互的信息处理方法及装置、计算机设备及可读介质
CN105468249B (zh) 智能互动系统及其控制方法
WO2015032304A1 (zh) 具有攻击者身份识别能力的在线笔迹身份认证方法
CN108563331A (zh) 动作匹配结果确定装置、方法、可读存储介质及交互装置
CN112613490B (zh) 一种行为识别方法、装置、机器可读介质及设备
CN111563497B (zh) 基于移动轨迹的框题方法、装置、电子设备及存储介质
EP4020964A1 (en) Camera control method and apparatus, and terminal device
US11461576B2 (en) Information processing method and related electronic device
CN114245021B (zh) 互动拍摄方法、电子设备、存储介质及计算机程序产品
CN111079662A (zh) 一种人物识别方法、装置、机器可读介质及设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19774694

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19774694

Country of ref document: EP

Kind code of ref document: A1