WO2022044399A1 - Terminal device and support method for improving form - Google Patents

Terminal device and support method for improving form Download PDF

Info

Publication number
WO2022044399A1
WO2022044399A1 PCT/JP2021/010525 JP2021010525W WO2022044399A1 WO 2022044399 A1 WO2022044399 A1 WO 2022044399A1 JP 2021010525 W JP2021010525 W JP 2021010525W WO 2022044399 A1 WO2022044399 A1 WO 2022044399A1
Authority
WO
WIPO (PCT)
Prior art keywords
skeleton data
instructor
user
data
unit
Prior art date
Application number
PCT/JP2021/010525
Other languages
French (fr)
Japanese (ja)
Inventor
英和 鈴木
洋介 芳賀
知恵子 遠藤
裕明 梨本
恵児 西巻
Original Assignee
株式会社Jvcケンウッド
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Jvcケンウッド filed Critical 株式会社Jvcケンウッド
Publication of WO2022044399A1 publication Critical patent/WO2022044399A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B69/00Training appliances or apparatus for special sports
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B71/00Games or sports accessories not covered in groups A63B1/00 - A63B69/00
    • A63B71/06Indicating or scoring devices for games or players, or for other sports activities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis

Definitions

  • the present invention relates to a terminal device for improving a form of sports or fitness, and a form improvement support method.
  • the above method requires multiple motion sensors and a dedicated data acquisition system. Further, in the above-mentioned method, it is necessary to give feedback to the subject after analyzing the form data acquired from a plurality of motion sensors, and it is difficult to give feedback in real time during the practice of the form.
  • This embodiment was made in view of such a situation, and the purpose is to provide a technique capable of providing support for improving a form in real time with a simple system configuration.
  • the terminal device of an embodiment of the present embodiment has an image data acquisition unit that acquires image data of a user reflected from an imaging unit, and an image in an image based on the acquired image data.
  • a skeletal data generation unit that generates skeletal data of a person connecting a plurality of feature points including joint positions in the person area in which the user is reflected, and a skeletal structure of the instructor generated by a terminal device used by the instructor.
  • the size of the skeletal data acquisition unit for acquiring data, the skeletal data of the instructor acquired by the skeletal data acquisition unit, and the skeletal data of the user instructed by the instructor generated by the skeletal data generation unit are normalized to the skeletal data of the instructor and the skeletal data of the user.
  • a display control unit for controlling the display on the display unit by superimposing the data on the display unit.
  • Another aspect of this embodiment is a form improvement support method.
  • a step of acquiring image data of a user imaged from an imaging unit and a plurality of feature points including joint positions of a person area in the image in which the user is imaged are obtained based on the acquired image data.
  • the step of normalizing at least one of the skeletal data and the image of the user captured by the image pickup unit were normalized so that the size of the skeletal data of the user being instructed by the instructor corresponds. It has a step of superimposing the skeleton data of the instructor and the skeleton data of the user and controlling the display unit to display the skeleton data.
  • Yet another aspect of this embodiment is a form improvement support program.
  • This program acquires image data of the user from the image pickup unit, and based on the acquired image data, a plurality of feature points including joint positions of the person area in which the user is reflected in the image.
  • the process of normalizing at least one of the skeletal data and the image of the user captured by the image pickup unit are normalized so that the size of the skeletal data of the user instructed by the instructor corresponds to the size of the skeletal data.
  • the computer is made to execute a process of superimposing the skeletal data of the instructor and the skeletal data of the user and controlling the display on the display unit.
  • FIGS. 4 (a)-(c) show the normalization of the received instructor's skeleton data, the image of the student displayed on the display of the terminal device used by the student, and the student's skeleton data. It is a figure which shows an example of comparison with the skeleton data of a leader. It is a flowchart which shows the flow of the form improvement support process using the terminal apparatus which concerns on embodiment.
  • FIG. 1 is a diagram showing a configuration of a form improvement support system 1 according to an embodiment.
  • the form improvement support system 1 includes a plurality of terminal devices 3, and the plurality of terminal devices 3 are connected via the network 2.
  • One of the plurality of terminal devices 3 is used by the instructor, and the other is used by the student who receives the instruction from the instructor.
  • an instructor is a person who can instruct a student in the correct exercise form in a particular discipline, including instructors, coaches, and school teachers.
  • the plurality of terminal devices 3 are connected by a LAN (Local Area Network). If the student and the instructor are in the same place, the student can receive direct instruction from the instructor, but if there are many students, the instructor's instruction may not be sufficient. Also, it may be difficult to see the correct form of the instructor from behind. By using the terminal device 3, students can intuitively and accurately grasp the difference between their own form and the form of the instructor who should be a model, and efficiently acquire the correct form. Can be done.
  • LAN Local Area Network
  • the plurality of terminal devices 3 are connected by a wide area network such as the Internet or a dedicated line.
  • the student can receive the guidance of the instructor remotely while staying at home or in the office.
  • a popular instructor at the headquarters can also remotely instruct students in a satellite studio.
  • FIG. 1 is a diagram assuming an example in which the instructor and the student are 1: N (N is an integer of 2 or more), but the instructor and the student may have an N: N relationship. In that case, the leaders may be geographically separated from each other. In the case of individual instruction, the instructor and the student have a 1: 1 relationship.
  • the terminal device 3 may be a dedicated terminal device for supporting the improvement of the exercise form, or may be a general-purpose terminal device such as a tablet, a notebook PC, or a smartphone. In the following explanation, a dedicated machine is assumed. For example, when used in a sports club, the terminal device 3 may be rented out to the student from the sports club.
  • the terminal device 3 includes a processing unit 10, a storage unit 20, a communication unit 30, an imaging unit 40, an operation unit 50, a display unit 60, an audio output unit 70, and a vibration device 80.
  • the processing unit 10 includes an image data acquisition unit 11, a skeleton data generation unit 12, a skeleton data transmission / reception unit 13, an operation reception unit 14, a normalization unit 15, a display control unit 16, an alignment unit 17, a deviation amount calculation unit 18, and a message.
  • the generation unit 19 and the notification control unit 110 are included.
  • the function of the processing unit 10 can be realized only by the collaboration of hardware resources and software resources, or by hardware resources only.
  • hardware resources CPU, GPU, ROM, RAM, DSP (Digital Signal Processor), ISP (Image Signal Processor), ASIC (Application Specific Integrated Circuit), FPGA (Field Programmable Gate Array), SoC (System on a Chip) , Other LSIs can be used.
  • Programs such as operating systems, middleware, and applications can be used as software resources.
  • the storage unit 20 has a non-volatile recording medium (for example, NAND flash memory, HDD (Hard Disk Drive)), and includes a dictionary data holding unit 21.
  • a removable recording medium for example, an SD card may be attached to the storage unit 20.
  • the communication unit 30 performs signal processing for connecting to the network 2. Specifically, it performs signal processing based on communication protocols such as TCP / IP (Transmission Control Protocol / Internet Protocol) and Ethernet (registered trademark).
  • the communication unit 30 can be connected to the network 2 via a router (not shown).
  • the router and the terminal device 3 can be connected by LAN cable or Wi-Fi (registered trademark).
  • the mobile phone network may be used to connect to the network 2.
  • the communication unit 30 also performs signal processing for short-range wireless communication with a device existing at a short distance (vibration device 80 in this embodiment).
  • a short-range wireless communication Bluetooth (registered trademark), Wi-Fi, infrared communication, etc. can be used.
  • the imaging unit 40 is a camera for photographing a student or an instructor who is exercising (hereinafter, both are collectively referred to as a user). It is desirable that the image pickup unit 40 is arranged at a position designated by the user and away from the user. Therefore, the image pickup unit 40 and the main body of the terminal device 3 may be connected by a cable (for example, a USB cable) or wirelessly (for example, Wi-Fi). Further, an external Web camera may be used as the image pickup unit 40.
  • the image pickup unit 40 includes a solid-state image sensor and a signal processing circuit.
  • a CMOS (Complementary Metal Oxide Semiconductor) image sensor or a CCD (Charge Coupled Device) can be used as the solid-state image sensor.
  • the solid-state image sensor converts the incident light into an electrical image signal and outputs it to a signal processing circuit.
  • the signal processing circuit performs signal processing such as A / D conversion and noise removal on the image signal input from the solid-state image sensor, and outputs the image data as digital value to the processing unit 10.
  • the operation unit 50 has a touch panel and physical buttons, receives a user's operation, and outputs the operation to the processing unit 10.
  • the display unit 60 has a display such as a liquid crystal display, an organic EL display, and a micro LED display, and displays images and characters supplied from the processing unit 10. In this embodiment, it is assumed that a touch panel display integrated with the touch panel function is used.
  • a large-screen external display may be connected to the main body of the terminal device 3.
  • a large screen TV or PC monitor may be connected to the main body of the terminal device 3 by using a cable (for example, HDMI (registered trademark) cable, USB cable) or wireless (for example, Wi-Fi). ..
  • the voice output unit 70 has a speaker and outputs the voice supplied from the processing unit 10.
  • a voice message for improving the exercise form is output.
  • the voice message may be notified to the user by using a wireless earphone wirelessly (for example, Bluetooth) connected to the main body of the terminal device 3.
  • the vibration device 80 is a band-type vibration device worn on a user's arm or foot (for example, both wrists and both ankles), and vibrates in response to a signal supplied from the processing unit 10.
  • the main body of the terminal device 3 and the vibration device 80 are wirelessly (for example, Bluetooth, Wi-Fi) connected.
  • the vibration device 80 is optional and can be omitted.
  • the image data acquisition unit 11 acquires image data showing the entire body of the user from the image pickup unit 40.
  • the skeleton data generation unit 12 is a person who connects a plurality of feature points (also referred to as key points) including a plurality of joint positions in a person area in which the user is shown in the image with a line. Generate skeleton data for.
  • the skeleton data generation unit 12 can generate skeleton data using an existing posture estimation model of a person (for example, OpenPose).
  • the skeleton data generation unit 12 generates skeleton data of a person with a plurality of feature points extracted from the person image and coordinate data defining the relationship between the feature points.
  • the operation reception unit 14 accepts the user's operation.
  • the operation reception unit 14 receives input and change of various setting information. For example, as setting information, selection of instructor mode or student mode and setting information of a communication destination (for example, IP address, port number) are accepted. It also accepts inputs such as name, gender, age, identification number uniquely assigned to each user, and body shape information. Accepts height and weight inputs as body shape information. Further, as auxiliary information of body shape information, input such as shoulder width, waist width, arm length, and leg length may be accepted. This makes it possible to estimate the length and width of each part of the person reflected in the image with high accuracy.
  • setting information selection of instructor mode or student mode and setting information of a communication destination (for example, IP address, port number) are accepted. It also accepts inputs such as name, gender, age, identification number uniquely assigned to each user, and body shape information. Accepts height and weight inputs as body shape information. Further, as auxiliary information of body shape information, input such as shoulder width, waist width, arm length, and
  • the operation reception unit 14 accepts the selection of the exercise item, the selection of the form to be acquired in the selected item, and the selection of the joint position to be the reference position based on the operation of the instructor.
  • FIG. 2 is a diagram showing an example of a selection screen displayed on the display unit 60 of the terminal device 3 used by the instructor.
  • the instructor selects an item from the item selection field 61, selects a form from the form selection field 62, and selects a joint from the reference position selection field 63.
  • the form selection field 62 a plurality of forms of the item selected in the item selection field 61 are displayed as selection candidates.
  • the form selection field 62 since yoga is selected in the item selection field 61, the form selection field 62 includes a cat pose, a cow face pose, ..., a board pose, a warrior pose A, and so on. Warrior pose B, ... is displayed as a selection candidate.
  • the warrior pose A is selected in the form selection field 62.
  • the instructor selects a new pose each time the pose changes.
  • a plurality of patterns of pitching form, a plurality of patterns of batting form, a plurality of patterns of catching form, ... are displayed as selection candidates in the form selection field 62.
  • the instructor selects a joint to be the reference position of the form selected in the form selection field 62 from the reference position selection field 63.
  • the hip joint is selected.
  • the selected reference position is used as a reference point when collating the skeletal data of the instructor with the skeletal data of the student.
  • the joint closest to the center of gravity of the body may be selected as the joint that should be the reference position for each form.
  • a joint that serves as a fulcrum of the moving part may be selected.
  • the shoulder joint may be selected. The instructor can select the joint that is judged to be optimal, regardless of the center of gravity of the body or the fulcrum of movement, in consideration of the visibility when collating the skeletal data.
  • the joint that should be the reference position for each form may be preset. In this case, when a form is selected from the form selection field 62 by the instructor, the joint associated with the selected form is automatically determined. In this case, the instructor does not have to select the joint that should be the reference position.
  • the dictionary data holding unit 21 in the storage unit 20 uses a classifier generated by machine learning a large number of skeletal data based on a large number of images showing the posture and movement of the instructor as dictionary data for each form of each item. Hold.
  • the joint position to be the reference position of the form may be registered in advance. In this case, the above-mentioned reference position can be automatically determined.
  • the skeleton data transmission / reception unit 13 transmits the skeleton data and the reference position information of the instructor generated by the skeleton data generation unit 12 to the student terminal device 3 via the network 2.
  • the skeleton data transmission / reception unit 13 receives the skeleton data and the reference position information of the instructor transmitted from the instructor's terminal device 3 via the network 2.
  • the display control unit 16 superimposes the skeleton data of the user on the image of the user captured by the image pickup unit 40 and displays it on the display unit 60.
  • FIG. 3 (a)-(b) are diagrams showing an image showing the instructor Pi displayed on the display unit 60 of the terminal device 3 used by the instructor, and an example of the skeleton data Si of the instructor Pi. Is.
  • the display control unit 16 superimposes and displays the skeleton data Si generated based on the image data of the area on the image area in which the instructor Pi is reflected in the image.
  • the skeleton data Si changes in real time following the movement of the instructor Pi in the image.
  • the above-mentioned instructor selects a joint to be a reference position by touching a specific feature point of the skeleton data Si displayed on the touch panel display that integrally constitutes the display unit 60 and the operation unit 50. It may be done by.
  • the hip joint can be selected by touching the point A in FIG.
  • the process of the processing unit 10 described below is a process that is executed in the student mode and not executed in the instructor mode.
  • the normalization unit 15 generates at least one skeleton data so that the size of the skeleton data of the instructor received by the skeleton data transmission / reception unit 13 and the size of the student's own skeleton data generated by the skeleton data generation unit 12 correspond to each other. Normalize.
  • the normalization unit 15 normalizes the instructor's skeleton data so that the received instructor's skeleton data fits the size of the student's own skeleton data. In this case, it is possible to prevent the fitting of the student's image displayed on the display unit 60 of the terminal device 3 used by the student and the student's skeleton data from being misaligned.
  • the scale factor when normalizing the skeletal data of the instructor may be determined based on the ratio of the height of the instructor to the height of the student, or the size of the image area in which the instructor is shown and the student are shown. It may be determined based on the ratio of the size of the image area.
  • the skeleton data of the standard body shape may be prepared in advance, and the normalization unit 15 may normalize the skeleton data of the instructor and the skeleton data of the student according to the skeleton data of the standard body shape.
  • the alignment unit 17 aligns the normalized skeleton data of the instructor with the skeleton data of the student. Specifically, the alignment unit 17 determines the reference position of the student's skeleton data based on the reference position information received from the instructor's terminal device 3, and sets the instructor's skeleton data reference position at the reference position. Place the leader's skeletal data to match.
  • the reference position of the skeleton data may be specified by image recognition.
  • the image recognition unit (not shown) of the processing unit 10 uses the classifier of each form held in the dictionary data holding unit 21 in the image area in which the instructor is reflected, and is used by the instructor. Detects the exercise form that is being performed. For the detection of the motion form, for example, a Haar-like feature amount, a HOG (Histogram of Gradients) feature amount, an LBP (Local Binary Patterns) feature amount, or the like can be used.
  • the image recognition unit identifies the reference position associated with the specified form. The same process is executed on the student terminal device 3 side.
  • the work of selecting the form can be omitted. Further, if the detection accuracy of each form by image recognition is high, the work of selecting an item can be omitted.
  • the deviation amount calculation unit 18 calculates the deviation amount between the instructor's skeleton data and the student's skeleton data for each body part. Based on the calculated deviation amount, the message generation unit 19 generates a message for encouraging the student to improve the form in order to bring the student's form closer to the instructor's form. Specifically, the deviation amount calculation unit 18 calculates the direction and distance in which the student should move each joint in order to make the deviation amount of each joint between the instructor's skeleton data and the student's skeleton data zero. .. The message generation unit 19 generates a message including the direction and distance in which the student should move each joint. For example, it generates messages such as "wrist 10 cm back" and "right knee 10 cm forward".
  • the notification control unit 110 controls to notify the student of the generated message by at least one of video, audio, and vibration.
  • the notification control unit 110 displays a message telop in the screen of the display unit 60.
  • the notification control unit 110 causes the voice output unit 70 to output a voice message.
  • the notification control unit 110 notifies the student of the magnitude and direction of the deviation from the instructor's form by the vibration interval or strength of the vibration device 80.
  • the vibration may be increased when the distance from the instructor's form is increased, and the vibration may be decreased when the distance from the instructor's form is increased.
  • the vibration interval may be shortened as the distance from the instructor's form increases, and the vibration interval may increase as the distance from the instructor's form increases.
  • FIGS. 4A-(c) show normalization of the received skeleton data of the instructor, an image showing the student Ps displayed on the display unit 60 of the terminal device 3 used by the student, and the student Ps. It is a figure which shows an example of the comparison between the skeleton data of the leader Pi, and the skeleton data Si of the instructor Pi.
  • FIGS. 3 (a)-(b) and 4 (a)-(c) show an example in which the instructor Pi is taller than the student Ps. Therefore, as shown in FIG. 4A, the normalization unit 15 of the student terminal device 3 reduces the skeleton data received from the instructor's terminal device 3 so as to match the size of the student skeleton data.
  • the display control unit 16 superimposes and displays the skeleton data Ss of the student Ps generated based on the image data of the area on the image area in which the student Ps is reflected in the image. do. Further, the display control unit 16 has the skeleton data of the instructor Pi normalized so that the reference point A of the skeleton data Ss of the student Ps and the reference point A of the skeleton data Sin of the normalized instructor Pi coincide with each other. Sin is superimposed and displayed on the screen. Further, the display control unit 16 superimposes message telops 64a-64d on the screen in order to encourage the student to improve the form according to the control by the notification control unit 110.
  • FIG. 5 is a flowchart showing the flow of form improvement support processing using the terminal device 3 according to the embodiment.
  • the user activates the terminal device 3 and sets and registers the user information from the operation unit 50 (S10).
  • the information to be set and registered is whether it is an instructor or a student, a name, an identification number, a height, a weight, and the like.
  • the instructor authority S11: Yes
  • the user sets the transmission destination of the skeleton data from the operation unit 50 (S12). Multiple destinations can be set.
  • the information to be set is the IP address and port number, the student's name, and so on.
  • the user sets the form of the item to be learned and the joint to be the reference position from the operation unit 50 (S13).
  • the processes of steps S12 and S13 are skipped.
  • the image pickup unit 40 actually photographs one's own figure, and a photography test is performed to see if the skeleton can be estimated normally (S14).
  • the actual shooting is started (S15).
  • the skeleton data generation unit 12 generates user skeleton data from the shot image (S17).
  • the display control unit 16 superimposes and displays the generated skeleton data on the image area of the image being photographed by the user (S18).
  • the skeleton data transmission / reception unit 13 transmits the skeleton data to the set destination terminal device 3 (S20). Transition to step S16.
  • the normalization unit 15 receives the instructor's skeleton data. , Normalize according to your height (S22).
  • the alignment unit 17 specifies a reference point for the skeleton data of the instructor (S23). Further, the alignment unit 17 specifies a reference point of its own skeleton data (S24).
  • the display control unit 16 superimposes and displays the skeleton data of the instructor so that the reference points match (S25).
  • the skeleton data generation unit 12 detects the orientation of the student's head from the image data in the image area in which the student is reflected (S26).
  • the deviation amount calculation unit 18 calculates the deviation amount of each joint position between the skeleton data of the instructor and the skeleton data of the student (S27).
  • the message generation unit 19 generates a message for encouraging the student to improve the form based on the amount of deviation of each joint position (S28).
  • the notification control unit 110 notifies the student of the generated message by at least one of video, audio, and vibration (S29).
  • the direction and distance (in cm) in which the joint should be moved are displayed in the vicinity of the joint.
  • the direction in which the joint should be moved is defined by the up, down, left, right, front and back with respect to the direction of the head. Transition to step S16.
  • the shooting is completed (S16: No)
  • the form improvement support process is completed.
  • a motion sensor and a dedicated data acquisition system are not required to acquire skeleton data. Since there is no need to attach sensors to the body, both teachers and students can teach or learn the correct form in a more natural way. Since the system can be configured with only a simple device using a camera, the cost of the entire system can be kept low.
  • the skeleton data of the instructor and the skeleton data of the student can be compared and fed back to the student in real time. This allows students to quickly master the correct form. By superimposing the skeletal data of the student and the instructor on the form image of the actual student, the difference between the actual body movement and the correct body movement can be easily and intuitively confirmed.
  • the reference position of the form to be compared can be changed according to the type of exercise. For example, if a baseball pitcher wants to focus on the shoulder-to-shoulder movement of the pitching form, by setting the starting point for comparison to the shoulder, he can concentrate on the shoulder movement without being disturbed by the movement of the entire body. You can learn the form.
  • the skeleton data of the instructor is transmitted to the student terminal device 3 in real time, and the skeleton data comparison processing and display processing are performed by the student terminal device 3.
  • the skeleton data comparison processing and display processing are performed by the student terminal device 3.
  • both a dynamic form with movement and a static form without movement can be learned.
  • the difference in posture between the student and the instructor can be fed back to the student in terms of the part and the quantitative value, the student can easily learn the posture.
  • Feedback to the student can be given via voice or vibration as well as screen display, so the student does not have to stare at the screen and can concentrate on learning posture.
  • the locus information of the motion sensor can be normalized and compared and displayed, but it is difficult for the student to actually imagine the movement of his / her body. Further, in Patent Document 1, it is possible to superimpose and display 3D graphics animation, but it is difficult to compare forms after fixing the position of the superimposing reference to a specific joint part. .. On the other hand, in the present embodiment, in addition to the normalization of the physique, the forms can be compared more accurately by selecting the reference position for comparing the forms according to the type of exercise. Further, in Patent Document 1, it is not considered to compare the form of one instructor with the form of a plurality of students, and it is difficult to realize it due to the system configuration. On the other hand, in the present embodiment, the form of one instructor and the form of a plurality of students can be easily compared.
  • the skeleton data of the instructor may be recorded as a moving image and the recorded data may be distributed to the student terminal device 3.
  • the student can play the instructor's video at his / her favorite time. At that time, it is also possible to specify a specific period of the video and play it.
  • the student may input the time, or the reproduction period may be designated by automatic detection. For example, the sound of kicking the ball may be automatically detected and the skeleton data for several seconds before and after that may be compared.
  • the playback period may be automatically set by automatically detecting a specific movement of the instructor (hand gooper, opening the mouth wide, etc.). If the form matches, you can also notify the student.
  • the playback speed of the instructor's video may be variable. In that case, for example, it is possible to adjust the form in detail by slow playback.
  • the instructor and the student use the same terminal device 3.
  • the instructor may use the terminal device 3 having high specifications.
  • the skeleton data generation process and the skeleton data comparison process may be performed on the cloud server.
  • the present invention can be used when instructing sports and fitness form improvement.
  • 1 form improvement support system, 2 network, 3 terminal device, 10 processing unit, 11 image data acquisition unit, 12 skeleton data generation unit, 13 skeleton data transmission / reception unit, 14 operation reception unit, 15 normalization unit, 16 display control unit, 17 alignment unit, 18 deviation amount calculation unit, 19 message generation unit, 110 notification control unit, 20 storage unit, 21 dictionary data retention unit, 30 communication unit, 40 image pickup unit, 50 operation unit, 60 display unit, 70 voice output. Department, 80 vibration device.

Abstract

An image data acquiring part 11 acquires, from an imaging part 40, data of an image in which a user is captured. A skeleton data generating part 12 generates, on the basis of the acquired data of the image, skeleton data of a person in which multiple characteristic points including joint positions in a person region in which the user is captured in the image are connected by a line. A skeleton data transmitting/receiving part 13 acquires skeleton data of an instructor generated by a terminal device 3 used by the instructor. A normalizing part 15 normalizes at least one set of skeleton data such that the size of the skeleton data of the instructor acquired by the skeleton data acquiring part corresponds to the size of the skeleton data of a user, who is being instructed by the instructor, generated by the skeleton data generating part 12. A display controlling part 16 performs control such that the skeleton data of the user and the skeleton data of the instructor having been normalized are overlapped on the image in which the user is captured by the imaging part 40, and the same is displayed on a display part 60.

Description

端末装置、及びフォーム改善支援方法Terminal device and form improvement support method
 本発明は、スポーツやフィットネスのフォームを改善するための端末装置、及びフォーム改善支援方法に関する。 The present invention relates to a terminal device for improving a form of sports or fitness, and a form improvement support method.
 野球、テニス、ゴルフなどのスポーツ、並びにヨガ、エアロビクスなどのフィットネスでは、正しいフォームを獲得することが重要である。正しいフォームを獲得することを目的として、被験者にモーションセンサを装着させ、被験者モデルの軌跡情報を生成し、リファレンスモデルの軌跡情報と比較させる方法が提案されている(例えば、特許文献1参照)。 In sports such as baseball, tennis and golf, as well as fitness such as yoga and aerobics, it is important to obtain the correct form. For the purpose of acquiring a correct form, a method has been proposed in which a subject is equipped with a motion sensor, locus information of a subject model is generated, and the locus information of a reference model is compared with the locus information (see, for example, Patent Document 1).
特開2012-120579号公報Japanese Unexamined Patent Publication No. 2012-12579
 上述した方法では、複数のモーションセンサと専用のデータ取得システムが必要となる。また上述した方法では、複数のモーションセンサから取得したフォームデータを解析した後に被験者にフィードバックする必要があり、フォームの練習中にリアルタイムにフィードバックすることが難しかった。 The above method requires multiple motion sensors and a dedicated data acquisition system. Further, in the above-mentioned method, it is necessary to give feedback to the subject after analyzing the form data acquired from a plurality of motion sensors, and it is difficult to give feedback in real time during the practice of the form.
 本実施形態はこうした状況に鑑みてなされたものであり、その目的は、シンプルなシステム構成で、リアルタイムにフォームの改善支援を行うことができる技術を提供することにある。 This embodiment was made in view of such a situation, and the purpose is to provide a technique capable of providing support for improving a form in real time with a simple system configuration.
 上記課題を解決するために、本実施形態のある態様の端末装置は、撮像部からユーザが映った画像データを取得する画像データ取得部と、取得された画像データをもとに、画像内の前記ユーザが映った人物領域の、関節位置を含む複数の特徴点を線で結んだ人物の骨格データを生成する骨格データ生成部と、指導者が使用する端末装置で生成された指導者の骨格データを取得する骨格データ取得部と、前記骨格データ取得部により取得された指導者の骨格データと、前記骨格データ生成部により生成された当該指導者に指導されているユーザの骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化する正規化部と、前記撮像部により撮像されているユーザが映った画像に、正規化された前記指導者の骨格データと前記ユーザの骨格データを重畳させて、表示部に表示させるよう制御する表示制御部と、を備える。 In order to solve the above-mentioned problems, the terminal device of an embodiment of the present embodiment has an image data acquisition unit that acquires image data of a user reflected from an imaging unit, and an image in an image based on the acquired image data. A skeletal data generation unit that generates skeletal data of a person connecting a plurality of feature points including joint positions in the person area in which the user is reflected, and a skeletal structure of the instructor generated by a terminal device used by the instructor. The size of the skeletal data acquisition unit for acquiring data, the skeletal data of the instructor acquired by the skeletal data acquisition unit, and the skeletal data of the user instructed by the instructor generated by the skeletal data generation unit Correspondingly, the normalization unit that normalizes at least one of the skeletal data, and the image of the user captured by the imaging unit are normalized to the skeletal data of the instructor and the skeletal data of the user. Is provided with a display control unit for controlling the display on the display unit by superimposing the data on the display unit.
 本実施形態の別の態様は、フォーム改善支援方法である。この方法は、撮像部からユーザが映った画像データを取得するステップと、取得された画像データをもとに、画像内の前記ユーザが映った人物領域の、関節位置を含む複数の特徴点を線で結んだ人物の骨格データを生成するステップと、指導者が使用する端末装置で生成された指導者の骨格データを取得するステップと、取得された指導者の骨格データと、生成された当該指導者に指導されているユーザの骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化するステップと、前記撮像部により撮像されているユーザが映った画像に、正規化された前記指導者の骨格データと前記ユーザの骨格データを重畳させて、表示部に表示させるよう制御するステップと、を有する。 Another aspect of this embodiment is a form improvement support method. In this method, a step of acquiring image data of a user imaged from an imaging unit and a plurality of feature points including joint positions of a person area in the image in which the user is imaged are obtained based on the acquired image data. The step of generating the skeleton data of the person connected by the line, the step of acquiring the skeleton data of the instructor generated by the terminal device used by the instructor, the acquired skeleton data of the instructor, and the generated relevant The step of normalizing at least one of the skeletal data and the image of the user captured by the image pickup unit were normalized so that the size of the skeletal data of the user being instructed by the instructor corresponds. It has a step of superimposing the skeleton data of the instructor and the skeleton data of the user and controlling the display unit to display the skeleton data.
 本実施形態のさらに別の態様は、フォーム改善支援プログラムである。このプログラムは、撮像部からユーザが映った画像データを取得する処理と、取得された画像データをもとに、画像内の前記ユーザが映った人物領域の、関節位置を含む複数の特徴点を線で結んだ人物の骨格データを生成する処理と、指導者が使用する端末装置で生成された指導者の骨格データを取得する処理と、取得された指導者の骨格データと、生成された当該指導者に指導されているユーザの骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化する処理と、前記撮像部により撮像されているユーザが映った画像に、正規化された前記指導者の骨格データと前記ユーザの骨格データを重畳させて、表示部に表示させるよう制御する処理と、をコンピュータに実行させる。 Yet another aspect of this embodiment is a form improvement support program. This program acquires image data of the user from the image pickup unit, and based on the acquired image data, a plurality of feature points including joint positions of the person area in which the user is reflected in the image. The process of generating the skeletal data of the person connected by the line, the process of acquiring the skeletal data of the instructor generated by the terminal device used by the instructor, the acquired skeletal data of the instructor, and the generated corresponding The process of normalizing at least one of the skeletal data and the image of the user captured by the image pickup unit are normalized so that the size of the skeletal data of the user instructed by the instructor corresponds to the size of the skeletal data. The computer is made to execute a process of superimposing the skeletal data of the instructor and the skeletal data of the user and controlling the display on the display unit.
 なお、以上の構成要素の任意の組合せ、本実施形態の表現を方法、装置、システム、記録媒体、コンピュータプログラムなどの間で変換したものもまた、本実施形態の態様として有効である。 It should be noted that any combination of the above components and the conversion of the expression of the present embodiment between methods, devices, systems, recording media, computer programs, etc. are also effective as aspects of the present embodiment.
 本実施形態によれば、シンプルなシステム構成で、リアルタイムにフォームの改善支援を行うことができる。 According to this embodiment, it is possible to support the improvement of the form in real time with a simple system configuration.
実施の形態に係るフォーム改善支援システムの構成を示す図である。It is a figure which shows the structure of the form improvement support system which concerns on embodiment. 指導者が使用している端末装置の表示部に表示される選択画面の一例を示す図である。It is a figure which shows an example of the selection screen displayed on the display part of the terminal apparatus used by an instructor. 図3(a)-(b)は、指導者が使用している端末装置の表示部に表示される指導者が映った画像と、指導者の骨格データの一例を示す図である。3 (a)-(b) are diagrams showing an image of the instructor displayed on the display unit of the terminal device used by the instructor and an example of the skeleton data of the instructor. 図4(a)-(c)は、受信した指導者の骨格データの正規化と、生徒が使用している端末装置の表示部に表示される生徒が映った画像と、生徒の骨格データと指導者の骨格データとの比較の一例を示す図である。FIGS. 4 (a)-(c) show the normalization of the received instructor's skeleton data, the image of the student displayed on the display of the terminal device used by the student, and the student's skeleton data. It is a figure which shows an example of comparison with the skeleton data of a leader. 実施の形態に係る端末装置を用いたフォーム改善支援処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the form improvement support process using the terminal apparatus which concerns on embodiment.
 図1は、実施の形態に係るフォーム改善支援システム1の構成を示す図である。フォーム改善支援システム1は複数の端末装置3を備え、複数の端末装置3はネットワーク2を介して接続される。複数の端末装置3の一つは指導者に使用され、その他は当該指導者から指導を受ける生徒により使用される。本明細書では、指導者とは特定の種目において正しい運動フォームを生徒に指導できる人物を指し、インストラクター、コーチ、学校の先生を含む。 FIG. 1 is a diagram showing a configuration of a form improvement support system 1 according to an embodiment. The form improvement support system 1 includes a plurality of terminal devices 3, and the plurality of terminal devices 3 are connected via the network 2. One of the plurality of terminal devices 3 is used by the instructor, and the other is used by the student who receives the instruction from the instructor. As used herein, an instructor is a person who can instruct a student in the correct exercise form in a particular discipline, including instructors, coaches, and school teachers.
 指導者と生徒が同じ場所(スタジオ等)にいる場合、複数の端末装置3はLAN(Local Area Network)により接続される。生徒と指導者が同じ場所にいる場合、生徒は指導者から直接指導を受けることができるが、生徒が大勢の場合、指導者の指導が行き届かない場合がある。また、後ろの方からは指導者の正しいフォームが見えにくい場合もある。端末装置3を使用することにより、生徒は自分のフォームと、手本とすべき指導者のフォームとの違いを、直感的かつ正確に把握できるようになり、効率的に正しいフォームを獲得することができる。 When the instructor and the student are in the same place (studio, etc.), the plurality of terminal devices 3 are connected by a LAN (Local Area Network). If the student and the instructor are in the same place, the student can receive direct instruction from the instructor, but if there are many students, the instructor's instruction may not be sufficient. Also, it may be difficult to see the correct form of the instructor from behind. By using the terminal device 3, students can intuitively and accurately grasp the difference between their own form and the form of the instructor who should be a model, and efficiently acquire the correct form. Can be done.
 指導者と生徒が地理的に離れた場所にいる場合、複数の端末装置3は、インターネットや専用線などの広域ネットワークにより接続される。この場合、生徒は自宅やオフィスに居ながらにして、遠隔で指導者の指導を受けることができる。また、本部にいる人気の指導者が遠隔で、サテライトスタジオにいる生徒を指導することもできる。 When the instructor and the student are geographically separated, the plurality of terminal devices 3 are connected by a wide area network such as the Internet or a dedicated line. In this case, the student can receive the guidance of the instructor remotely while staying at home or in the office. A popular instructor at the headquarters can also remotely instruct students in a satellite studio.
 図1は指導者と生徒が1:N(Nは2以上の整数)の例を想定した図であるが、指導者と生徒がN:Nの関係でもよい。その場合、複数の指導者は地理的に離れたそれぞれ別の場所にいてもよい。また、個人指導の場合、指導者と生徒が1:1の関係になる。 FIG. 1 is a diagram assuming an example in which the instructor and the student are 1: N (N is an integer of 2 or more), but the instructor and the student may have an N: N relationship. In that case, the leaders may be geographically separated from each other. In the case of individual instruction, the instructor and the student have a 1: 1 relationship.
 端末装置3は、運動フォームの改善を支援するための専用の端末装置であってもよいし、タブレット、ノートPC、スマートフォンなどの汎用の端末装置であってもよい。以下の説明では専用機を想定する。例えばスポーツクラブで利用される場合、端末装置3は、スポーツクラブから生徒に貸し出されてもよい。 The terminal device 3 may be a dedicated terminal device for supporting the improvement of the exercise form, or may be a general-purpose terminal device such as a tablet, a notebook PC, or a smartphone. In the following explanation, a dedicated machine is assumed. For example, when used in a sports club, the terminal device 3 may be rented out to the student from the sports club.
 端末装置3は、処理部10、記憶部20、通信部30、撮像部40、操作部50、表示部60、音声出力部70及び振動デバイス80を備える。処理部10は、画像データ取得部11、骨格データ生成部12、骨格データ送受信部13、操作受付部14、正規化部15、表示制御部16、位置合わせ部17、ずれ量算出部18、メッセージ生成部19及び通知制御部110を含む。 The terminal device 3 includes a processing unit 10, a storage unit 20, a communication unit 30, an imaging unit 40, an operation unit 50, a display unit 60, an audio output unit 70, and a vibration device 80. The processing unit 10 includes an image data acquisition unit 11, a skeleton data generation unit 12, a skeleton data transmission / reception unit 13, an operation reception unit 14, a normalization unit 15, a display control unit 16, an alignment unit 17, a deviation amount calculation unit 18, and a message. The generation unit 19 and the notification control unit 110 are included.
 処理部10の機能は、ハードウェア資源とソフトウェア資源の協働、又はハードウェア資源のみにより実現できる。ハードウェア資源として、CPU、GPU、ROM、RAM、DSP(Digital Signal Processor)、ISP(Image Signal Processor)、ASIC(Application Specific Integrated Circuit)、FPGA(Field Programmable Gate Array)、SoC(System on a Chip)、その他のLSIを利用できる。ソフトウェア資源としてオペレーティングシステム、ミドルウェア、アプリケーション等のプログラムを利用できる。 The function of the processing unit 10 can be realized only by the collaboration of hardware resources and software resources, or by hardware resources only. As hardware resources, CPU, GPU, ROM, RAM, DSP (Digital Signal Processor), ISP (Image Signal Processor), ASIC (Application Specific Integrated Circuit), FPGA (Field Programmable Gate Array), SoC (System on a Chip) , Other LSIs can be used. Programs such as operating systems, middleware, and applications can be used as software resources.
 記憶部20は、不揮発性の記録媒体(例えば、NAND型フラッシュメモリ、HDD(Hard Disk Drive))を有し、辞書データ保持部21を含む。記憶部20には、着脱自在なリムーバブル記録媒体(例えば、SDカード)が装着されてもよい。 The storage unit 20 has a non-volatile recording medium (for example, NAND flash memory, HDD (Hard Disk Drive)), and includes a dictionary data holding unit 21. A removable recording medium (for example, an SD card) may be attached to the storage unit 20.
 通信部30は、ネットワーク2に接続するための信号処理を行う。具体的には、TCP/IP(Transmission Control Protocol /Internet Protocol)やイーサネット(登録商標)などの通信プロトコルに準拠した信号処理を行う。通信部30は、ルータ(不図示)を介してネットワーク2に接続することができる。ルータと端末装置3間は、LANケーブまたはWi-Fi(登録商標)で接続することができる。なお、携帯電話網を利用してネットワーク2に接続してもよい。 The communication unit 30 performs signal processing for connecting to the network 2. Specifically, it performs signal processing based on communication protocols such as TCP / IP (Transmission Control Protocol / Internet Protocol) and Ethernet (registered trademark). The communication unit 30 can be connected to the network 2 via a router (not shown). The router and the terminal device 3 can be connected by LAN cable or Wi-Fi (registered trademark). The mobile phone network may be used to connect to the network 2.
 通信部30は、近距離に存在する機器(本実施の形態では振動デバイス80)と近距離無線通信するための信号処理も行う。近距離無線通信として、Bluetooth(登録商標)、Wi-Fi、赤外線通信などを使用することができる。 The communication unit 30 also performs signal processing for short-range wireless communication with a device existing at a short distance (vibration device 80 in this embodiment). As short-range wireless communication, Bluetooth (registered trademark), Wi-Fi, infrared communication, etc. can be used.
 撮像部40は、運動中の生徒または指導者(以下、両者を総称してユーザという)を撮影するためのカメラである。撮像部40は、ユーザから指定された距離、離れた位置に配置されることが望ましい。そのため、撮像部40と端末装置3の本体との間は、ケーブル(例えば、USBケーブル)または無線(例えば、Wi-Fi)で接続されてもよい。また撮像部40として、外付けのWebカメラを使用してもよい。 The imaging unit 40 is a camera for photographing a student or an instructor who is exercising (hereinafter, both are collectively referred to as a user). It is desirable that the image pickup unit 40 is arranged at a position designated by the user and away from the user. Therefore, the image pickup unit 40 and the main body of the terminal device 3 may be connected by a cable (for example, a USB cable) or wirelessly (for example, Wi-Fi). Further, an external Web camera may be used as the image pickup unit 40.
 撮像部40は、固体撮像素子及び信号処理回路を含む。固体撮像素子には例えば、CMOS(Complementary Metal Oxide Semiconductor)イメージセンサやCCD(Charge Coupled Devices)を使用することができる。固体撮像素子は入射光を電気的な画像信号に変換し、信号処理回路に出力する。信号処理回路は、固体撮像素子から入力される画像信号に対して、A/D変換、ノイズ除去などの信号処理を施し、デジタル値の画像データとして処理部10に出力する。 The image pickup unit 40 includes a solid-state image sensor and a signal processing circuit. For example, a CMOS (Complementary Metal Oxide Semiconductor) image sensor or a CCD (Charge Coupled Device) can be used as the solid-state image sensor. The solid-state image sensor converts the incident light into an electrical image signal and outputs it to a signal processing circuit. The signal processing circuit performs signal processing such as A / D conversion and noise removal on the image signal input from the solid-state image sensor, and outputs the image data as digital value to the processing unit 10.
 操作部50は、タッチパネルや物理ボタンを有し、ユーザの操作を受け付けて処理部10に出力する。表示部60は、液晶ディスプレイ、有機ELディスプレイ、マイクロLEDディスプレイ等のディスプレイを有し、処理部10から供給される画像や文字を表示する。本実施の形態では、タッチパネル機能と一体化されたタッチパネルディスプレイを使用することを想定する。 The operation unit 50 has a touch panel and physical buttons, receives a user's operation, and outputs the operation to the processing unit 10. The display unit 60 has a display such as a liquid crystal display, an organic EL display, and a micro LED display, and displays images and characters supplied from the processing unit 10. In this embodiment, it is assumed that a touch panel display integrated with the touch panel function is used.
 ユーザが運動中の自分の映像を見るためには、ディスプレイのサイズが大きいほうが好ましい。そこで端末装置3の本体に大画面の外付けディスプレイを接続してもよい。例えば、端末装置3の本体に、ケーブル(例えば、HDMI(登録商標)ケーブル、USBケーブル)または無線(例えば、Wi-Fi)を使用して、大画面のテレビやPCモニタを接続してもよい。 It is preferable that the size of the display is large so that the user can see his / her own image while exercising. Therefore, a large-screen external display may be connected to the main body of the terminal device 3. For example, a large screen TV or PC monitor may be connected to the main body of the terminal device 3 by using a cable (for example, HDMI (registered trademark) cable, USB cable) or wireless (for example, Wi-Fi). ..
 音声出力部70はスピーカを有し、処理部10から供給される音声を出力する。本実施の形態では、運動フォームを改善するための音声メッセージを出力する。なお、端末装置3の本体と無線(例えば、Bluetooth)で接続されたワイヤレスイヤホンを使用して、当該音声メッセージをユーザに通知してもよい。 The voice output unit 70 has a speaker and outputs the voice supplied from the processing unit 10. In this embodiment, a voice message for improving the exercise form is output. The voice message may be notified to the user by using a wireless earphone wirelessly (for example, Bluetooth) connected to the main body of the terminal device 3.
 振動デバイス80は、ユーザの腕や足(例えば、両手首と両足首)に装着されるバンド型の振動デバイスであり、処理部10から供給される信号に応じて振動する。端末装置3の本体と振動デバイス80との間は、無線(例えば、Bluetooth、Wi-Fi)で接続される。なお、振動デバイス80はオプションであり省略可能である。 The vibration device 80 is a band-type vibration device worn on a user's arm or foot (for example, both wrists and both ankles), and vibrates in response to a signal supplied from the processing unit 10. The main body of the terminal device 3 and the vibration device 80 are wirelessly (for example, Bluetooth, Wi-Fi) connected. The vibration device 80 is optional and can be omitted.
 画像データ取得部11は、撮像部40からユーザの全身が映った画像データを取得する。骨格データ生成部12は、取得された画像データをもとに、画像内のユーザが映った人物領域の、複数の関節位置を含む複数の特徴点(キーポイントともいう)を線で結んだ人物の骨格データを生成する。骨格データ生成部12は、既存の人物の姿勢推定モデル(例えば、OpenPose)を使用して骨格データを生成することができる。骨格データ生成部12は人物の骨格データを、人物画像から抽出した複数の特徴点と、特徴点間の関係性を規定した座標データで生成する。 The image data acquisition unit 11 acquires image data showing the entire body of the user from the image pickup unit 40. Based on the acquired image data, the skeleton data generation unit 12 is a person who connects a plurality of feature points (also referred to as key points) including a plurality of joint positions in a person area in which the user is shown in the image with a line. Generate skeleton data for. The skeleton data generation unit 12 can generate skeleton data using an existing posture estimation model of a person (for example, OpenPose). The skeleton data generation unit 12 generates skeleton data of a person with a plurality of feature points extracted from the person image and coordinate data defining the relationship between the feature points.
 操作受付部14は、ユーザの操作を受け付ける。操作受付部14は、各種の設定情報の入力や変更を受け付ける。例えば設定情報として、指導者モードか生徒モードかの選択や、通信先の設定情報(例えば、IPアドレス、ポート番号)を受け付ける。また、氏名、性別、年齢、各ユーザにユニークに割り当てられた識別番号、体型情報などの入力を受け付ける。体型情報として身長と体重の入力を受け付ける。さらに体型情報の補助情報として、肩幅、腰幅、腕の長さ、脚の長さなどの入力を受け付けてもよい。これにより、画像内に映った人物の各部位の長さや幅を高精度に推定することができる。 The operation reception unit 14 accepts the user's operation. The operation reception unit 14 receives input and change of various setting information. For example, as setting information, selection of instructor mode or student mode and setting information of a communication destination (for example, IP address, port number) are accepted. It also accepts inputs such as name, gender, age, identification number uniquely assigned to each user, and body shape information. Accepts height and weight inputs as body shape information. Further, as auxiliary information of body shape information, input such as shoulder width, waist width, arm length, and leg length may be accepted. This makes it possible to estimate the length and width of each part of the person reflected in the image with high accuracy.
 操作受付部14は、指導者の操作にもとづき、運動種目の選択、選択した種目において獲得すべきフォームの選択、基準位置とすべき関節位置の選択を受け付ける。 The operation reception unit 14 accepts the selection of the exercise item, the selection of the form to be acquired in the selected item, and the selection of the joint position to be the reference position based on the operation of the instructor.
 図2は、指導者が使用している端末装置3の表示部60に表示される選択画面の一例を示す図である。指導者は生徒への指導に先立ち、または生徒への指導中に、種目選択欄61から種目を選択し、フォーム選択欄62からフォームを選択し、基準位置選択欄63から関節を選択する。 FIG. 2 is a diagram showing an example of a selection screen displayed on the display unit 60 of the terminal device 3 used by the instructor. Prior to or during instruction to the student, the instructor selects an item from the item selection field 61, selects a form from the form selection field 62, and selects a joint from the reference position selection field 63.
 フォーム選択欄62には、種目選択欄61で選択された種目の複数のフォームが選択候補として表示される。図2に示す例では、種目選択欄61でヨガが選択されているため、フォーム選択欄62には、猫のポーズ、牛の顔のポーズ、・・・、板のポーズ、戦士のポーズA、戦士のポーズB、・・・が選択候補として表示されている。図2に示す例では、フォーム選択欄62で戦士のポーズAが選択されている。指導者は、ポーズが変わるたびに新たなポーズを選択する。 In the form selection field 62, a plurality of forms of the item selected in the item selection field 61 are displayed as selection candidates. In the example shown in FIG. 2, since yoga is selected in the item selection field 61, the form selection field 62 includes a cat pose, a cow face pose, ..., a board pose, a warrior pose A, and so on. Warrior pose B, ... is displayed as a selection candidate. In the example shown in FIG. 2, the warrior pose A is selected in the form selection field 62. The instructor selects a new pose each time the pose changes.
 なお、種目選択欄61において例えば、野球が選択された場合、複数パターンのピッチングフォーム、複数パターンのバッティングフォーム、複数パターンのキャッチングフォーム、・・・がフォーム選択欄62に選択候補として表示される。 For example, when baseball is selected in the item selection field 61, a plurality of patterns of pitching form, a plurality of patterns of batting form, a plurality of patterns of catching form, ... Are displayed as selection candidates in the form selection field 62.
 指導者は、基準位置選択欄63から、フォーム選択欄62で選択されたフォームの基準位置とすべき関節を選択する。図2に示す例では、股関節が選択されている。選択された基準位置は、指導者の骨格データと生徒の骨格データを照合する際の基準点として使用される。 The instructor selects a joint to be the reference position of the form selected in the form selection field 62 from the reference position selection field 63. In the example shown in FIG. 2, the hip joint is selected. The selected reference position is used as a reference point when collating the skeletal data of the instructor with the skeletal data of the student.
 各フォームの基準位置とすべき関節として、静止したフォーム(ポーズ)の場合、体の重心に一番近い関節が選択されてもよい。動きのあるフォームの場合、動いている部位の支点となる関節が選択されてもよい。例えば、野球のピッチングフォームの場合、肩関節が選択されてもよい。なお指導者は、体の重心や動きの支点に関係なく、骨格データを照合した場合の見やすさなどを考慮し、最適と判断した関節を選択することができる。 In the case of a stationary form (pose), the joint closest to the center of gravity of the body may be selected as the joint that should be the reference position for each form. In the case of a moving form, a joint that serves as a fulcrum of the moving part may be selected. For example, in the case of a baseball pitching form, the shoulder joint may be selected. The instructor can select the joint that is judged to be optimal, regardless of the center of gravity of the body or the fulcrum of movement, in consideration of the visibility when collating the skeletal data.
 なお、各フォームの基準位置とすべき関節は、予め設定されていてもよい。この場合、指導者によりフォーム選択欄62からフォームが選択されると、選択されたフォームに紐付けられている関節が自動的に決定される。この場合、指導者は基準位置とすべき関節を選択する必要がない。 The joint that should be the reference position for each form may be preset. In this case, when a form is selected from the form selection field 62 by the instructor, the joint associated with the selected form is automatically determined. In this case, the instructor does not have to select the joint that should be the reference position.
 記憶部20内の辞書データ保持部21は、各種目のフォームごとに、指導者の姿勢や動きが映った多数の画像に基づく多数の骨格データを機械学習して生成した識別器を辞書データとして保持する。各種目の各フォームの辞書データには、当該フォームの基準位置とすべき関節位置が予め登録されていてもよい。この場合、上述した基準位置の自動決定が可能となる。 The dictionary data holding unit 21 in the storage unit 20 uses a classifier generated by machine learning a large number of skeletal data based on a large number of images showing the posture and movement of the instructor as dictionary data for each form of each item. Hold. In the dictionary data of each form of each item, the joint position to be the reference position of the form may be registered in advance. In this case, the above-mentioned reference position can be automatically determined.
 骨格データ送受信部13は、指導者モードの場合、骨格データ生成部12により生成された指導者の骨格データと基準位置情報を、ネットワーク2を介して生徒の端末装置3に送信する。骨格データ送受信部13は、生徒モードの場合、指導者の端末装置3からネットワーク2を介して送信されてくる指導者の骨格データと基準位置情報を受信する。 In the case of the instructor mode, the skeleton data transmission / reception unit 13 transmits the skeleton data and the reference position information of the instructor generated by the skeleton data generation unit 12 to the student terminal device 3 via the network 2. In the student mode, the skeleton data transmission / reception unit 13 receives the skeleton data and the reference position information of the instructor transmitted from the instructor's terminal device 3 via the network 2.
 表示制御部16は、撮像部40により撮像されているユーザが映った画像に、当該ユーザの骨格データを重畳させて表示部60に表示させる。 The display control unit 16 superimposes the skeleton data of the user on the image of the user captured by the image pickup unit 40 and displays it on the display unit 60.
 図3(a)-(b)は、指導者が使用している端末装置3の表示部60に表示される指導者Piが映った画像と、指導者Piの骨格データSiの一例を示す図である。図3(a)に示すように表示制御部16は、画像内の指導者Piが映った画像領域に、当該領域の画像データをもとに生成された骨格データSiを重畳させて表示する。骨格データSiは、画像内の指導者Piの動きに追従してリアルタイムに変化する。 3 (a)-(b) are diagrams showing an image showing the instructor Pi displayed on the display unit 60 of the terminal device 3 used by the instructor, and an example of the skeleton data Si of the instructor Pi. Is. As shown in FIG. 3A, the display control unit 16 superimposes and displays the skeleton data Si generated based on the image data of the area on the image area in which the instructor Pi is reflected in the image. The skeleton data Si changes in real time following the movement of the instructor Pi in the image.
 なお、上述した指導者による基準位置とすべき関節の選択は、表示部60と操作部50を一体的に構成しているタッチパネルディスプレイに表示されている骨格データSiの特定の特徴点にタッチすることにより行ってもよい。例えば、図3のA点をタッチすることにより、股関節を選択することができる。 The above-mentioned instructor selects a joint to be a reference position by touching a specific feature point of the skeleton data Si displayed on the touch panel display that integrally constitutes the display unit 60 and the operation unit 50. It may be done by. For example, the hip joint can be selected by touching the point A in FIG.
 図1に戻る。以下に説明する処理部10の処理は、生徒モードのときに実行され、指導者モードでは実行されない処理である。正規化部15は、骨格データ送受信部13により受信された指導者の骨格データと、骨格データ生成部12により生成された生徒自身の骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化する。本実施の形態では、正規化部15は、受信された指導者の骨格データを生徒自身の骨格データのサイズに合わせるように、指導者の骨格データを正規化する。この場合、生徒が使用している端末装置3の表示部60に表示される生徒の画像と、生徒の骨格データのフィッティングがずれることを防止することができる。 Return to Fig. 1. The process of the processing unit 10 described below is a process that is executed in the student mode and not executed in the instructor mode. The normalization unit 15 generates at least one skeleton data so that the size of the skeleton data of the instructor received by the skeleton data transmission / reception unit 13 and the size of the student's own skeleton data generated by the skeleton data generation unit 12 correspond to each other. Normalize. In this embodiment, the normalization unit 15 normalizes the instructor's skeleton data so that the received instructor's skeleton data fits the size of the student's own skeleton data. In this case, it is possible to prevent the fitting of the student's image displayed on the display unit 60 of the terminal device 3 used by the student and the student's skeleton data from being misaligned.
 指導者の骨格データを正規化する際の縮尺率は、指導者の身長と生徒の身長との比率をもとに決定してもよいし、指導者が映った画像領域のサイズと生徒が映った画像領域のサイズの比率をもとに決定してもよい。なお、標準体型の骨格データを予め準備し、正規化部15は、指導者の骨格データと生徒の骨格データを標準体型の骨格データにそれぞれ合わせて正規化してもよい。 The scale factor when normalizing the skeletal data of the instructor may be determined based on the ratio of the height of the instructor to the height of the student, or the size of the image area in which the instructor is shown and the student are shown. It may be determined based on the ratio of the size of the image area. The skeleton data of the standard body shape may be prepared in advance, and the normalization unit 15 may normalize the skeleton data of the instructor and the skeleton data of the student according to the skeleton data of the standard body shape.
 位置合わせ部17は、正規化された指導者の骨格データと生徒の骨格データの位置合わせを行う。具体的には位置合わせ部17は、指導者の端末装置3から受信した基準位置情報をもとに生徒の骨格データの基準位置を決定し、当該基準位置に指導者の骨格データの基準位置を合わせるように、指導者の骨格データを配置する。 The alignment unit 17 aligns the normalized skeleton data of the instructor with the skeleton data of the student. Specifically, the alignment unit 17 determines the reference position of the student's skeleton data based on the reference position information received from the instructor's terminal device 3, and sets the instructor's skeleton data reference position at the reference position. Place the leader's skeletal data to match.
 なお、骨格データの基準位置は画像認識により特定されてもよい。指導者の端末装置3において処理部10の画像認識部(不図示)は、指導者が映った画像領域において、辞書データ保持部21に保持されている各フォームの識別器を用いて、指導者が行っている運動フォームを検出する。運動フォームの検出には例えば、Haar-like特徴量、HOG(Histogram of Gradients)特徴量、LBP(Local Binary Patterns)特徴量などを用いることができる。画像認識部は、特定したフォームに紐付いている基準位置を特定する。生徒の端末装置3側でも同様の処理を実行する。 The reference position of the skeleton data may be specified by image recognition. In the terminal device 3 of the instructor, the image recognition unit (not shown) of the processing unit 10 uses the classifier of each form held in the dictionary data holding unit 21 in the image area in which the instructor is reflected, and is used by the instructor. Detects the exercise form that is being performed. For the detection of the motion form, for example, a Haar-like feature amount, a HOG (Histogram of Gradients) feature amount, an LBP (Local Binary Patterns) feature amount, or the like can be used. The image recognition unit identifies the reference position associated with the specified form. The same process is executed on the student terminal device 3 side.
 この場合、指導者の端末装置3から生徒の端末装置3に基準位置情報を送信する必要がない。また、指導者が操作受付部14から、基準位置とすべき関節を選択する作業に加えて、フォームを選択する作業も省略することができる。さらに、画像認識による各フォームの検出精度が高ければ、種目を選択する作業も省略することができる。 In this case, it is not necessary to transmit the reference position information from the instructor's terminal device 3 to the student's terminal device 3. Further, in addition to the work of the instructor selecting the joint to be the reference position from the operation reception unit 14, the work of selecting the form can be omitted. Further, if the detection accuracy of each form by image recognition is high, the work of selecting an item can be omitted.
 ずれ量算出部18は体の部位ごとに、指導者の骨格データと生徒の骨格データのずれ量を算出する。メッセージ生成部19は、算出されたずれ量をもとに、生徒のフォームを指導者のフォームに近づけさせるための、生徒にフォームの改善を促すためのメッセージを生成する。具体的には、ずれ量算出部18は、指導者の骨格データと生徒の骨格データとの各関節のずれ量をゼロにするための、生徒が各関節を移動させるべき方向と距離を算出する。メッセージ生成部19は、生徒が各関節を移動させるべき方向と距離を含むメッセージを生成する。例えば、「手首を10cm後ろへ」、「右膝を10cm前へ」といったメッセージを生成する。 The deviation amount calculation unit 18 calculates the deviation amount between the instructor's skeleton data and the student's skeleton data for each body part. Based on the calculated deviation amount, the message generation unit 19 generates a message for encouraging the student to improve the form in order to bring the student's form closer to the instructor's form. Specifically, the deviation amount calculation unit 18 calculates the direction and distance in which the student should move each joint in order to make the deviation amount of each joint between the instructor's skeleton data and the student's skeleton data zero. .. The message generation unit 19 generates a message including the direction and distance in which the student should move each joint. For example, it generates messages such as "wrist 10 cm back" and "right knee 10 cm forward".
 通知制御部110は、生成されたメッセージを、映像、音声、振動の少なくとも一つで生徒に通知するよう制御する。映像で通知する場合、通知制御部110は、表示部60の画面内にメッセージテロップを表示させる。音声で通知する場合、通知制御部110は、音声出力部70に音声メッセージを出力させる。振動で通知する場合、通知制御部110は、振動デバイス80の振動の間隔または強弱で、指導者のフォームとのずれの大きさと方向を生徒に通知する。例えば、指導者のフォームとの距離が離れると振動を強くし、近づくと振動を弱くしてもよい。また、指導者のフォームとの距離が離れると振動の間隔を短くし、近づくと振動の間隔を長くしてもよい。 The notification control unit 110 controls to notify the student of the generated message by at least one of video, audio, and vibration. When notifying by video, the notification control unit 110 displays a message telop in the screen of the display unit 60. When notifying by voice, the notification control unit 110 causes the voice output unit 70 to output a voice message. When notifying by vibration, the notification control unit 110 notifies the student of the magnitude and direction of the deviation from the instructor's form by the vibration interval or strength of the vibration device 80. For example, the vibration may be increased when the distance from the instructor's form is increased, and the vibration may be decreased when the distance from the instructor's form is increased. In addition, the vibration interval may be shortened as the distance from the instructor's form increases, and the vibration interval may increase as the distance from the instructor's form increases.
 図4(a)-(c)は、受信した指導者の骨格データの正規化と、生徒が使用している端末装置3の表示部60に表示される生徒Psが映った画像と、生徒Psの骨格データと指導者Piの骨格データSiとの比較の一例を示す図である。 FIGS. 4A-(c) show normalization of the received skeleton data of the instructor, an image showing the student Ps displayed on the display unit 60 of the terminal device 3 used by the student, and the student Ps. It is a figure which shows an example of the comparison between the skeleton data of the leader Pi, and the skeleton data Si of the instructor Pi.
 図3(a)-(b)及び図4(a)-(c)では、指導者Piのほうが生徒Psより身長が高い例を示している。そこで図4(a)に示すように、生徒の端末装置3の正規化部15は、指導者の端末装置3から受信した骨格データを、生徒の骨格データのサイズに合わせるように縮小する。 FIGS. 3 (a)-(b) and 4 (a)-(c) show an example in which the instructor Pi is taller than the student Ps. Therefore, as shown in FIG. 4A, the normalization unit 15 of the student terminal device 3 reduces the skeleton data received from the instructor's terminal device 3 so as to match the size of the student skeleton data.
 図4(b)に示すように表示制御部16は、画像内の生徒Psが映った画像領域に、当該領域の画像データをもとに生成された生徒Psの骨格データSsを重畳させて表示する。さらに表示制御部16は、生徒Psの骨格データSsの基準点Aと、正規化された指導者Piの骨格データSinの基準点Aが一致するように、正規化された指導者Piの骨格データSinを重畳させて画面に表示する。さらに表示制御部16は、通知制御部110による制御に従い、生徒にフォームの改善を促すためのメッセージテロップ64a-64dを画面に重畳する。 As shown in FIG. 4B, the display control unit 16 superimposes and displays the skeleton data Ss of the student Ps generated based on the image data of the area on the image area in which the student Ps is reflected in the image. do. Further, the display control unit 16 has the skeleton data of the instructor Pi normalized so that the reference point A of the skeleton data Ss of the student Ps and the reference point A of the skeleton data Sin of the normalized instructor Pi coincide with each other. Sin is superimposed and displayed on the screen. Further, the display control unit 16 superimposes message telops 64a-64d on the screen in order to encourage the student to improve the form according to the control by the notification control unit 110.
 図5は、実施の形態に係る端末装置3を用いたフォーム改善支援処理の流れを示すフローチャートである。ユーザが端末装置3を起動して、操作部50からユーザ情報を設定登録する(S10)。設定登録する情報は、指導者か生徒か、氏名、識別番号、身長、体重などである。指導者権限で登録した場合(S11:Yes)、ユーザは操作部50から、骨格データの送信先を設定する(S12)。送信先は複数設定することが可能である。設定する情報は、IPアドレスとポート番号、生徒の氏名などである。ユーザは操作部50から、学習する種目のフォームと基準位置とすべき関節を設定する(S13)。ステップS11において生徒権限で登録した場合(S11:No)、ステップS12、S13の処理はスキップされる。 FIG. 5 is a flowchart showing the flow of form improvement support processing using the terminal device 3 according to the embodiment. The user activates the terminal device 3 and sets and registers the user information from the operation unit 50 (S10). The information to be set and registered is whether it is an instructor or a student, a name, an identification number, a height, a weight, and the like. When registered with the instructor authority (S11: Yes), the user sets the transmission destination of the skeleton data from the operation unit 50 (S12). Multiple destinations can be set. The information to be set is the IP address and port number, the student's name, and so on. The user sets the form of the item to be learned and the joint to be the reference position from the operation unit 50 (S13). When registering with student authority in step S11 (S11: No), the processes of steps S12 and S13 are skipped.
 次に実際に自分の姿を撮像部40で撮影して、正常に骨格推定が可能か否か撮影テストを行う(S14)。テストが完了したら実際の撮影を開始する(S15)。撮影中(S16:Yes)、骨格データ生成部12は、撮影された画像からユーザの骨格データを生成する(S17)。表示制御部16は、生成された骨格データを、撮影中の画像の当該ユーザが映った画像領域に重畳させて表示する(S18)。 Next, the image pickup unit 40 actually photographs one's own figure, and a photography test is performed to see if the skeleton can be estimated normally (S14). When the test is completed, the actual shooting is started (S15). During shooting (S16: Yes), the skeleton data generation unit 12 generates user skeleton data from the shot image (S17). The display control unit 16 superimposes and displays the generated skeleton data on the image area of the image being photographed by the user (S18).
 指導者権限でログインしている場合(S19:Yes)、骨格データ送受信部13は骨格データを、設定された送信先の端末装置3に送信する(S20)。ステップS16に遷移する。 When logged in with the instructor authority (S19: Yes), the skeleton data transmission / reception unit 13 transmits the skeleton data to the set destination terminal device 3 (S20). Transition to step S16.
 生徒権限でログインしている場合において(S19:No)、指導者の端末装置3から指導者の骨格データを受信すると(S21:Yes)、正規化部15は、受信した指導者の骨格データを、自分の身長に合わせて正規化する(S22)。位置合わせ部17は、指導者の骨格データの基準点を特定する(S23)。また位置合わせ部17は、自分の骨格データの基準点を特定する(S24)。表示制御部16は、基準点が一致するように、指導者の骨格データを重畳させて表示する(S25)。 When the instructor's skeleton data is received from the instructor's terminal device 3 (S21: Yes) when logged in with student authority (S19: No), the normalization unit 15 receives the instructor's skeleton data. , Normalize according to your height (S22). The alignment unit 17 specifies a reference point for the skeleton data of the instructor (S23). Further, the alignment unit 17 specifies a reference point of its own skeleton data (S24). The display control unit 16 superimposes and displays the skeleton data of the instructor so that the reference points match (S25).
 骨格データ生成部12は、生徒が映った画像領域の画像データから生徒の頭の向きを検出する(S26)。ずれ量算出部18は、指導者の骨格データと生徒の骨格データの各関節位置のずれ量を算出する(S27)。メッセージ生成部19は、各関節位置のずれ量をもとに、生徒にフォームの改善を促すためのメッセージを生成する(S28)。通知制御部110は、生成されたメッセージを映像、音声、振動の少なくとも一つで生徒に通知する(S29)。画面に表示する場合は、関節の近傍に関節を動かすべき方向と距離(cm単位)を表示する。関節を動かすべき方向は、頭の向きを基準とした上下左右前後で規定される。ステップS16に遷移する。撮影が終了すると(S16:No)、フォーム改善支援処理を終了する。 The skeleton data generation unit 12 detects the orientation of the student's head from the image data in the image area in which the student is reflected (S26). The deviation amount calculation unit 18 calculates the deviation amount of each joint position between the skeleton data of the instructor and the skeleton data of the student (S27). The message generation unit 19 generates a message for encouraging the student to improve the form based on the amount of deviation of each joint position (S28). The notification control unit 110 notifies the student of the generated message by at least one of video, audio, and vibration (S29). When displaying on the screen, the direction and distance (in cm) in which the joint should be moved are displayed in the vicinity of the joint. The direction in which the joint should be moved is defined by the up, down, left, right, front and back with respect to the direction of the head. Transition to step S16. When the shooting is completed (S16: No), the form improvement support process is completed.
 以上説明したように本実施の形態によれば、骨格データの取得にモーションセンサと専用のデータ取得システムが不要となる。体にセンサ類を取り付ける必要がないため、指導者も生徒も、より自然な状態で正しいフォームを指導または学習できる。カメラを使ったシンプルな装置のみでシステムを構成可能であるため、システム全体のコストを安価に抑えることができる。 As described above, according to this embodiment, a motion sensor and a dedicated data acquisition system are not required to acquire skeleton data. Since there is no need to attach sensors to the body, both teachers and students can teach or learn the correct form in a more natural way. Since the system can be configured with only a simple device using a camera, the cost of the entire system can be kept low.
 また、フォームの練習中に、指導者の骨格データと生徒の骨格データを比較して、リアルタイムに生徒にフィードバックすることができる。これにより、生徒は正しいフォームを迅速に習得することができる。実際の生徒のフォーム映像に、生徒と指導者の骨格データを重畳させることによって、実際の身体の動きと正しい身体の動きとの違いを容易かつ直感的に確認することができる。 Also, while practicing the form, the skeleton data of the instructor and the skeleton data of the student can be compared and fed back to the student in real time. This allows students to quickly master the correct form. By superimposing the skeletal data of the student and the instructor on the form image of the actual student, the difference between the actual body movement and the correct body movement can be easily and intuitively confirmed.
 また、本実施の形態に係る端末装置3では、比較するフォームの基準位置を運動の種類に応じて変更することができる。例えば野球のピッチャーが投球フォームの肩から先の動きを重点的に学習したい場合、比較する起点を肩に設定することによって、他の身体全体の動きに邪魔されずに肩の動きのみに集中してフォームを学習することができる。 Further, in the terminal device 3 according to the present embodiment, the reference position of the form to be compared can be changed according to the type of exercise. For example, if a baseball pitcher wants to focus on the shoulder-to-shoulder movement of the pitching form, by setting the starting point for comparison to the shoulder, he can concentrate on the shoulder movement without being disturbed by the movement of the entire body. You can learn the form.
 また、本実施の形態に係るフォーム改善支援システム1では、指導者の骨格データをリアルタイムに生徒の端末装置3に送信し、生徒の端末装置3で骨格データの比較処理と表示処理を行う。これにより、指導者の端末装置3に負荷が集中することなく、スムーズに複数の生徒が同時にフォームの学習をすることができる。 Further, in the form improvement support system 1 according to the present embodiment, the skeleton data of the instructor is transmitted to the student terminal device 3 in real time, and the skeleton data comparison processing and display processing are performed by the student terminal device 3. As a result, a plurality of students can smoothly learn the form at the same time without concentrating the load on the instructor's terminal device 3.
 また、本実施の形態に係る端末装置3では、動きのある動的なフォームも動きのない静的なフォームも両方学習することができる。また、生徒と指導者の姿勢のずれを、部位と定量的な値で生徒にフィードバックできるため、生徒は姿勢の学習を容易に行うことができる。生徒へのフィードバックは画面表示の他、音声または振動を介しても行うことができるため、生徒は画面を凝視する必要がなく、姿勢の学習に集中することができる。また、本実施の形態に係る端末装置3を使用することにより、同一場所もしくは遠隔地の生徒をリアルタイムに指導することが可能である。 Further, in the terminal device 3 according to the present embodiment, both a dynamic form with movement and a static form without movement can be learned. In addition, since the difference in posture between the student and the instructor can be fed back to the student in terms of the part and the quantitative value, the student can easily learn the posture. Feedback to the student can be given via voice or vibration as well as screen display, so the student does not have to stare at the screen and can concentrate on learning posture. Further, by using the terminal device 3 according to the present embodiment, it is possible to instruct students in the same place or a remote place in real time.
 上記特許文献1では、モーションセンサの軌跡情報を正規化して比較表示することができるが、生徒が自分の身体の動きを実際にイメージすることが困難であった。また上記特許文献1では、3Dグラフィックスのアニメーションを重ね合わせて表示することができるが、重ね合わせる基準の位置を、特定の関節部位に固定した上でフォームの比較を行うことは困難であった。これに対して本実施の形態では、体格の正規化に加えて、フォームを比較する基準位置を運動の種類に応じて選択することにより、より正確にフォームを比較することができる。また上記特許文献1では、指導者一人のフォームと、複数の生徒のフォームを比較することは考慮されておらず、システム構成上、それを実現することは困難であった。これに対して本実施の形態では、指導者一人のフォームと、複数の生徒のフォームを容易に比較することができる。 In Patent Document 1, the locus information of the motion sensor can be normalized and compared and displayed, but it is difficult for the student to actually imagine the movement of his / her body. Further, in Patent Document 1, it is possible to superimpose and display 3D graphics animation, but it is difficult to compare forms after fixing the position of the superimposing reference to a specific joint part. .. On the other hand, in the present embodiment, in addition to the normalization of the physique, the forms can be compared more accurately by selecting the reference position for comparing the forms according to the type of exercise. Further, in Patent Document 1, it is not considered to compare the form of one instructor with the form of a plurality of students, and it is difficult to realize it due to the system configuration. On the other hand, in the present embodiment, the form of one instructor and the form of a plurality of students can be easily compared.
 以上、本発明を実施の形態をもとに説明した。この実施の形態は例示であり、それらの各構成要素や各処理プロセスの組合せにいろいろな変形例が可能なこと、またそうした変形例も本発明の範囲にあることは当業者に理解されるところである。 The present invention has been described above based on the embodiments. It is understood by those skilled in the art that this embodiment is an example, and that various modifications are possible for each of these components and combinations of each processing process, and that such modifications are also within the scope of the present invention. be.
 上述した実施の形態では、指導者のフォームをリアルタイムに学習する例を説明した。この点、指導者の骨格データを動画で録画しておき、録画データを生徒の端末装置3に配信してもよい。この場合、生徒は自分の好きな時間に指導者の動画を再生することができる。その際、動画の特定の期間を指定して再生することもできる。指定の方法としては生徒が時間を入力してもよいし、自動検知により再生期間が指定されてもよい。例えば、ボールをける音を自動検知して、その前後数秒間の骨格データを比較してもよい。また、指導者の特定の動き(手のグーパー、口を大きく開ける等)を自動検知して自動的に再生期間が設定されてもよい。フォームが合致した場合は、その旨を生徒に通知することもできる。また、指導者の動画の再生速度を可変できるようにしてもよい。その場合、例えば、スロー再生で詳細にフォーム合わせをすることも可能である。 In the above-described embodiment, an example of learning the instructor's form in real time was described. In this regard, the skeleton data of the instructor may be recorded as a moving image and the recorded data may be distributed to the student terminal device 3. In this case, the student can play the instructor's video at his / her favorite time. At that time, it is also possible to specify a specific period of the video and play it. As a method of designation, the student may input the time, or the reproduction period may be designated by automatic detection. For example, the sound of kicking the ball may be automatically detected and the skeleton data for several seconds before and after that may be compared. In addition, the playback period may be automatically set by automatically detecting a specific movement of the instructor (hand gooper, opening the mouth wide, etc.). If the form matches, you can also notify the student. In addition, the playback speed of the instructor's video may be variable. In that case, for example, it is possible to adjust the form in detail by slow playback.
 上述した実施の形態では、指導者と生徒が同じ端末装置3を使用する例を想定した。この点、指導者が高スペックな端末装置3を使用してもよい。その場合、指導者の端末装置3で集中的に骨格データの比較処理を行うことも可能である。また、骨格データの生成処理と骨格データの比較処理をクラウドサーバ上で行ってもよい。 In the above-described embodiment, an example is assumed in which the instructor and the student use the same terminal device 3. In this respect, the instructor may use the terminal device 3 having high specifications. In that case, it is also possible to intensively compare the skeleton data with the instructor's terminal device 3. Further, the skeleton data generation process and the skeleton data comparison process may be performed on the cloud server.
 本発明は、スポーツやフィットネスのフォーム改善の指導をする際に利用可能である。 The present invention can be used when instructing sports and fitness form improvement.
 1 フォーム改善支援システム、 2 ネットワーク、 3 端末装置、 10 処理部、 11 画像データ取得部、 12 骨格データ生成部、 13 骨格データ送受信部、 14 操作受付部、 15 正規化部、 16 表示制御部、 17 位置合わせ部、 18 ずれ量算出部、 19 メッセージ生成部、 110 通知制御部、 20 記憶部、 21 辞書データ保持部、 30 通信部、 40 撮像部、 50 操作部、 60 表示部、 70 音声出力部、 80 振動デバイス。 1 form improvement support system, 2 network, 3 terminal device, 10 processing unit, 11 image data acquisition unit, 12 skeleton data generation unit, 13 skeleton data transmission / reception unit, 14 operation reception unit, 15 normalization unit, 16 display control unit, 17 alignment unit, 18 deviation amount calculation unit, 19 message generation unit, 110 notification control unit, 20 storage unit, 21 dictionary data retention unit, 30 communication unit, 40 image pickup unit, 50 operation unit, 60 display unit, 70 voice output. Department, 80 vibration device.

Claims (5)

  1.  撮像部からユーザが映った画像データを取得する画像データ取得部と、
     取得された画像データをもとに、画像内の前記ユーザが映った人物領域の、関節位置を含む複数の特徴点を線で結んだ人物の骨格データを生成する骨格データ生成部と、
     指導者が使用する端末装置で生成された指導者の骨格データを取得する骨格データ取得部と、
     前記骨格データ取得部により取得された指導者の骨格データと、前記骨格データ生成部により生成された当該指導者に指導されているユーザの骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化する正規化部と、
     前記撮像部により撮像されているユーザが映った画像に、正規化された前記指導者の骨格データと前記ユーザの骨格データを重畳させて、表示部に表示させるよう制御する表示制御部と、
     を備える、端末装置。
    An image data acquisition unit that acquires image data of the user from the image pickup unit, and an image data acquisition unit.
    Based on the acquired image data, a skeleton data generation unit that generates skeletal data of a person who connects a plurality of feature points including joint positions in the person area in which the user is shown in the image with a line, and a skeleton data generation unit.
    The skeleton data acquisition unit that acquires the skeleton data of the instructor generated by the terminal device used by the instructor,
    At least one skeleton data so that the size of the skeleton data of the instructor acquired by the skeleton data acquisition unit and the skeleton data of the user instructed by the instructor generated by the skeleton data generation unit correspond to each other. And the normalization part that normalizes
    A display control unit that controls the display unit to superimpose the normalized skeleton data of the instructor and the skeleton data of the user on the image of the user captured by the image pickup unit.
    A terminal device.
  2.  正規化された前記指導者の骨格データと前記ユーザの骨格データの位置を合わせる位置合わせ部をさらに備え、
     前記位置合わせ部は、特定の関節位置を基準位置として、前記指導者の骨格データと前記ユーザの骨格データの位置を合わせる、
     請求項1に記載の端末装置。
    Further provided with an alignment unit that aligns the normalized skeleton data of the leader with the skeleton data of the user.
    The alignment unit aligns the skeleton data of the instructor with the skeleton data of the user with a specific joint position as a reference position.
    The terminal device according to claim 1.
  3.  前記ユーザの操作に基づき運動種目の選択を受け付ける操作受付部をさらに備え、
     前記位置合わせ部は、選択された運動種目に応じて前記基準位置を選択する、
     請求項2に記載の端末装置。
    Further equipped with an operation reception unit that accepts the selection of an exercise item based on the user's operation.
    The alignment unit selects the reference position according to the selected exercise item.
    The terminal device according to claim 2.
  4.  前記指導者の骨格データと前記ユーザの骨格データのずれ量をもとに、前記ユーザのフォームを前記指導者のフォームに近づけさせるためのメッセージを生成するメッセージ生成部と、
     生成されたメッセージを、映像、音声、振動の少なくとも一つで前記ユーザに通知するよう制御する通知制御部と、
     をさらに備える、請求項1から3のいずれか1項に記載の端末装置。
    A message generation unit that generates a message for bringing the user's form closer to the instructor's form based on the amount of deviation between the instructor's skeleton data and the user's skeleton data.
    A notification control unit that controls the user to be notified of the generated message by at least one of video, audio, and vibration.
    The terminal device according to any one of claims 1 to 3, further comprising.
  5.  撮像部からユーザが映った画像データを取得するステップと、
     取得された画像データをもとに、画像内の前記ユーザが映った人物領域の、関節位置を含む複数の特徴点を線で結んだ人物の骨格データを生成するステップと、
     指導者が使用する端末装置で生成された指導者の骨格データを取得するステップと、
     取得された指導者の骨格データと、生成された当該指導者に指導されているユーザの骨格データのサイズが対応するように、少なくとも一方の骨格データを正規化するステップと、
     前記撮像部により撮像されているユーザが映った画像に、正規化された前記指導者の骨格データと前記ユーザの骨格データを重畳させて、表示部に表示させるよう制御するステップと、
     を有する、フォーム改善支援方法。
    The step of acquiring the image data of the user from the image pickup unit,
    Based on the acquired image data, a step of generating skeleton data of a person connecting a plurality of feature points including joint positions in the person area in which the user is shown in the image with a line, and
    Steps to acquire the skeleton data of the instructor generated by the terminal device used by the instructor,
    A step to normalize at least one of the skeleton data so that the size of the acquired skeleton data of the instructor and the generated skeleton data of the user instructed by the instructor correspond to each other.
    A step of superimposing the normalized skeleton data of the instructor and the skeleton data of the user on the image of the user captured by the image pickup unit, and controlling the display unit to display the normalized skeleton data.
    Form improvement support method.
PCT/JP2021/010525 2020-08-27 2021-03-16 Terminal device and support method for improving form WO2022044399A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020143973A JP2022039120A (en) 2020-08-27 2020-08-27 Terminal device and form improvement support method
JP2020-143973 2020-08-27

Publications (1)

Publication Number Publication Date
WO2022044399A1 true WO2022044399A1 (en) 2022-03-03

Family

ID=80355063

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/010525 WO2022044399A1 (en) 2020-08-27 2021-03-16 Terminal device and support method for improving form

Country Status (2)

Country Link
JP (1) JP2022039120A (en)
WO (1) WO2022044399A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006230630A (en) * 2005-02-23 2006-09-07 Nihon Knowledge Kk Practical skill analysis system and program
JP2013116311A (en) * 2011-11-04 2013-06-13 Konami Digital Entertainment Co Ltd Game device, detail presentation method, and program
KR101970687B1 (en) * 2018-04-11 2019-04-19 주식회사 큐랩 Fitness coaching system using personalized augmented reality technology
US20190295438A1 (en) * 2018-03-21 2019-09-26 Physera, Inc. Augmented reality guided musculoskeletal exercises
JP2020005192A (en) * 2018-06-29 2020-01-09 キヤノン株式会社 Information processing unit, information processing method, and program
KR20200056233A (en) * 2018-11-14 2020-05-22 주식회사 퓨전소프트 A motion accuracy judgment system using artificial intelligence posture analysis technology based on single camera

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006230630A (en) * 2005-02-23 2006-09-07 Nihon Knowledge Kk Practical skill analysis system and program
JP2013116311A (en) * 2011-11-04 2013-06-13 Konami Digital Entertainment Co Ltd Game device, detail presentation method, and program
US20190295438A1 (en) * 2018-03-21 2019-09-26 Physera, Inc. Augmented reality guided musculoskeletal exercises
KR101970687B1 (en) * 2018-04-11 2019-04-19 주식회사 큐랩 Fitness coaching system using personalized augmented reality technology
JP2020005192A (en) * 2018-06-29 2020-01-09 キヤノン株式会社 Information processing unit, information processing method, and program
KR20200056233A (en) * 2018-11-14 2020-05-22 주식회사 퓨전소프트 A motion accuracy judgment system using artificial intelligence posture analysis technology based on single camera

Also Published As

Publication number Publication date
JP2022039120A (en) 2022-03-10

Similar Documents

Publication Publication Date Title
US8371989B2 (en) User-participating type fitness lecture system and fitness training method using the same
JP5881136B2 (en) Information processing apparatus and method, and program
WO2021000708A1 (en) Fitness teaching method and apparatus, electronic device and storage medium
US7494430B2 (en) System and method for correcting golf swing using internet
US10070046B2 (en) Information processing device, recording medium, and information processing method
KR102266219B1 (en) Method of providing personal training service and system thereof
US20070146484A1 (en) Automated video system for context-appropriate object tracking
KR102232253B1 (en) Posture comparison and correction method using an application that checks two golf images and result data together
EP2203896B1 (en) Method and system for selecting the viewing configuration of a rendered figure
US20200094122A1 (en) Information processor, information processing method, and program
WO2011007545A1 (en) Training machine and computer-readable medium
JP2018512980A (en) Frameworks, devices and methods configured to enable delivery of interactive skill training content, including content with multiple expert knowledge variations
CN108079547B (en) Image processing apparatus, analysis system, image processing method, and recording medium
US11682157B2 (en) Motion-based online interactive platform
KR20120034394A (en) Practicing method of golf swing motion using motion overlap and practicing system of golf swing motion using the same
JP2009034360A (en) Training system, and apparatus for the same
JP2006181014A (en) Image analysis device and movement correction system
US20140118522A1 (en) Dance learning system using a computer
WO2022044399A1 (en) Terminal device and support method for improving form
JP2018050884A (en) Notification device, notification method and program
JP2006318385A (en) Image generation system, program, and information storage medium
JP7282519B2 (en) Image processing device or image processing server
WO2020262336A1 (en) Program, method, information processing device, and swing space
US20240042309A1 (en) Control Method and Electronic Device
WO2022215313A1 (en) Information processing method, information processing device, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21860831

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21860831

Country of ref document: EP

Kind code of ref document: A1