WO2012146184A1 - 一种运动参数确定方法、装置和运动辅助设备 - Google Patents
一种运动参数确定方法、装置和运动辅助设备 Download PDFInfo
- Publication number
- WO2012146184A1 WO2012146184A1 PCT/CN2012/074748 CN2012074748W WO2012146184A1 WO 2012146184 A1 WO2012146184 A1 WO 2012146184A1 CN 2012074748 W CN2012074748 W CN 2012074748W WO 2012146184 A1 WO2012146184 A1 WO 2012146184A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- motion
- time
- sampling
- acceleration
- sampling time
- Prior art date
Links
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01P—MEASURING LINEAR OR ANGULAR SPEED, ACCELERATION, DECELERATION, OR SHOCK; INDICATING PRESENCE, ABSENCE, OR DIRECTION, OF MOVEMENT
- G01P15/00—Measuring acceleration; Measuring deceleration; Measuring shock, i.e. sudden change of acceleration
- G01P15/18—Measuring acceleration; Measuring deceleration; Measuring shock, i.e. sudden change of acceleration in two or more dimensions
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/10—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration
- G01C21/12—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning
- G01C21/16—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation
- G01C21/165—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments
- G01C21/1654—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 by using measurements of speed or acceleration executed aboard the object being navigated; Dead reckoning by integrating acceleration or speed, i.e. inertial navigation combined with non-inertial navigation instruments with electromagnetic compass
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01P—MEASURING LINEAR OR ANGULAR SPEED, ACCELERATION, DECELERATION, OR SHOCK; INDICATING PRESENCE, ABSENCE, OR DIRECTION, OF MOVEMENT
- G01P21/00—Testing or calibrating of apparatus or devices covered by the preceding groups
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0219—Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/028—Microscale sensors, e.g. electromechanical sensors [MEMS]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/002—Monitoring the patient using a local or closed circuit, e.g. in a room or building
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7242—Details of waveform analysis using integration
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2102/00—Application of clubs, bats, rackets or the like to the sporting activity ; particular sports involving the use of balls and clubs, bats, rackets, or the like
- A63B2102/32—Golf
Definitions
- the present invention relates to motion recognition technology, and more particularly to a motion parameter determination method, apparatus and motion assisting apparatus.
- the trajectory and attitude recognition of spatial acceleration motion refers to detecting the position and rotation angle of each moment in the motion of the object, and at the same time obtaining the real-time velocity of the object.
- Combining spatial acceleration trajectory and gesture recognition technology with human motion, detecting the movement of various parts of the human body can be widely used in sports, games, movies, medical simulation or motion skill training.
- the current motion recognition technologies mainly exist in the following categories:
- the RGB camera, depth sensor and microphone array are used to sample the three-dimensional motion, facial motion and sound of the user.
- the present invention provides a motion parameter determining method, device and motion assisting Equipment, in order to reduce the impact of the external environment on accuracy.
- a method for determining a motion parameter comprising:
- the attitude transformation matrix ⁇ , determining and recording the attitude transformation matrix of the current time relative to the motion start time t 0 ;
- a motion parameter determining device comprising: a data storage unit, wherein the motion data comprises: a sampled by a three-axis acceleration sensor The acceleration of the object, the angular velocity of the identified object sampled by the three-axis gyroscope, and the angle of the identified object sampled by the three-axis magnetic field sensor with respect to the three-dimensional geomagnetic coordinate system;
- a data storage unit configured to store the motion data
- the motion still detecting unit is configured to perform motion stationary detection by using the acceleration of each sampling moment stored by the data storage unit, and determine a motion starting time ⁇ of a motion state.
- An initial posture determining unit is configured to start the motion start time t according to the data storage unit.
- the angle of the motion determine the start time of the motion ⁇ .
- the initial pose matrix relative to the three-dimensional geomagnetic coordinate system;
- a motion parameter determining unit for starting the moment ⁇ from the motion. The next sampling time starts to the end time of the motion ⁇ , and the acceleration of each sampling time is determined as the current sampling time;
- the motion parameter determining unit specifically includes:
- the attitude transformation determining module is configured to use an angular velocity of the current sampling time and a previous sampling time stored by the data storage unit, and the previous sampling time relative to the motion starting time ⁇ .
- the attitude transformation matrix ⁇ determines and records the current time relative to the motion start time ⁇ .
- Attitude transformation matrix 7
- the gravity-affecting module is configured to use the r m fc to adjust the acceleration a Cur at the current sampling time, and remove the acceleration of the current sampling time by the influence of the gravity acceleration to obtain the actual acceleration a ur at the current sampling time.
- An action assisting device comprising: a sensing device and the above-described motion Parameter determining device
- the sensing device is configured to sample motion data of each sampling moment of the identified object, and send the motion data to the motion parameter determining device, where the motion data includes: an acceleration of the identified object, an angular velocity of the identified object, and the identified object The angle with respect to the three-dimensional geomagnetic coordinate system.
- the motion parameter determining method, device and motion assisting device provided by the present invention do not need to be based on vision, and the accuracy is less affected by environmental influences, especially light.
- FIG. 1 is a schematic structural diagram of an identification system according to an embodiment of the present invention.
- FIG. 2 is a schematic structural diagram of a motion assisting device according to an embodiment of the present invention
- FIG. 2 is a schematic diagram of a corner of a three-axis magnetic field sensor according to an embodiment of the present invention
- FIG. 3 is a data sent by a processor according to an embodiment of the present invention
- FIG. 4 is a flowchart of a method for determining a motion parameter according to an embodiment of the present invention
- FIG. 5 is a schematic structural diagram of a motion parameter determining apparatus according to an embodiment of the present invention.
- the present invention may employ an identification system as shown in FIG. 1a, and mainly includes: a MEMS sensing device 100, a processor 110, a data transmission interface 120, and a motion parameter determining device 130, and may further include: a motion recognition device 140, a parameter display device 150 and expert evaluation device 160.
- the MEMS sensing device 100, the processor 110 and the data transmission interface 120 may be packaged as a terminal device disposed on the identified object, for example, the package is a portable motion detecting device disposed on the player's glove, the club, the joint At the same time, the weight of the portable motion detecting device can be only a few tens of grams, and it hardly affects the motion of the other object.
- the MEMS sensing device 100 may include: a triaxial acceleration sensor 101, a three-axis gyroscope 102, and a three-axis magnetic field sensor 103.
- the triaxial acceleration sensor 101 is configured to sample the acceleration of the identified object at each sampling moment, and the acceleration is an acceleration in three-dimensional space, that is, the acceleration data corresponding to each sampling moment includes acceleration values of the X-axis, the Y-axis, and the Z-axis.
- the three-axis gyroscope 102 is used for sampling the angular velocity of the identified object at each sampling moment, and the angular velocity is the angular velocity in three-dimensional space, that is, the angular velocity data corresponding to each sampling moment includes the angular velocity values of the X-axis, the Y-axis, and the Z-axis. .
- the triaxial magnetic field sensor 103 is configured to sample the rotation angle of the identified object relative to the three-dimensional geomagnetic coordinate system at each sampling moment, and the rotation angle data corresponding to each sampling moment includes: Roll, ⁇ , and /3 ⁇ 4, where is the X-axis of the identified object.
- the angle between the XY plane in the three-dimensional geomagnetic coordinate system, w is the angle between the Y-axis of the identified object and the XY plane in the three-dimensional geomagnetic coordinate system and the positive direction of the Y-axis in the three-dimensional geomagnetic coordinate system, i3 ⁇ 4 is the identified object
- the angle between the Y-axis and the XY plane in the three-dimensional geomagnetic coordinate system as shown in Fig.
- Xmag, Ymag, and Zmag are the X-axis, Y-axis, and Z-axis of the three-dimensional geomagnetic coordinate system, respectively, and Xsen, Ysen, and Zsen are respectively identified.
- the processor 110 reads the motion data sampled by the three-axis acceleration sensor 101, the three-axis gyro 102, and the three-axis magnetic field sensor 103 in the MEMS sensing device 100 according to a certain frequency, and sends the motion data to the motion parameter determining device according to a certain transmission protocol. 130.
- Figure 3 shows a format of a packet containing motion data sent by the processor.
- the check field may contain check information for ensuring data integrity and security, and the header field may include a protocol header used to transmit the motion data.
- the processor 110 can also be configured to receive the configuration sent by the data transmission interface 120.
- the instruction parses the configuration command, and configures the MEMS sensing device 100 according to the parsed configuration information, for example, configuration of sampling precision, configuration of sampling frequency and range, etc., and can also be used for receiving motion data. Perform calibration.
- the processor 110 can use a low power processor to effectively extend the battery life.
- the triaxial acceleration sensor 101, the three-axis gyroscope 102, and the three-axis magnetic field sensor 103 in the MEMS sensing device 100 can communicate with the processor 110 via a serial bus or an AD interface.
- the data transmission interface 120 supports both wired and wireless communication transmission methods.
- the wired interface can use various protocols such as USB, serial port, parallel port, and FireWire;
- the wireless interface can use Bluetooth, infrared, and other protocols.
- Taking the USB interface 121 and/or the Bluetooth module 122 as an example is shown in FIG.
- the USB interface 121 can realize charging when the MEMS sensing device 100, the processor 110, and the data transmission interface 120 are packaged as one terminal device and two-way communication with other devices.
- the Bluetooth module 122 can implement two-way communication between the above terminal device and the Bluetooth master device.
- the motion parameter determining device 130, the motion recognition device 140, the parameter display device 150, and the expert evaluation device 160 may be connected to the processor 110 in the terminal device through a USB interface (not shown in FIG. 1), or may be used as a Bluetooth master.
- the device is connected to the processor 110 in the above terminal device via the Bluetooth module 122.
- the motion parameter determining means 130 performs motion recognition using the received motion data to determine a motion parameter including at least one of acceleration information, speed information, and position information.
- the motion recognition device 140 can identify the motion type of the motion using the motion parameter determined by the motion parameter determination device 130, thereby extracting the motion parameter corresponding to a motion of a certain motion type.
- the parameter display device 150 displays the motion parameters determined by the motion parameter determining device 130 in some form (the connection relationship of the case is not shown in the figure) or the motion recognition device
- the extracted motion parameters are displayed in some form, for example, displaying the position information of the recognized object in the form of a 3D trajectory, displaying the speed information of the recognized object, and the like in the form of a table or a curve.
- the parameter display device 150 can be any terminal having a display function, such as a computer, a mobile phone, a PDA, or the like.
- the expert evaluation device 160 determines the motion parameter determined by the motion parameter determining device 130 (the connection relationship of the case is not shown in FIG. 1a), or evaluates the motion of the identified object according to the display result of the parameter display device 150, and the evaluation may From a real expert, it can also be an evaluation automatically given by the device based on a pre-excavated database of motion parameters.
- the MEMS sensing device 100 and the motion parameter determining device 130 may be packaged as an action assisting device. As shown in FIG. 1B, the motion parameter determining device 130 may directly acquire the motion data sampled by the MEMS sensing device 100. And determine the motion parameters of each sampled moment of the identified object.
- the motion data may also be read from the MEMS sensor 100 by the processor 110 at a set frequency and transmitted to the motion parameter determining device 130 in accordance with a preset transmission protocol.
- the data transmission interface 120 can be configured as the external interface connection parameter determining device 130, and the data transmission interface 120 can also be the USB interface 121 or the Bluetooth interface 122.
- the data transmission interface 120 can transmit the motion data determined by the motion parameter determining device 130 to an external device of the motion assisting device, such as a motion recognition device, a motion display device, or an expert evaluation device.
- the data transmission interface 120 may be disposed between the processor and the motion parameter determining means 130 in the manner shown in FIG.
- the motion parameter determining apparatus 130 by way of an embodiment.
- the motion parameter determination method is described. As shown in FIG. 4, the motion parameter determining method may include the following steps:
- Step 401 Acquire motion data of each sampling moment, where the motion data includes: an acceleration of the identified object sampled by the triaxial acceleration sensor, an angular velocity of the identified object sampled by the triaxial gyroscope, and a sampled by the triaxial magnetic field sensor. The angle of the object relative to the three-dimensional geomagnetic coordinate system is identified.
- the obtained motion data may be interpolated, for example, linearly. Interpolation or spline interpolation.
- Step 402 Perform preprocessing on the acquired motion data.
- the pre-processing in this step is to filter the acquired motion data to reduce the noise of the motion data sampled by the MEMS sensor.
- filtering methods can be used, for example, 16-point Fast Fourier Transform (FFT) filtering can be used, and the specific filtering method is not limited here.
- FFT Fast Fourier Transform
- the above interpolation processing and preprocessing have no fixed order and can be executed in any order. Alternatively, the two can also be executed one by one.
- Step 403 Perform data calibration on the pre-processed motion data.
- the acceleration sampled by the three-axis acceleration sensor is mainly calibrated, and the zero-drift of the three-axis acceleration sensor is utilized.
- the obtained acceleration at each sampling time is removed from the zero drift, and the acceleration at each sampling time after calibration is obtained.
- the zero drift of the three-axis acceleration sensor It is obtained by sampling the acceleration of a stationary object.
- step 402 and the step 403 are preferred steps in the embodiment of the present invention, and the motion data acquired in step 401 is directly cached without performing step 402 and step 403.
- Step 404 Cache the motion data of each sampled time after calibration.
- the motion data of the newly obtained N sampling moments is stored in the buffer area, that is, the buffered motion data includes: the latest one sampling time to the motion data of the first N-1 sampling moments, that is, the N sampling moments are buffered in the buffer area.
- the motion data when the motion data with the new sampling moment is buffered into the buffer area, the motion data of the earliest sampling moment overflows.
- N may be an integer of 3 or more, and is usually set to an integer power of 2, for example, the value of N is 16 or 32 to maintain motion data with a buffer length of 0.1 s to 0.2 s in the buffer area.
- the data structure of the buffer area is a queue, which is arranged in order according to the sampling time, and the motion data of the latest sampling moment is placed at the end of the queue.
- Step 405 Perform motion stationary detection by using the acceleration of each sampling moment to determine a starting moment ⁇ of a motion state. And the end time t e .
- the starting moment is ⁇ .
- the critical sampling instant is the stationary state to the motion state
- the ending time t e is the critical sampling instant of the motion state to the stationary state.
- each sampling moment is judged according to a preset motion time determination strategy, if ⁇ . Meet the motion time determination strategy, and the sampling time ⁇ . -1 does not satisfy the motion time determination strategy, then determines ⁇ . For the beginning of the sport. If ⁇ satisfies the motion time determination strategy, and the sampling time ⁇ +1 does not satisfy the motion time determination strategy, it is determined that ⁇ is the motion end time.
- the motion time is determined policy may be: If the sampling time ⁇ to its previous acceleration ⁇ sampling instants after the modulo variance ⁇ ⁇ greater than or equal to a preset acceleration variance threshold and the sampling time t acceleration x modulo The resulting ⁇ . If the preset motion acceleration threshold is greater than or equal to, the sampling time ⁇ is considered as the motion moment. That is to say, if a certain sampling moment satisfies the above-mentioned motion time strategy, it is considered that the sampling moment enters the motion state, otherwise it is still at rest. Status. Where T is the default positive integer (
- the above-described motion time determination strategy can effectively filter short-term jitter, preventing short-term rest and pause to cut off complete motion.
- the acceleration variance threshold and the motion acceleration threshold can be flexibly set according to the degree of motion of the recognized object. The more intense the motion of the identified object, the higher the acceleration variance threshold and the motion acceleration threshold can be set.
- Steps 406 to 411 are executed at each sampling time between the end time and the end time ⁇ as the current sampling time.
- Step 406 Determine the motion start time ⁇ according to the motion data sampled by the three-axis magnetic field sensor in the buffer area.
- Step 407 Determine the attitude change matrix of the previous sampling time to the current sampling time according to the angular velocity data sampled by the three-axis gyroscope at the current sampling time and the previous sampling time when the identified object is in the motion state.
- the interval between adjacent sampling instants determines the previous sampling
- Step 408 Determine and record the pose transformation matrix T t r of the current time relative to the identified object using the pose transformation matrix ⁇ ⁇ and 7 of the previous sampling instant with respect to t.
- attitude transformation matrix ⁇ ⁇ of the previous sampling moment of the record which can be:
- step 407 r m fcM ⁇
- C W r represents the current sampling time
- Mt represents the motion start time ⁇ .
- r + represents the attitude change matrix from the sampling time J to the sampling time.
- the object in the static state can be used to determine the gravity acceleration in the three-dimensional geomagnetic coordinate system.
- the three-axis acceleration sensor can be used to continuously smash the object in a stationary state.
- the sampling time is sampled, and the average value of the gravity acceleration in the geomagnetic coordinate system of consecutive M sampling moments is taken as the actual gravity acceleration in the current geomagnetic coordinate system, that is, ⁇ can be according to the formula
- ⁇ is a preset positive integer
- ⁇ is the initial sampling moment for sampling an object in a stationary state
- I ⁇ i mi a is the acceleration sampled by the three-axis acceleration sensor at the sampling time
- 7 ⁇ is the attitude matrix of the above-mentioned stationary object at the sampling time, which is sampled according to the three-axis magnetic field sensor. The angle of the moment is determined as follows:
- Step 411 For ⁇ .
- the actual acceleration to the current sampling time is integrated to obtain the real-time speed of the current sampling time, right.
- the real-time speed to the current sampling time is integrated to obtain the position of the current sampling time.
- the method for obtaining the real-time speed and position by means of integration in this step is a well-known technique, and will not be described in detail here .
- At least one of acceleration, real-time speed, and position at each sampling instant between the start time t 0 and the end time t e is stored in the database as a motion parameter of a motion.
- the motion state of both ends is considered to be a motion state.
- the time interval between the sampling time ⁇ ' at the end of the motion state of the previous period is less than the preset duration threshold, and the attitude matrix of ⁇ ' is taken as the initial pose matrix r m fcM of t 0 ; otherwise, ⁇ is determined according to the formula (1).
- the initial pose matrix r m fcM is determined according to the formula (1).
- the apparatus may include: a motion data acquiring unit 500, a data storing unit 510, a motion still detecting unit 520, and an initial posture.
- the determining unit 530 and the motion parameter determining unit 540 may include: a motion data acquiring unit 500, a data storing unit 510, a motion still detecting unit 520, and an initial posture.
- the motion data acquiring unit 500 acquires the motion data of each sampling moment and sends it to the data storage unit 510, wherein the motion data includes: an acceleration of the identified object sampled by the triaxial acceleration sensor, and an angular velocity of the identified object sampled by the triaxial gyroscope And the angle of the identified object sampled by the three-axis magnetic field sensor with respect to the three-dimensional geomagnetic coordinate system.
- Data storage unit 510 will store the motion data.
- the data storage unit 510 can store the motion data of the newly obtained N sampling moments into the buffer area; where N is an integer greater than 3, and the motion data in the buffer area is sequentially arranged according to the sampling time, and the latest sampling time is The motion data is queued at the end of the queue in the buffer. That is, the data storage unit 510 stores the motion data of the latest one sampling time to the first N-1 sampling times into the buffer area.
- the motion still detecting unit 520 utilizes the sampling moments stored by the data storage unit 510.
- the acceleration performs motion stationary detection to determine the motion start time ⁇ of a motion state. And the end of the sport.
- the initial posture determining unit 530 is based on the motion start time ⁇ stored by the data storage unit 510. The angle of the sample is taken to determine the start of motion ⁇ . Initial attitude moment relative to the three-dimensional geomagnetic coordinate system
- the motion parameter determination unit 540 starts from the motion start time ⁇ .
- the next sampling time starts to the end time of motion ⁇ , and the acceleration of each sampling time is determined as the current sampling time.
- the motion parameter determining unit 540 may specifically include: a posture transformation determining module 541, a real-time posture determining module 542, and a de-gravity influencing module 543.
- the attitude transformation determining module 541 is based on the current sampling time stored by the data storage unit 510 and the angular velocity sampled by the previous sampling time, and the previous sampling time relative to the motion starting time ⁇ .
- the attitude transformation matrix ⁇ ⁇ determines and records the pose transformation matrix ⁇ of the current time relative to the motion start time t 0 .
- the de-gravity influence module 543 uses r m fc to adjust the acceleration of the current sampling time, and then removes the acceleration of the current sampling time to remove the influence of the gravitational acceleration to obtain the actual acceleration a m Mcur at the current sampling time.
- the apparatus may further include: at least one of a pre-processing unit 550 and a filter processing unit 560.
- a pre-processing unit 550 When the two units are included at the same time, the processing of the two units has no fixed order and can be performed in any order.
- Figure la shows an example of including both units.
- the pre-processing unit 550 performs the interpolation processing on the motion data transmitted by the motion data acquiring unit 500 to the data storage unit 510.
- the pre-processing unit 550 can improve the calculation accuracy of the subsequent calculation of motion parameters such as acceleration, velocity, and position when the sampling frequency of the MEMS sensor is not high enough.
- the interpolation processing method used may include but is not limited to: linear interpolation or spline interpolation.
- the filter processing unit 560 performs filtering processing on the motion data transmitted from the motion data acquiring unit 500 to the data storage unit 510 to eliminate noise of the motion data.
- a specific filtering method can be selected, and a plurality of filtering methods can be used, for example, 16-point FFT filtering.
- the apparatus may further include: a data calibration unit 570 for using a zero drift pair of the three-axis acceleration sensor
- the motion data acquired by the motion data acquiring unit 500 to the data storage unit 510 performs data calibration, and the accelerations of the obtained sampling moments are all removed by zero drift.
- the motion still detecting unit 520 may perform the motion stationary detection according to the sequence of the sampling moments, and determine the sampling time according to the preset motion time determining strategy for each sampling moment, if the sampling time ⁇ . Meet the motion time determination strategy, and the sampling time ⁇ . -1 does not satisfy the motion time determination strategy, then determines ⁇ . For the motion start time; if the sampling time ⁇ satisfies the motion time determination strategy, and the sampling time ⁇ +1 does not satisfy the motion time determination strategy, it is determined as the motion end time.
- said motion timing determined policy may be: variance ⁇ if the sampling time ⁇ to its previous ⁇ acceleration sampling time modulo the ⁇ greater than or equal to a preset acceleration variance threshold and the sampling time t acceleration x modulo give ⁇ . If the preset motion acceleration threshold is greater than or equal to, the sampling moment is determined as the motion moment, where ⁇ is a preset positive integer.
- the initial posture determining unit 530 may specifically include: the motion interval determining module 531 and Initial pose determination module 532.
- the motion interval judging module 531 judges whether the time interval between the motion start time t 0 and the motion end time ' of the previous motion state is less than a preset time threshold.
- the initial posture determination module 532 starts the time ⁇ when the determination result of the motion interval determination module 531 is NO. With respect to the initial attitude matrix r m fcM of the three-dimensional geomagnetic coordinate system, when the determination result of the motion interval determination module 531 is YES, the attitude matrix with respect to the three-dimensional geomagnetic coordinate system is determined as the motion start time ⁇ .
- the initial pose matrix T Mt relative to the three-dimensional geomagnetic coordinate system.
- the attitude transformation determining module 541 may specifically include: a first posture transformation determining sub-module 5411 and a second posture transformation determining sub-module 5412.
- the first posture transformation determining sub-module 5411 determines the posture change matrix from the previous sampling time to the current sampling time according to the current sampling time stored by the data storage unit 510 and the angular velocity sampled by the previous sampling time.
- the second attitude transformation determination sub-module 5412 acquires the previous sampling time of the record relative to the motion start time.
- the apparatus may further include: a gravity motion parameter determining unit 580.
- the gravity motion parameter determining unit 580 may specifically include: a data acquiring module 581 and a gravity acceleration determining module 582.
- the data acquisition module 581 acquires the acceleration and angle at which the object in the stationary state is sampled at successive M sampling times. That is, the three-axis acceleration sensor and the three-axis magnetic field sensor sample the object in the stationary state, and the data acquisition module 581 obtains the acceleration and the angle of the M sampling moments therefrom.
- Roll, Faw, and Pitchj are the angles at which the three-axis magnetic field sensor samples the sampling instant.
- the motion parameter determining unit 540 may further include: a speed determining module 544 and a position determining module 545.
- the speed determination module 544 pairs the motion start time ⁇ .
- the actual acceleration to the current sampling time is integrated to obtain the real-time speed of the current sampling time.
- the position determination module 545 pairs the motion start time ⁇ .
- the real-time speed to the current sampling time is integrated to obtain the position of the current sampling time.
- At least one of acceleration, real-time speed, and position at each sampling instant between the start time t 0 and the end time ⁇ is stored in the database as a motion parameter of a motion.
- the following applications can be further applied: 1) transmitting motion parameters such as position information and posture information at each sampling time to the motion recognition device (such as the motion recognition device 140 in FIG. 1a), and the motion recognition device recognizes the motion type of the motion according to the motion parameter, thereby extracting A motion parameter corresponding to a motion of a certain type of motion.
- the MEMS sensor is disposed on the golf glove, and after determining the motion parameters of the golf glove by using the motion parameter determining method and apparatus provided by the present invention, the motion parameters are provided to the motion recognition device, because the player performs the golf swing action.
- other actions such as taking a break, answering a call, and the like may be performed, and the motion recognition device can recognize and extract the motion parameters corresponding to a complete golf swing action.
- the parameter display device such as the parameter display device 150 in FIG. 1a
- the device may display in the form of a table according to the position information of each sampling moment, or display the 3D motion track of the identified object, and/or display the table according to the speed information of each sampling time, or display the identified object in a curved form.
- Speed information Through the parameter display device, the user can view the specific motion details of the identified object, such as the real-time speed of the motion, the position, the time distribution of the position, and the time distribution of the speed.
- the expert evaluation device may be a device with an automatic evaluation function.
- the expert evaluation device may search for a pre-excavated motion parameter database, where the motion parameter database stores evaluation information corresponding to various motion parameters, acceleration at each moment, The real-time speed and position information are given corresponding evaluations.
- the expert evaluation device can also be a user interface, and the motion parameters are provided to the expert through the user interface, and the expert according to the motion parameters
- the user interface can obtain the evaluation information input by the expert, and send the evaluation information to the terminal device for viewing and reference by the user of the terminal device.
- the MEMS sensing device is taken as an example for description, but the present invention is not limited thereto, and other sensing devices other than the MEMS sensing device may be used as long as it can be realized.
- the motion data sample described in the embodiment of the present invention may be used.
Landscapes
- Engineering & Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Automation & Control Theory (AREA)
- Biophysics (AREA)
- Surgery (AREA)
- Dentistry (AREA)
- Physiology (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Length Measuring Devices With Unspecified Measuring Means (AREA)
- Gyroscopes (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014506744A JP5966208B2 (ja) | 2011-04-29 | 2012-04-26 | 運動パラメーター確定方法、装置及び運動支援装置 |
KR20137020213A KR101509472B1 (ko) | 2011-04-29 | 2012-04-26 | 운동파라미터 확정방법, 장치와 운동지원방법 |
EP20120776967 EP2704098A4 (en) | 2011-04-29 | 2012-04-26 | METHOD AND DEVICE FOR DETERMINING MOTION PARAMETER AND AUXILIARY MOTION EQUIPMENT |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201110111559A CN102184549B (zh) | 2011-04-29 | 2011-04-29 | 一种运动参数确定方法、装置和运动辅助设备 |
CN201110111559.8 | 2011-04-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012146184A1 true WO2012146184A1 (zh) | 2012-11-01 |
Family
ID=44570718
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2012/074748 WO2012146184A1 (zh) | 2011-04-29 | 2012-04-26 | 一种运动参数确定方法、装置和运动辅助设备 |
Country Status (8)
Country | Link |
---|---|
US (1) | US8725452B2 (zh) |
EP (1) | EP2704098A4 (zh) |
JP (1) | JP5966208B2 (zh) |
KR (1) | KR101509472B1 (zh) |
CN (1) | CN102184549B (zh) |
AU (1) | AU2011244873B1 (zh) |
CA (1) | CA2757672C (zh) |
WO (1) | WO2012146184A1 (zh) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9589207B2 (en) | 2013-11-21 | 2017-03-07 | Mo' Motion Ventures | Jump shot and athletic activity analysis system |
US10664690B2 (en) | 2013-11-21 | 2020-05-26 | Mo' Motion Ventures | Jump shot and athletic activity analysis system |
Families Citing this family (104)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9636550B2 (en) | 2009-11-19 | 2017-05-02 | Wilson Sporting Goods Co. | Football sensing |
US10821329B2 (en) | 2009-11-19 | 2020-11-03 | Wilson Sporting Goods Co. | Football sensing |
US10751579B2 (en) | 2009-11-19 | 2020-08-25 | Wilson Sporting Goods Co. | Football sensing |
US10668333B2 (en) | 2009-11-19 | 2020-06-02 | Wilson Sporting Goods Co. | Football sensing |
US8941723B2 (en) | 2010-08-26 | 2015-01-27 | Blast Motion Inc. | Portable wireless mobile device motion capture and analysis system and method |
US9607652B2 (en) | 2010-08-26 | 2017-03-28 | Blast Motion Inc. | Multi-sensor event detection and tagging system |
US9626554B2 (en) | 2010-08-26 | 2017-04-18 | Blast Motion Inc. | Motion capture system that combines sensors with different measurement ranges |
US9076041B2 (en) | 2010-08-26 | 2015-07-07 | Blast Motion Inc. | Motion event recognition and video synchronization system and method |
US9604142B2 (en) | 2010-08-26 | 2017-03-28 | Blast Motion Inc. | Portable wireless mobile device motion capture data mining system and method |
US9235765B2 (en) | 2010-08-26 | 2016-01-12 | Blast Motion Inc. | Video and motion event integration system |
US9646209B2 (en) | 2010-08-26 | 2017-05-09 | Blast Motion Inc. | Sensor and media event detection and tagging system |
US9247212B2 (en) | 2010-08-26 | 2016-01-26 | Blast Motion Inc. | Intelligent motion capture element |
US9619891B2 (en) | 2010-08-26 | 2017-04-11 | Blast Motion Inc. | Event analysis and tagging system |
US9261526B2 (en) | 2010-08-26 | 2016-02-16 | Blast Motion Inc. | Fitting system for sporting equipment |
US9940508B2 (en) | 2010-08-26 | 2018-04-10 | Blast Motion Inc. | Event detection, confirmation and publication system that integrates sensor data and social media |
US9320957B2 (en) | 2010-08-26 | 2016-04-26 | Blast Motion Inc. | Wireless and visual hybrid motion capture system |
US9406336B2 (en) | 2010-08-26 | 2016-08-02 | Blast Motion Inc. | Multi-sensor event detection system |
US9418705B2 (en) | 2010-08-26 | 2016-08-16 | Blast Motion Inc. | Sensor and media event detection system |
US9396385B2 (en) | 2010-08-26 | 2016-07-19 | Blast Motion Inc. | Integrated sensor and video motion analysis method |
US9401178B2 (en) | 2010-08-26 | 2016-07-26 | Blast Motion Inc. | Event analysis system |
CN102184549B (zh) * | 2011-04-29 | 2012-10-10 | 闫文闻 | 一种运动参数确定方法、装置和运动辅助设备 |
BR112014025395B1 (pt) * | 2012-04-13 | 2022-07-26 | Wi-Tronix, Llc | Métodos para registrar, processar e transmitir dados a partir de um bem móvel |
CN102842156B (zh) * | 2012-09-21 | 2015-07-08 | 电子科技大学 | 获取车辆行驶状况数据、评估车辆行驶状况的方法及装置 |
CN102923572B (zh) * | 2012-09-24 | 2015-08-26 | 苏州市思玛特电力科技有限公司 | 一种吊车负载空间摆角检测技术及装置 |
US9283457B2 (en) | 2012-11-09 | 2016-03-15 | Wilson Sporting Goods Co. | Sport performance system with ball sensing |
US10159884B2 (en) | 2012-11-09 | 2018-12-25 | Wilson Sporting Goods Co. | Basketball make-miss shot sensing |
US9656143B2 (en) | 2012-11-09 | 2017-05-23 | Wilson Sporting Goods Co. | Basketball shot determination system |
US9901801B2 (en) | 2012-11-09 | 2018-02-27 | Wilson Sporting Goods Co. | Basketball sensing apparatus |
US9623311B2 (en) | 2012-11-09 | 2017-04-18 | Wilson Sporting Goods Co. | Basketball sensing apparatus |
US9724570B2 (en) | 2012-11-09 | 2017-08-08 | Wilson Sporting Goods Co. | Ball lighting |
US9656142B2 (en) | 2012-11-09 | 2017-05-23 | Wilson Sporting Goods Co. | Basketball shot determination system |
US9844704B2 (en) | 2012-11-09 | 2017-12-19 | Wilson Sporting Goods Co. | Basketball sensing apparatus |
US9656140B2 (en) | 2012-11-09 | 2017-05-23 | Wilson Sporting Goods Co. | Sport performance system with ball sensing |
EP2920670A4 (en) * | 2012-11-16 | 2016-06-29 | Nokia Technologies Oy | TRANSMISSION OF MOTION DATA |
US10912131B2 (en) * | 2012-12-03 | 2021-02-02 | Samsung Electronics Co., Ltd. | Method and mobile terminal for controlling bluetooth low energy device |
JP6268945B2 (ja) * | 2013-02-04 | 2018-01-31 | 株式会社リコー | 慣性装置、方法及びプログラム |
US9384671B2 (en) | 2013-02-17 | 2016-07-05 | Ronald Charles Krosky | Instruction production |
CN104035685A (zh) * | 2013-03-07 | 2014-09-10 | 龙旗科技(上海)有限公司 | 一种基于动作传感的手持终端解锁方法 |
US20140274486A1 (en) | 2013-03-15 | 2014-09-18 | Wilson Sporting Goods Co. | Ball sensing |
US11040244B2 (en) | 2013-03-15 | 2021-06-22 | Karsten Manufacturing Corporation | Impact and sound analysis for golf equipment |
CN105378432B (zh) * | 2013-03-15 | 2019-06-18 | 谷歌有限责任公司 | 用于姿态校正的系统和方法 |
US10391358B2 (en) | 2013-03-15 | 2019-08-27 | Karsten Manufacturing Corporation | Impact and sound analysis for golf equipment |
US9612253B2 (en) * | 2013-03-15 | 2017-04-04 | Nike, Inc. | Impact and sound analysis for golf equipment |
US9342737B2 (en) * | 2013-05-31 | 2016-05-17 | Nike, Inc. | Dynamic sampling in sports equipment |
US9597554B2 (en) | 2013-08-07 | 2017-03-21 | Wilson Sporting Goods Co. | Racquet hit notification |
JP6152763B2 (ja) * | 2013-09-19 | 2017-06-28 | カシオ計算機株式会社 | 運動支援装置及び運動支援方法、運動支援プログラム |
CN103519820A (zh) * | 2013-09-29 | 2014-01-22 | 上海致趣网络科技有限公司 | 穿戴式设备运动检测方法 |
US9833683B2 (en) | 2013-10-16 | 2017-12-05 | Wilson Sporting Goods Co. | Golf ball and caddie system |
US10220286B2 (en) | 2013-10-16 | 2019-03-05 | Wilson Sporting Goods Co. | Golf ball and caddie system |
EP2899547B1 (en) * | 2013-12-05 | 2016-09-28 | Huawei Device Co., Ltd. | Method and apparatus for determining acceleration of vehicle |
US10281484B2 (en) * | 2014-05-02 | 2019-05-07 | Qualcomm Incorporated | Motion direction determination and application |
CN113303790A (zh) * | 2014-05-30 | 2021-08-27 | 日东电工株式会社 | 用于对用户活动进行分类和/或对用户步数进行计数的设备和方法 |
KR102233301B1 (ko) | 2014-06-12 | 2021-03-31 | 순위안 카이화 (베이징) 테크놀로지 컴퍼니 리미티드 | 스포츠 기구 내에 삽입된 제거 가능한 동작 센서 |
CN104112301B (zh) * | 2014-06-23 | 2015-07-22 | 深圳市一体数科科技有限公司 | 基于智能终端和服务器端的汽车驾驶行为分析方法及系统 |
KR101545654B1 (ko) * | 2014-06-26 | 2015-08-20 | 주식회사 아이파이브 | 개인별 맞춤형 운동 시스템 및 개인별 맞춤형 운동 방법 |
CN104154913A (zh) * | 2014-06-28 | 2014-11-19 | 南京诺导电子科技有限公司 | 全姿态环境下自主室内定位算法 |
CN104154915B (zh) * | 2014-07-04 | 2017-10-20 | 北京搜狗科技发展有限公司 | 一种基于智能穿戴设备的音频操作方法和智能穿戴设备 |
US9916001B2 (en) | 2014-07-08 | 2018-03-13 | Wilson Sporting Goods Co. | Sport equipment input mode control |
US9409074B2 (en) | 2014-08-27 | 2016-08-09 | Zepp Labs, Inc. | Recommending sports instructional content based on motion sensor data |
US9587943B2 (en) * | 2014-10-04 | 2017-03-07 | Honeywell International Inc. | High rate rotation sensing |
WO2016096443A1 (en) | 2014-12-18 | 2016-06-23 | Koninklijke Philips N.V. | Activity classification and communication system for wearable medical device |
JP6474128B2 (ja) * | 2015-03-18 | 2019-02-27 | アルプスアルパイン株式会社 | 地磁気センサと加速度センサを搭載した電子機器 |
KR102390876B1 (ko) * | 2015-03-27 | 2022-04-26 | 삼성전자주식회사 | 가속도 센서를 이용하여 사용자의 활동을 인식하는 방법 및 장치 |
US9889358B2 (en) | 2015-06-04 | 2018-02-13 | Jeffrey Kyle Greenwalt | Systems and methods utilizing a ball including one or more sensors to improve pitching performance |
US9808692B2 (en) | 2015-06-04 | 2017-11-07 | Jeffrey Kyle Greenwalt | Ball including one or more sensors to improve pitching performance |
DE202015103582U1 (de) | 2015-07-07 | 2015-08-20 | Oliver Baltzer | Vorrichtung zum Erkennen eines Aufschlags |
CN105043383A (zh) * | 2015-07-10 | 2015-11-11 | 哈尔滨医科大学 | 一种姿态矫正方法及装置 |
US11565163B2 (en) | 2015-07-16 | 2023-01-31 | Blast Motion Inc. | Equipment fitting system that compares swing metrics |
US11577142B2 (en) | 2015-07-16 | 2023-02-14 | Blast Motion Inc. | Swing analysis system that calculates a rotational profile |
US10974121B2 (en) | 2015-07-16 | 2021-04-13 | Blast Motion Inc. | Swing quality measurement system |
US9694267B1 (en) | 2016-07-19 | 2017-07-04 | Blast Motion Inc. | Swing analysis method using a swing plane reference frame |
US10124230B2 (en) | 2016-07-19 | 2018-11-13 | Blast Motion Inc. | Swing analysis method using a sweet spot trajectory |
EP3323084A4 (en) | 2015-07-16 | 2019-07-10 | Blast Motion Inc. | SYSTEM FOR MULTISENSOR EVENT RECOGNITION AND MARKING |
US10172541B2 (en) * | 2015-09-25 | 2019-01-08 | J-Mex Inc. | Motion recognition device and method |
CN106708252A (zh) * | 2015-11-12 | 2017-05-24 | 昆山罗伯斯特机器人科技有限公司 | 一种智能无线控制手势手套 |
CN105407238B (zh) * | 2015-12-08 | 2018-11-02 | 北京百度网讯科技有限公司 | 一种基于传感器的骑行监测方法及装置 |
US10215542B2 (en) | 2015-12-09 | 2019-02-26 | Virtual Clays, LLC | System for analyzing performance of an activity involving using an implement to strike a moving target object effectively |
US10161954B2 (en) * | 2016-01-22 | 2018-12-25 | Htc Corporation | Motion detecting device and detecting method for repetitive motion |
CN105549746B (zh) * | 2016-01-28 | 2018-10-30 | 广州成潮智能科技有限公司 | 一种基于加速度传感芯片的动作识别方法 |
TWI729064B (zh) * | 2016-01-28 | 2021-06-01 | 日商日本鼎意股份有限公司 | 包括內置有感應器的球的系統、行動終端的程式及經由行動終端監控球的動向之方法 |
US10265602B2 (en) | 2016-03-03 | 2019-04-23 | Blast Motion Inc. | Aiming feedback system with inertial sensors |
CN105957183B (zh) * | 2016-04-21 | 2018-05-29 | 深圳市元征科技股份有限公司 | 一种车辆静止状态检测方法及装置 |
CN106097843A (zh) * | 2016-08-30 | 2016-11-09 | 广东工业大学 | 一种运动学实验系统及方法 |
CN106444809B (zh) * | 2016-10-12 | 2024-04-16 | 湖南绿野航空科技有限公司 | 一种无人机飞行控制器 |
JP2018082771A (ja) * | 2016-11-21 | 2018-05-31 | カシオ計算機株式会社 | 運動解析装置、運動解析方法及びプログラム |
CN107049324B (zh) * | 2016-11-23 | 2019-09-17 | 深圳大学 | 一种肢体运动姿态的判断方法及装置 |
CN108426573B (zh) * | 2017-02-14 | 2023-04-07 | 中兴通讯股份有限公司 | 一种终端设备的行人步态检测方法及终端设备 |
CN108663067A (zh) * | 2017-03-30 | 2018-10-16 | 杭州维圣智能科技有限公司 | 一种运动传感器的自适应校准方法和系统 |
US10786728B2 (en) | 2017-05-23 | 2020-09-29 | Blast Motion Inc. | Motion mirroring system that incorporates virtual environment constraints |
CN107329860B (zh) * | 2017-05-25 | 2020-07-14 | 福建网龙计算机网络信息技术有限公司 | Vr设备传感器的数据采样率检测方法及其系统 |
CN107883953B (zh) * | 2017-09-26 | 2021-05-25 | 广州新维感信息技术有限公司 | 一种vr手柄静态检测算法、vr手柄及存储介质 |
CN107976559A (zh) * | 2017-10-11 | 2018-05-01 | 常州信息职业技术学院 | 基于静止点检测的滤除重力加速度方法 |
WO2019099053A1 (en) * | 2017-11-19 | 2019-05-23 | Mocini Jeffrey | Belt-mounted slope sensor system |
CN110966978B (zh) * | 2018-09-28 | 2021-10-26 | 千寻位置网络有限公司 | 自行车轮胎变形的检测方法及装置、自行车 |
KR102361148B1 (ko) * | 2018-10-30 | 2022-02-10 | 에이케이시스 주식회사 | 응용프로그램 구동 환경을 제공하는 가상 안드로이드 장치를 이용한 제어 시스템 |
US20200215376A1 (en) * | 2019-01-07 | 2020-07-09 | Spencer Bishop | Smartbell |
CN110246561A (zh) * | 2019-04-30 | 2019-09-17 | 华为技术有限公司 | 一种移动距离计算方法、装置以及系统 |
CN111399669B (zh) * | 2020-03-27 | 2023-08-18 | 歌尔股份有限公司 | 耳机的位姿变化识别方法、装置、无线耳机及存储介质 |
CN111693019B (zh) * | 2020-05-20 | 2021-04-20 | 西安交通大学 | 一种姿态传感装置及数据融合、姿态解算方法 |
CN112288766B (zh) * | 2020-10-28 | 2024-05-28 | 中国科学院深圳先进技术研究院 | 运动评估方法、装置、系统及存储介质 |
CN113074724B (zh) * | 2021-03-26 | 2023-05-02 | 歌尔股份有限公司 | 运动时间计算方法、装置、设备及计算机可读存储介质 |
CN113376404B (zh) * | 2021-05-19 | 2022-11-18 | 中寰卫星导航通信有限公司 | 一种标定验证系统、设备、方法及存储介质 |
CN114176576B (zh) * | 2021-12-11 | 2024-05-24 | 江苏智恒文化科技有限公司 | 基于加速度识别人体运动状态的方法 |
CN114967942B (zh) * | 2022-03-31 | 2023-04-18 | 河北迥然科技有限公司 | 一种运动姿态分析方法、终端及存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007124257A2 (en) * | 2006-04-21 | 2007-11-01 | Motorola, Inc. | Method and system for personal inertial navigation measurements |
CN101299271A (zh) * | 2008-06-12 | 2008-11-05 | 复旦大学 | 一种机动目标状态方程的多项式预测模型及跟踪方法 |
CN101488222A (zh) * | 2008-01-16 | 2009-07-22 | 中国科学院自动化研究所 | 一种基于运动目标表象和运动信息的摄像机自标定方法 |
CN101886927A (zh) * | 2010-06-25 | 2010-11-17 | 武汉大学 | 基于惯性传感器和地磁传感器的三维运动跟踪系统与方法 |
CN101964047A (zh) * | 2009-07-22 | 2011-02-02 | 深圳泰山在线科技有限公司 | 一种基于多跟踪点的人体动作识别方法 |
CN102184549A (zh) * | 2011-04-29 | 2011-09-14 | 韩铮 | 一种运动参数确定方法、装置和运动辅助设备 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002037827A2 (en) * | 2000-10-30 | 2002-05-10 | Naval Postgraduate School | Method and apparatus for motion tracking of an articulated rigid body |
KR100631035B1 (ko) * | 2004-06-03 | 2006-10-02 | 이기영 | 구기 스포츠의 스윙 폼 교정장치 |
JP4622441B2 (ja) * | 2004-10-13 | 2011-02-02 | 横浜ゴム株式会社 | ゴルフスウィング解析システムおよびそのプログラム |
KR100718126B1 (ko) * | 2005-02-05 | 2007-05-15 | 삼성전자주식회사 | 동작 인식 기반 입력 장치의 사용자 인터페이스 방법 및 그장치 |
NL1030440C2 (nl) * | 2005-11-16 | 2007-05-21 | Univ Twente | Bewegingsvolgsysteem. |
US8082122B2 (en) * | 2006-12-12 | 2011-12-20 | Samsung Electronics Co., Ltd. | Mobile device having a motion detector |
FR2934043B1 (fr) * | 2008-07-18 | 2011-04-29 | Movea Sa | Procede d'estimation ameliore de l'orientation d'un objet et centrale d'attitude mettant en oeuvre un tel procede |
JP5604779B2 (ja) * | 2008-09-17 | 2014-10-15 | 富士通株式会社 | 携帯端末装置、スイングの測定方法及び測定プログラム |
CN101726295B (zh) * | 2008-10-24 | 2011-09-07 | 中国科学院自动化研究所 | 考虑加速度补偿和基于无迹卡尔曼滤波的惯性位姿跟踪方法 |
JP4702475B2 (ja) * | 2008-12-25 | 2011-06-15 | ソニー株式会社 | 入力装置、ハンドヘルド装置及び制御方法 |
JP5182708B2 (ja) * | 2009-06-17 | 2013-04-17 | ダンロップスポーツ株式会社 | ゴルフスイングの分析方法 |
KR101586007B1 (ko) * | 2009-06-25 | 2016-01-21 | 삼성전자주식회사 | 데이터 처리 장치 및 방법 |
JP2011033489A (ja) * | 2009-08-03 | 2011-02-17 | Aichi Micro Intelligent Corp | モーションキャプチャ用のマーカ |
KR20110060481A (ko) * | 2009-11-30 | 2011-06-08 | 울산대학교 산학협력단 | 자세각 검출 장치 및 그 자세각 검출 방법 |
JP4590018B1 (ja) * | 2010-02-26 | 2010-12-01 | エンパイア テクノロジー ディベロップメント エルエルシー | 特徴量変換装置、および特徴量変換方法 |
CN101915580B (zh) * | 2010-07-14 | 2012-09-12 | 中国科学院自动化研究所 | 一种基于微惯性和地磁技术的自适应三维姿态定位方法 |
-
2011
- 2011-04-29 CN CN201110111559A patent/CN102184549B/zh active Active
- 2011-10-07 US US13/269,236 patent/US8725452B2/en active Active
- 2011-10-28 AU AU2011244873A patent/AU2011244873B1/en not_active Ceased
- 2011-11-09 CA CA2757672A patent/CA2757672C/en active Active
-
2012
- 2012-04-26 JP JP2014506744A patent/JP5966208B2/ja active Active
- 2012-04-26 WO PCT/CN2012/074748 patent/WO2012146184A1/zh active Application Filing
- 2012-04-26 EP EP20120776967 patent/EP2704098A4/en not_active Withdrawn
- 2012-04-26 KR KR20137020213A patent/KR101509472B1/ko active IP Right Grant
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007124257A2 (en) * | 2006-04-21 | 2007-11-01 | Motorola, Inc. | Method and system for personal inertial navigation measurements |
CN101488222A (zh) * | 2008-01-16 | 2009-07-22 | 中国科学院自动化研究所 | 一种基于运动目标表象和运动信息的摄像机自标定方法 |
CN101299271A (zh) * | 2008-06-12 | 2008-11-05 | 复旦大学 | 一种机动目标状态方程的多项式预测模型及跟踪方法 |
CN101964047A (zh) * | 2009-07-22 | 2011-02-02 | 深圳泰山在线科技有限公司 | 一种基于多跟踪点的人体动作识别方法 |
CN101886927A (zh) * | 2010-06-25 | 2010-11-17 | 武汉大学 | 基于惯性传感器和地磁传感器的三维运动跟踪系统与方法 |
CN102184549A (zh) * | 2011-04-29 | 2011-09-14 | 韩铮 | 一种运动参数确定方法、装置和运动辅助设备 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2704098A4 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9589207B2 (en) | 2013-11-21 | 2017-03-07 | Mo' Motion Ventures | Jump shot and athletic activity analysis system |
US10664690B2 (en) | 2013-11-21 | 2020-05-26 | Mo' Motion Ventures | Jump shot and athletic activity analysis system |
US11227150B2 (en) | 2013-11-21 | 2022-01-18 | Mo' Motion Ventures | Jump shot and athletic activity analysis system |
Also Published As
Publication number | Publication date |
---|---|
US20120278023A1 (en) | 2012-11-01 |
US8725452B2 (en) | 2014-05-13 |
CA2757672C (en) | 2014-10-28 |
CN102184549A (zh) | 2011-09-14 |
KR20130116910A (ko) | 2013-10-24 |
EP2704098A1 (en) | 2014-03-05 |
CN102184549B (zh) | 2012-10-10 |
AU2011244873B1 (en) | 2012-09-06 |
EP2704098A4 (en) | 2014-10-01 |
JP2014512237A (ja) | 2014-05-22 |
JP5966208B2 (ja) | 2016-08-10 |
CA2757672A1 (en) | 2012-10-29 |
KR101509472B1 (ko) | 2015-04-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2012146184A1 (zh) | 一种运动参数确定方法、装置和运动辅助设备 | |
WO2012146182A1 (zh) | 一种球类运动的动作识别方法、装置和动作辅助设备 | |
US8989441B2 (en) | Data acquisition method and device for motion recognition, motion recognition system and computer readable storage medium | |
US8425292B2 (en) | System and method for analyzing postures | |
EP3060317B1 (en) | Information processing device, recording medium, and information processing method | |
US20150285834A1 (en) | Sensor, computing device, and motion analyzing apparatus | |
WO2011003218A1 (zh) | 一种加速度动作识别系统及其方法 | |
CN108769380B (zh) | 撞击角度获取方法及相关产品 | |
US20130202108A1 (en) | Method and device for generation of secret key | |
EP3528024B1 (en) | Information processing device, information processing method, and program | |
WO2015100923A1 (zh) | 用户信息获取方法及移动终端 | |
US20170203181A1 (en) | Display method, display apparatus, motion analysis system, motion analysis program, and recording medium | |
CN113687590B (zh) | Vr系统中头戴与手柄的时间同步方法、系统 | |
CN107148614A (zh) | 信息处理设备、信息处理方法和程序 | |
JP2013042246A (ja) | 携帯端末及び位置及び方位推定プログラム | |
CN107734269A (zh) | 一种图像处理方法及移动终端 | |
US11721027B2 (en) | Transforming sports implement motion sensor data to two-dimensional image for analysis | |
KR20150009626A (ko) | 사용자 모션을 이용한 원격 컨트롤 모듈 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12776967 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012776967 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20137020213 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2014506744 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |