WO2015098304A1 - 解析装置、記録媒体および解析方法 - Google Patents
解析装置、記録媒体および解析方法 Download PDFInfo
- Publication number
- WO2015098304A1 WO2015098304A1 PCT/JP2014/079389 JP2014079389W WO2015098304A1 WO 2015098304 A1 WO2015098304 A1 WO 2015098304A1 JP 2014079389 W JP2014079389 W JP 2014079389W WO 2015098304 A1 WO2015098304 A1 WO 2015098304A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- play
- analysis
- user
- play events
- extracted
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/003—Repetitive work cycles; Sequence of movements
- G09B19/0038—Sports
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
Definitions
- This disclosure relates to an analysis apparatus, a recording medium, and an analysis method.
- the present disclosure proposes a new and improved analysis device, recording medium, and analysis method capable of analyzing data obtained from a series of sports plays as a series or set.
- an acquisition function for acquiring data indicating a play event defined based on a motion of a user who plays a sport, and a plurality of play events classified into the same type are extracted from the play events.
- an analysis apparatus including a processor that realizes an extraction function and an analysis function for analyzing the data indicating the plurality of extracted play events.
- an acquisition function that acquires data indicating a play event defined based on a motion of a user who plays a sport, and a plurality of play events classified into the same type from the above play events are provided.
- a recording medium storing a program for causing a computer to implement an extraction function for extraction and an analysis function for analyzing the data indicating the plurality of extracted play events.
- data indicating a play event defined based on a motion of a user who plays a sport is acquired, and a plurality of play events classified into the same type are extracted from the play events.
- an analysis method including analyzing the data indicating the plurality of extracted play events.
- data obtained from a series of sports plays can be analyzed as a series or a set.
- FIG. 5 is a diagram for describing an example of analysis processing according to an embodiment of the present disclosure.
- FIG. It is a figure which shows notionally the function of the filter part in the example shown in FIG.
- FIG. It is a figure which shows the example of the analysis with respect to the play event group extracted in the example of FIG.
- FIG. 11 shows the example of the information which expresses the result of the analysis shown in FIG. 11 in a two-dimensional plane. It is a figure which shows the example of the information which expresses the result of the analysis shown in FIG. 11 in a two-dimensional plane.
- the following explanation is made using a specific example of sports (tennis), but the scope of application of the present technology is not limited to the exemplified sports.
- the present technology can be applied to any sport as long as a play event can be defined based on the motion of a user playing the sport.
- FIG. 1 is a diagram illustrating an example of a system configuration according to an embodiment of the present disclosure.
- the system 10 includes a sensor device 100, a smartphone 200, and a server 300.
- the sensor device 100 is mounted on a tennis racket R.
- the sensor device 100 includes, for example, a motion sensor (for example, an acceleration sensor, a gyro sensor, a geomagnetic sensor, etc.).
- a motion sensor for example, an acceleration sensor, a gyro sensor, a geomagnetic sensor, etc.
- the sensor device 100 directly detects the motion of the racket R.
- the sensor device 100 indirectly detects the motion of the racket R. It can be said that the user's motion is detected.
- the sensor device 100 is indirectly attached to the user and detects the user's motion.
- the sensor device 100 may be attached to, for example, a user's clothes or shoes. In this case as well, the sensor device 100 directly detects the motion of clothes and shoes, but since the clothing and shoes move with the user, the sensor device can indirectly detect the motion of the user. .
- the sensor device 100 may be directly attached to the user, for example, by being wound around an arm with a band. In this case, the sensor device 100 can directly detect the user's motion. Even if the sensor device 100 directly detects a user's motion, or indirectly detects the user's motion, the sensor device 100 provides if the detected motion reflects the user's motion. Based on the detection result, it is possible to define a play event corresponding to the motion of the user playing the sport.
- the sensor device 100 may include a vibration sensor.
- the data detected by the vibration sensor can easily identify the section corresponding to the play event (for example, the section before and after the impact of the ball).
- the data detected by the vibration sensor may also be used for play event analysis in the same manner as the data detected by the motion sensor.
- the sensor device 100 may further include a sensor for acquiring environment information of a user who plays sports, such as temperature, humidity, brightness, and position. Data detected by various sensors included in the sensor device 100 is preprocessed as necessary, and then transmitted to the smartphone 200 by wireless communication such as Bluetooth (registered trademark).
- the smartphone 200 is disposed near a user who is playing sports, for example.
- the smartphone 200 receives data transmitted from the sensor device 100 through wireless communication such as Bluetooth (registered trademark), temporarily stores and processes the data as necessary, and then performs data transmission through network communication. Is transmitted to the server 300.
- the smartphone 200 may receive the result of the analysis performed by the server 300 based on the transmitted data and output it to the user via a display, a speaker, or the like. Note that the user does not have to play sports when the analysis result is output.
- the output of the analysis result may be executed by an information processing terminal used by the user, for example, various personal computers or tablet terminals, a game machine, a television, or the like other than the smartphone 200.
- the smart phone 200 does not necessarily need to be arrange
- the sensor device 100 accumulates the detected data in an internal storage area (memory or external storage device).
- data may be transmitted from the sensor device 100 to the smartphone 200 by wireless communication such as Bluetooth (registered trademark).
- data may be transmitted when the sensor device 100 and the smartphone 200 are connected by wire via USB or the like.
- a removable recording medium may be used for data transfer from the sensor device 100 to the smartphone 200.
- the server 300 communicates with the smartphone 200 via the network, and receives data detected by various sensors included in the sensor device 100.
- the server 300 executes an analysis process using the received data, and generates various types of information regarding sports play.
- the server 300 defines a play event based on data obtained by a motion sensor and directly or indirectly indicating a motion of a user who plays a sport.
- a play event corresponds to one shot using a racket R, for example.
- a play event for example, a user's play represented by motion data can be grasped as a series of plays having a meaning such as ⁇ serve, stroke, volley,.
- the server 300 may extract a plurality of play events classified into the same type from the play events by play event analysis processing, and analyze the data indicating the extracted play events.
- the information generated by the analysis processing of the server 300 is transmitted to the smartphone 200, for example, and output to the user via the display or speaker of the smartphone 200.
- the server 300 may transmit information to an information processing terminal other than the smartphone 200 and output the information to the user.
- the server 300 performs analysis processing based on data received for each of a plurality of users, generates information based on a result of comparing the play patterns generated for each user, and the information processing terminal for each user May be sent to.
- FIG. 2 is a block diagram schematically illustrating a device configuration of a system according to an embodiment of the present disclosure.
- the sensor device 100 includes a sensor 110, a processing unit 120, and a transmission unit 130.
- the smartphone 200 includes a reception unit 210, a processing unit 220, a storage unit 230, a transmission unit 240, an imaging unit 250, an input unit 260, and an output unit 270.
- Server 300 includes a reception unit 310, a processing unit 320, a storage unit 330, and a transmission unit 340.
- a hardware configuration example (a hardware configuration example of a sensor device and an analysis device) for realizing each device will be described later.
- the processing unit 120 processes the data acquired by the sensor 110, and the transmission unit 130 transmits the processed data to the smartphone 200.
- the sensor 110 includes, for example, a motion sensor, and directly or indirectly detects a motion of a user who plays sports.
- the sensor 110 may further include a vibration sensor, a sensor for acquiring user environment information, and the like.
- the processing unit 120 is realized by a processor that operates according to a program, and preprocesses data acquired by the sensor 110 as necessary.
- the preprocessing can include, for example, sampling and noise removal.
- the preprocessing does not necessarily have to be executed.
- the transmission unit 130 is realized by a communication device, and transmits data to the smartphone 200 using wireless communication such as Bluetooth (registered trademark).
- the sensor device 100 may include a storage unit for temporarily storing data.
- the reception unit 210 receives data transmitted by the sensor device 100, and the transmission unit 240 transmits data to the server 300.
- the reception unit 210 and the transmission unit 240 are realized by a communication device that performs wireless communication such as Bluetooth (registered trademark) and wired or wireless network communication.
- the received data is transmitted after being temporarily stored in the storage unit 230 by the processing unit 220, for example. Further, the processing unit 220 may perform preprocessing on the received data.
- the processing unit 220 is realized by a processor that operates according to a program, and the storage unit 230 is realized by a memory or a storage.
- the receiving unit 210 may further receive information transmitted from the server 300.
- the received information is output from the output unit 270 to the user according to the control of the processing unit 220, for example.
- the output unit 270 includes a display and a speaker, for example.
- an image is acquired by the imaging unit 250.
- the imaging unit 250 is realized by, for example, a camera module that combines an imaging device with an optical system such as a lens.
- the image may include a user who plays sports as a subject.
- the image acquired by the imaging unit 250 is transmitted from the transmission unit 240 to the server 300 together with the data received by the reception unit 210, for example.
- the image may be used in the analysis process in the server 300 together with the data acquired by the sensor device 100, for example, or may be incorporated in information generated by the analysis process.
- the input unit 260 includes, for example, a touch panel, hardware buttons, and / or a microphone and a camera for receiving voice input and gesture input.
- the processing unit 220 may request information from the server 300 via the transmission unit 240 according to a user operation acquired via the input unit 260.
- the server 300 includes a reception unit 310, a processing unit 320, a storage unit 330, and a transmission unit 340.
- the receiving unit 310 is realized by a communication device, and receives data transmitted from the smartphone 200 using network communication such as the Internet.
- the processing unit 320 is realized by a processor such as a CPU, and processes received data. For example, the processing unit 320 may execute an analysis process on the received data, and may further accumulate the analyzed data in the storage unit 330 or output the data via the transmission unit 340. Alternatively, the processing unit 320 may only execute accumulation and output control of data already analyzed in the smartphone 200 or the like.
- the analysis process using the data acquired by the sensor device 100 is executed by the processing unit 320 of the server 300, but the analysis process may be executed by the processing unit 220 of the smartphone 200, It may be executed by the processing unit 120 of the sensor device 100.
- the system 10 is described as including the sensor device 100, the smartphone 200, and the server 300, for example, when an analysis process is executed by the processing unit 220 of the smartphone 200, the server 300 is included in the system 10. It does not have to be included.
- the server 300 may store the information obtained by the analysis process and provide a service shared between users.
- the smartphone 10 and the server 300 may not be included in the system 10.
- the sensor device 100 may be a dedicated sensor device attached to a user or a tool, for example, or a sensor module mounted on a portable information processing terminal may function as the sensor device 100. Therefore, the sensor device 100 can be the same device as the smartphone 200.
- FIG. 3 is a diagram for describing an example of analysis processing according to an embodiment of the present disclosure.
- the processor that executes the analysis process acquires, as inputs, motion data 401 and metadata 403 indicating a play event defined based on a motion of a user who plays sports.
- the processor extracts a plurality of play events classified into the same type from the play events indicated by the data by the function of the filter unit 405.
- the processor extracts the feature of the motion indicated by the data corresponding to the plurality of extracted play events by the function of the feature extraction unit 407, and performs the determination based on the feature by the function of the determination unit 409.
- the motion data 401 is data acquired by a motion sensor included in the sensor 110 included in the sensor device 100, and indicates a motion of a user who plays sports.
- the metadata 403 is data indicating a play event defined based on this motion.
- the motion data 401 and the metadata 403 correspond to each other.
- the metadata 403 defines the type and attribute of the play event represented by the corresponding motion data 401.
- play event types may include shot types such as serve, forehand stroke, backhand stroke, forehand volley, and the like.
- the attributes of the play event may include information such as the occurrence time (for example, the time when the impact between the racket and the ball occurs), the swing speed, and the like.
- the motion data 401 and the metadata 403 are examples of data indicating a play event, and data indicating a play event can be acquired in various other formats.
- each play event may be indicated by a single data in which motion data and metadata are combined, or motion data indicating a plurality of play events that occur in succession is integrated into one.
- Different sections of the motion data may be referred to by a plurality of metadata.
- the filter unit 405 extracts a plurality of play events classified into the same type from the play events indicated by the motion data 401 and the metadata 403.
- the types of play events may include shot types such as serve, forehand stroke, backhand stroke, and forehand volley. Therefore, the filter unit 405 extracts a play event of any shot type. Note that a single play event may be extracted. However, in this case, the analysis process executed for the extracted play event may be different from the processes in the feature extraction unit 407 and the determination unit 409 described later.
- the feature extraction unit 407 includes data indicating a plurality of play events extracted by the filter unit 405, that is, data corresponding to a plurality of play events extracted by the filter unit 405 out of the motion data 401 and the metadata 403. An analysis process for extracting features is executed. What features the feature extraction unit 407 extracts from play event data may differ depending on, for example, the analysis processing performed by the determination unit 409 that follows. Further, the extracted feature may be different depending on the type of play event extracted by the filter unit 405. As a feature extraction method, for example, k-means coding, an auto encoder, or the like can be used. Since these methods are already well known, detailed description thereof is omitted here.
- the determination unit 409 performs some determination on the plurality of play events extracted by the filter unit 405 based on the features extracted by the feature extraction unit 407. For example, when the plurality of extracted play events are single user play events, the determination unit 409 may determine the proficiency level of play indicated by these play events. More specifically, the determination unit 409 may evaluate the stability of play based on the extracted features, and determine the proficiency level based on the stability. Alternatively, the determination unit 409 uses a proficiency level determination device that learns based on teacher data for each proficiency level (advanced / intermediate / beginner) collected in advance for each type of play event (for example, the type of swing). The proficiency level may be determined.
- the determination unit 409 may determine the similarity of each user's play indicated by these play events. More specifically, the determination unit 409 calculates a distance in the feature space of each user's play based on the extracted feature (similar to the above-described stability evaluation), and the similarity based on the distance May be determined.
- the processor that executes the analysis processing may further realize a function of displaying information generated by the analysis on a display (for example, included in the output unit 270 of the smartphone 200).
- a display for example, included in the output unit 270 of the smartphone 200.
- information indicating the feature extracted by the feature extraction unit 407 may be displayed on the display.
- the analysis processing executed by the feature extraction unit 407 and the determination unit 409 in the above example may be replaced by analysis processing by another method, for example, analysis processing not based on feature extraction.
- analysis processing not based on feature extraction.
- the motion data 401 indicates the user's motion in a certain type of play event as a waveform such as acceleration
- the above-described stability and similarity are calculated based on the distance between waveforms in each of a predetermined number of sample sections. You may judge.
- a user's play habit may be detected.
- FIG. 4 is a diagram conceptually showing the function of the filter unit in the example shown in FIG. FIG. 4 shows an example in which the same type of play event group 1103 is extracted from a time series including three types of play events 1101. More specifically, the play events 1101 include a forehand stroke (FHST) play event 1101a, a backhand stroke (BHST) play event 1101b, and a serve (SRV) play event 1101c.
- the play event group 1103 includes a forehand stroke play event group 1103a, a backhand stroke play event group 1103b, and a serve play event group 1103c.
- the filter unit 405 extracts the play event groups 1103a to 1103c from the time series of the play events 1101.
- the play event groups 1103a to 1103c are extracted in parallel, and analysis processing for each play event group is executed by the feature extraction unit 407 and the determination unit 409, and analysis results for a plurality of play event types are output simultaneously. May be.
- only a part of the play event groups 1103a to 1103c may be extracted, and analysis processing for the extracted play event group may be executed by the feature extraction unit 407 and the determination unit 409.
- the play event group 1103 is extracted from the play event 1101 of a single user, but the same applies to the case of extracting the play event group from the play events of a plurality of users.
- metadata such as a user ID indicating which user's play event may be added to the play event data included in the play event group.
- FIG. 5 is a diagram showing an example of analysis for the play event group extracted in the example of FIG. Referring to FIG. 5, for each of the play event groups 1103a to 1103c extracted by the filter unit 405, the feature extraction unit 407 extracts the features of each play event, and the determination unit 409 distributes the distribution of the features of each play event (Distribution). ) To evaluate the stability.
- the feature of each play event is expressed as, for example, a multidimensional feature vector.
- such dispersion of feature vectors is expressed in a two-dimensional plane.
- a function for displaying information generated by analysis by a processor that executes analysis processing is realized on a display, such information indicating the dispersion of each play event in the feature space in a two-dimensional plane is displayed. May be.
- a technique for visualizing a multidimensional feature vector in a two-dimensional plane or a three-dimensional space is already well known, and a detailed description thereof will be omitted here.
- the distribution of the characteristics of each play event is greatly different for each play event group 1103. Therefore, the stability evaluated based on the variance is also greatly different for each play event group 1103.
- the stability may be used as an index indicating the proficiency level of play. Therefore, in the illustrated example, the proficiency level of play can be determined for each play event group 1103, that is, for each type of play event. If the level of proficiency is higher as stability is higher, the user in the illustrated example has a relatively high proficiency for forehand stroke, a relatively low proficiency for backhand stroke, and a proficiency level for serve. Presumed to be moderate.
- one of the advantages of the present embodiment is that information such as the proficiency level is generated for each type of play event, in the case of tennis, for example.
- FIG. 6 is a diagram for explaining the relationship between the distribution of the characteristics of the play events shown in the example of FIG. 5 and the proficiency level.
- the stability (Stability) of two play event groups 1103a1 and 1103a2 for the forehand stroke is evaluated based on the distribution.
- the play event group 1103a1 is a forehand stroke play event group for a senior tennis player
- the play event group 1103a2 is a forehand stroke play event group for a tennis beginner.
- the distribution of the characteristics of play events included in the play event group of the same type of shot becomes smaller as the proficiency level increases. Accordingly, the variance in the feature space of each play event included in the play event group 1103a1 of the advanced player is smaller than the variance in the feature space of each play event included in the play event group 1103a2 of the beginner. This is thought to be due to the fact that the higher the proficiency level, the smaller the variation in swing motion, and in many cases the same swing can be achieved.
- the characteristics of the play event defined based on the user's motion at the time of the shot differ depending on the type of shot. Therefore, even if the characteristic distribution of play events is calculated in a state where different types of play events (shots) are mixed, the relationship between the distribution, stability, and proficiency shown in FIG. 6 does not hold. Therefore, it is possible to evaluate the stability based on the characteristics as described above, and to determine the proficiency based on the stability.
- This embodiment is analyzed for each play event group in which the play events are classified into the same type. This is one of the advantages.
- FIG. 7 is a diagram showing another example of analysis based on the characteristics of the play event in the example of FIG.
- regions A1 to A3 in which features of play events of players (users) of various proficiency levels are distributed are defined.
- the area A1 is an area where the characteristics of the beginner player are distributed
- the area A2 is an area where the characteristics of the intermediate player are distributed
- the area A3 is an area where the characteristics of the advanced player are distributed.
- the determination unit 409 determines the proficiency level learned based on the teacher data for each proficiency level (advanced / intermediate / beginner) collected in advance for each type of play event (for example, the type of swing). This corresponds to the case where the proficiency level is judged using a vessel.
- the proficiency level determination unit is formed by learning using the distribution of the characteristics of the play event of the player of each proficiency level as teacher data. For example, the representation of a region on a two-dimensional plane as shown in FIG. 7 may also be used when a function for displaying information generated by analysis by a processor executing analysis processing on a display is realized.
- a point P indicating time series transition of a feature distribution of a certain user is shown in the feature space.
- the center of the user's feature distribution that was at the point P1 of the elementary area A1 at the beginning of service use transitions to the point P2 of the intermediate area A2 as it progresses, and further toward the advanced area A3. It is shown that it has moved and is at point P3 (still intermediate area A2).
- Such a representation of time series transition may also be used to display information generated by analysis.
- FIGS. 8 and 9 are diagrams showing further examples of the analysis result of a single user's play event in the present embodiment. Expressions of analysis results such as these examples may also be used for displaying information generated by analysis.
- a temporal change in the stability of the characteristic of the play event described above with reference to FIGS. 5 and 6 in the day is detected.
- the stability of the characteristics of the play event every hour for the day when the user was playing almost all day (approximately 10:00 to 17:00) It is expressed by a broken line from Ps to the end point Pe.
- information such as “time required for warming up” and “time until tiredness is affected by play” can be obtained from changes in the stability of play during the day. .
- the stability of the feature calculated for each type of play event may be averaged and output for each type of play event. Good. As described above, it is effective to execute analysis processing for each type of play event in order to calculate meaningful stability, but after the stability is calculated, the average exceeding the type of play event, etc. The process may be executed.
- the temporal change in the stability (Stability) of the characteristic of the play event described above with reference to FIGS. 5 and 6 is detected over a relatively long period of several months.
- the stability of the feature of the play event for each month is expressed with time as the horizontal axis. For example, it is possible to know how much play has improved in the past few months from changes in the stability of play over time. If the stability is expressed for each type of play event (shot) as in the illustrated example, it is possible to compare the progress of play corresponding to each type of play event. Similar expressions may also be used to compare stability between different users for the same type of play event.
- FIG. 10 is a diagram for describing an example of analysis of play events of a plurality of users according to an embodiment of the present disclosure.
- the processor that executes the analysis process executes the comparison process 413 based on the feature data 411 extracted by the feature extraction unit 407 from a plurality of user play events, and obtains an analysis result 415.
- the analysis result 415 can include, for example, a feature map of each player's play described below, a similar player map, and the like.
- additional information 417 for providing to the first user whose characteristic data 411 includes the characteristic data of the play event may be used.
- the result of the comparison process 413 indicates that the play is similar to the first user (similarly, the feature data of the play event is included in the feature data 411)
- the first user Additional information 417 may be provided for.
- the incidental information 417 may include, for example, advice on a specific type of play by the user (for example, forehand stroke in the case of tennis), the user's past opponents and their impressions, and the like.
- the second user whose feature data 411 indicates that the play is similar to the first user there is a high possibility that the same incidental information 417 as the first user is valid.
- FIG. 11 is a diagram for more specifically describing an example of analysis of play events of a plurality of users according to an embodiment of the present disclosure.
- the processor that executes the analysis process issues a query 419 to execute the data extraction process 423 from the motion data 421.
- An analysis result 427 is obtained from the extracted data by the analysis process 425. Further, the display 431 of the analysis result is realized by the display process 429 based on the analysis result 427.
- the type of play event to be extracted for example, for tennis, the type of shot, forehand stroke, backhand stroke, serve, etc.
- the user to be extracted are designated.
- the motion data 421 of many users is stored in the database of the storage unit 330 of the server 300, but analysis of play events in the present embodiment, for example, determination of play similarity, can be executed effectively. Since the number is limited, the target user may be limited. In addition, since motion data is also a kind of personal information, the extraction target user may be limited by limiting the disclosure target (up to friends, etc.).
- the processes of the feature extraction unit 407 and the determination unit 409 described above are executed. That is, the feature of the play event is extracted from the extracted motion data (limited to a specific type of play event by the query 419), and the similarity and the like are determined based on the extracted feature. For example, when calculating the distance in the play feature space of a plurality of players (users) in the analysis process 425, the process can be executed by a program as shown below.
- the analysis result 427 indicating the distance of each player in the play feature space may include a matrix C.
- the element C (i, j) indicates the distance in the feature space between the play of the i-th player and the play of the j-th player.
- the distance in the feature space of each user's play expressed by the matrix C is expressed in a two-dimensional plane by the process of solving the optimization problem as shown in the following Expression 1. .
- FIG. 12 and 13 are diagrams showing examples of information expressing the result of the analysis shown in FIG. 11 on a two-dimensional plane.
- the three-axis acceleration and angular velocity data are acquired using the sensor device 100 attached to the racket, and the motion of each user is obtained.
- the positional relationship of the features extracted from the data in the feature space is expressed in a two-dimensional plane.
- FIG. 12 shows the characteristics of a play event whose type (swing type) is a forehand stroke
- FIG. 13 shows the characteristics of a play event whose type is a serve.
- the representation of a region on a two-dimensional plane as shown in these drawings may also be used when a function for displaying information generated by analysis by a processor that executes analysis processing is realized on a display, for example. .
- FIG. 14 and FIG. 15 are diagrams illustrating further examples of analysis results of a plurality of users' play events in the present embodiment. Expressions of analysis results such as these examples may also be used for displaying information generated by analysis.
- the play (serve) indicated by the play event group 1103c1 for a certain user's serve is the play of any other user (for example, a professional player) to be compared. It is determined whether they are similar. In this case, the characteristics of the play event included in the user's play event group 1103c1 are compared with the characteristics of the play event included in each of the other user's play event groups 1103c2 to 1103c4, and each of the play event groups 1103c2 to 1103c4 is compared. Similarity is calculated. In the illustrated example, the similarity with the play event group 1103c3 is the highest, so that an analysis result 1105 indicating that the user is of the same type as the user of the play event group 1103c3 is output.
- an analysis result 1107 that maps the characteristics of each user's play event to a two-dimensional plane and displays it with icons indicating each user is shown. It is output.
- the analysis result 1107 is similar to the map described with reference to FIGS. 12 and 13, for example. However, by displaying it with an icon, for example, it becomes easier to understand that it is a similarity map of each user's play.
- FIG. 16 is a diagram illustrating an example of a hardware configuration of the sensor device according to the embodiment of the present disclosure.
- the sensor device 100 includes a sensor 101, a CPU (Central Processing Unit) 103, a ROM (Read Only Memory) 105, a RAM (Random Access Memory) 107, a user interface 109, and an external storage device. 111, a communication device 113, and an output device 115. These elements are connected to each other by, for example, a bus.
- a bus for example, a bus.
- the sensor 101 includes, for example, an acceleration sensor, an angular velocity sensor, a vibration sensor, a geomagnetic sensor, a temperature sensor, a pressure sensor (including a push switch), or a GPS (Global Positioning System) receiver.
- the sensor 101 may include a camera (image sensor) and a microphone (sound sensor).
- the CPU 103, the ROM 105, and the RAM 107 implement various functions in software by reading and executing program instructions recorded in the external storage device 111, for example.
- functions such as control of the entire sensor device 100 can be realized by the CPU 103, the ROM 105, and the RAM 107, for example.
- the user interface 109 is an input device such as a button or a touch panel that accepts a user operation to the sensor device 100.
- the user's operation can be, for example, an instruction to start or end transmission of sensor information from the sensor device.
- the external storage device 111 stores various types of information regarding the sensor device 100.
- the external storage device 111 may store, for example, program instructions for causing the CPU 103, the ROM 105, and the RAM 107 to realize functions in software, and the data acquired by the sensor 101 may be temporarily cached. Good.
- an external storage device 111 that is resistant to impact, such as a semiconductor memory. Note that the configuration corresponding to the internal storage area (memory or external storage device) that accumulates data detected by the sensor device 100 when the smartphone 200 is not arranged in the vicinity of the user who is playing sports is described above.
- ROM 105, RAM 107, and / or external storage device 111 is described above.
- the communication device 113 communicates with an analysis device 600 described later by various wired or wireless communication methods. Further, the communication device 113 may directly communicate with the analysis device 600 through inter-device communication, or may communicate with the analysis device 600 via a network such as the Internet.
- the output device 115 is configured by a device that can output information as light, sound, or an image.
- the output device 115 may output, for example, information notifying the detection of the time or play event in the sensor device 100, or based on the analysis result received from the analysis device 600 or the analysis result calculated in the sensor device 100. Thus, a visual or audible notification for the user may be output.
- the output device 115 includes, for example, a lamp such as an LED, a display such as an LCD, a speaker, or a vibrator.
- FIG. 17 is a diagram illustrating an example of a hardware configuration of an analysis apparatus according to an embodiment of the present disclosure.
- the analysis device 600 may realize the analysis device according to the embodiment of the present disclosure, for example, the smartphone 200 or the server 300 described above. As described above, the analysis device may be realized by the sensor device 100.
- the analysis device 600 can include a CPU 601, a ROM 603, a RAM 605, a user interface 609, an external storage device 611, a communication device 613, and an output device 615. These elements are connected to each other by, for example, a bus.
- the CPU 601, the ROM 603, and the RAM 605 realize various functions in software by reading and executing program instructions recorded in the external storage device 611, for example.
- the CPU 601, the ROM 603, and the RAM 605 can realize, for example, control of the entire analysis apparatus 600 and functions of the processing unit in the functional configuration described above.
- the user interface 609 is an input device such as a button or a touch panel that accepts a user operation to the analysis device 600.
- the external storage device 611 stores various types of information related to the analysis device 600.
- the CPU 601, the ROM 603, and the RAM 605 may store program instructions for realizing functions in software, and the sensor information received by the communication device 613 is temporarily cached in the external storage device 611. Also good.
- the external storage device 611 may store analysis result logs.
- the output device 615 is configured by a device capable of visually or audibly notifying information to the user.
- the output device 615 can be, for example, a display device such as an LCD (Liquid Crystal Display), or an audio output device such as a speaker or headphones.
- the output device 615 outputs the result obtained by the processing of the analysis device 600 as a video such as text or an image, or outputs it as a sound or sound.
- each component described above may be configured using a general-purpose member, or may be configured by hardware specialized for the function of each component. Such a configuration can be appropriately changed according to the technical level at the time of implementation.
- Embodiments of the present disclosure function, for example, an analysis apparatus (an information processing terminal such as a smartphone, a server, or a sensor apparatus), a system, an analysis apparatus or an information processing method executed by the system, and the analysis apparatus as described above And a non-transitory tangible medium on which the program is recorded.
- an analysis apparatus an information processing terminal such as a smartphone, a server, or a sensor apparatus
- a system an analysis apparatus or an information processing method executed by the system, and the analysis apparatus as described above
- a non-transitory tangible medium on which the program is recorded for example, an analysis apparatus (an information processing terminal such as a smartphone, a server, or a sensor apparatus), a system, an analysis apparatus or an information processing method executed by the system, and the analysis apparatus as described above
- a non-transitory tangible medium on which the program is recorded for example, an analysis apparatus (an information processing terminal such as a smartphone, a server, or a sensor apparatus), a system, an analysis apparatus or an information processing method executed by the
- an acquisition function for acquiring data indicating a play event defined based on a motion of a user who plays sports
- An extraction function for extracting a plurality of play events classified into the same type from the play events
- An analysis apparatus comprising a processor that realizes an analysis function for analyzing data indicating the plurality of extracted play events.
- the analysis device according to (1) wherein the analysis function analyzes characteristics of the motion corresponding to the plurality of extracted play events.
- the plurality of extracted play events are single user play events, The analysis device according to (2), wherein the analysis function evaluates the play of the single user indicated by the plurality of extracted play events based on the characteristics.
- the analysis device evaluates the stability of play of the single user based on the characteristics.
- the analysis device determines a proficiency level of play of the single user based on the stability.
- the analysis device detects a temporal change in the feature.
- the plurality of extracted play events include a plurality of user play events,
- the analysis device compares the play of the plurality of users indicated by each of the extracted plurality of play events based on the characteristics.
- the plurality of users includes a first user and a second user
- the analysis function also provides information provided to the first user regarding the play to the second user whose result of the comparison indicates that the play is similar to the first user.
- the analysis device according to (7).
- the analysis device according to any one of (1) to (8), wherein the processor further realizes a display control function for displaying information generated by the analysis on a display.
- the analysis function analyzes features of the motion corresponding to the plurality of extracted play events, The analysis device according to (9), wherein the display control function displays information indicating the feature.
- the plurality of extracted play events include a plurality of user play events, The analysis device according to (10), wherein the display control function displays information indicating the characteristics of the plurality of users on a two-dimensional plane.
- An acquisition function for acquiring data indicating a play event defined based on a motion of a user who plays sports;
- An extraction function for extracting a plurality of play events classified into the same type from the play events;
- a recording medium storing a program for causing a computer to realize an analysis function for analyzing data indicating the plurality of extracted play events.
- (13) obtaining data indicating a play event defined based on a motion of a user who plays sports; Extracting a plurality of play events classified into the same type from the play events; Analyzing the data indicating the plurality of extracted play events.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Business, Economics & Management (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Educational Technology (AREA)
- Entrepreneurship & Innovation (AREA)
- Educational Administration (AREA)
- Computing Systems (AREA)
- Bioinformatics & Computational Biology (AREA)
- Human Computer Interaction (AREA)
- Databases & Information Systems (AREA)
- Psychiatry (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Social Psychology (AREA)
- Evolutionary Biology (AREA)
- Physical Education & Sports Medicine (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
1.システム構成
2.解析処理の例
2-1.単一のユーザのプレーイベントの解析
2-2.複数のユーザのプレーイベントの解析
3.ハードウェア構成
4.補足
図1は、本開示の一実施形態に係るシステム構成の例を示す図である。図1を参照すると、システム10は、センサ装置100と、スマートフォン200と、サーバ300とを含む。
また、スマートフォン200は、必ずしも、スポーツをプレーしているユーザの近傍に配置されていなくてもよい。この場合、センサ装置100は、検出されたデータを内部の記憶領域(メモリまたは外部記憶装置)に蓄積しておく。例えば、スポーツのプレー後、センサ装置100とスマートフォン200とが接近したときに、Bluetooth(登録商標)などの無線通信によって、データがセンサ装置100からスマートフォン200に送信されてもよい。あるいは、スポーツのプレー後、センサ装置100とスマートフォン200とがUSBなどで有線接続されたときにデータが送信されてもよい。また、センサ装置100からスマートフォン200へのデータの受け渡しには、リムーバブル記録媒体が用いられてもよい。
図3は、本開示の一実施形態における解析処理の例について説明するための図である。図3を参照すると、解析処理を実行するプロセッサは、スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すモーションデータ401とメタデータ403とを入力として取得する。次に、プロセッサは、フィルタ部405の機能によって、データによって示されるプレーイベントの中から、同じ種類に分類される複数のプレーイベントを抽出する。さらに、プロセッサは、特徴抽出部407の機能によって、抽出された複数のプレーイベントに対応するデータによって示されるモーションの特徴を抽出し、判定部409の機能によって特徴に基づく判定を実施する。
図5は、図4の例において抽出されたプレーイベント群に対する解析の例を示す図である。図5を参照すると、フィルタ部405によって抽出されたプレーイベント群1103a~1103cのそれぞれについて、特徴抽出部407が各プレーイベントの特徴を抽出し、判定部409が各プレーイベントの特徴の分散(Distribution)に基づいて安定度(Stability)を評価する。
図10は、本開示の一実施形態における複数のユーザのプレーイベントの解析の例について説明するための図である。図10を参照すると、解析処理を実行するプロセッサは、複数のユーザのプレーイベントから特徴抽出部407によって抽出された特徴データ411に基づいて、比較処理413を実行し、解析結果415を得る。解析結果415は、例えば、以下で説明する各プレーヤのプレーの特徴マップや、類似プレーヤマップなどを含みうる。
for j=i+1,...,N
(プレーヤiとプレーヤjの特徴間の平均距離を算出)
(算出された距離をC(i,j)に代入)
end
end
(3.ハードウェア構成)
次に、図16および図17を参照して、本開示の実施形態に係るセンサ装置および解析装置(上述した例ではセンサ装置、スマートフォンまたはサーバ)を実現するためのハードウェア構成の例について説明する。
図16は、本開示の実施形態に係るセンサ装置のハードウェア構成の例を示す図である。図16を参照すると、センサ装置100は、センサ101と、CPU(Central Processing Unit)103と、ROM(Read Only Memory)105と、RAM(Random Access Memory)107と、ユーザインターフェース109と、外部記憶装置111と、通信装置113と、出力装置115とを含みうる。これらの要素は、例えばバスによって相互に接続される。
なお、上述した、スマートフォン200がスポーツをプレーしているユーザの近傍に配置されない場合における、センサ装置100において検出されたデータを蓄積する内部の記憶領域(メモリまたは外部記憶装置)に対応する構成は、ROM105、RAM107、および/または外部記憶装置111である。
図17は、本開示の実施形態に係る解析装置のハードウェア構成の例を示す図である。解析装置600は、本開示の実施形態に係る解析装置、例えば上記で説明したスマートフォン200またはサーバ300を実現しうる。なお、上述のように、解析装置は、センサ装置100によって実現されてもよい。
本開示の実施形態は、例えば、上記で説明したような解析装置(スマートフォンなどの情報処理端末、サーバ、またはセンサ装置)、システム、解析装置またはシステムで実行される情報処理方法、解析装置を機能させるためのプログラム、およびプログラムが記録された一時的でない有形の媒体を含みうる。
(1)スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得する取得機能と、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出する抽出機能と、
前記抽出された複数のプレーイベントを示すデータを解析する解析機能と
を実現するプロセッサを備える解析装置。
(2)前記解析機能は、前記抽出された複数のプレーイベントに対応する前記モーションの特徴を解析する、前記(1)に記載の解析装置。
(3)前記抽出された複数のプレーイベントは、単一のユーザのプレーイベントであり、
前記解析機能は、前記抽出された複数のプレーイベントによって示される前記単一のユーザのプレーを前記特徴に基づいて評価する、前記(2)に記載の解析装置。
(4)前記解析機能は、前記単一のユーザのプレーの安定度を前記特徴に基づいて評価する、前記(3)に記載の解析装置。
(5)前記解析機能は、前記単一のユーザのプレーの習熟度を前記安定度に基づいて判定する、前記(4)に記載の解析装置。
(6)前記解析機能は、前記特徴の時間的な変化を検出する、前記(3)~(5)のいずれか1項)に記載の解析装置。
(7)前記抽出された複数のプレーイベントは、複数のユーザのプレーイベントを含み、
前記解析機能は、前記抽出された複数のプレーイベントのそれぞれによって示される前記複数のユーザのプレーを前記特徴に基づいて比較する、前記(2)に記載の解析装置。
(8)前記複数のユーザは、第1のユーザおよび第2のユーザを含み、
前記解析機能は、前記プレーに関して第1のユーザに提供される情報を、前記比較の結果によって前記プレーが前記第1のユーザと類似していることが示される前記第2のユーザにも提供する、前記(7)に記載の解析装置。
(9)前記プロセッサは、前記解析によって生成された情報をディスプレイに表示させる表示制御機能をさらに実現する、前記(1)~(8)のいずれか1項)に記載の解析装置。
(10)前記解析機能は、前記抽出された複数のプレーイベントに対応する前記モーションの特徴を解析し、
前記表示制御機能は、前記特徴を示す情報を表示させる、前記(9)に記載の解析装置。
(11)前記抽出された複数のプレーイベントは、複数のユーザのプレーイベントを含み、
前記表示制御機能は、前記複数のユーザの前記特徴を2次元平面において示す情報を表示させる、前記(10)に記載の解析装置。
(12)スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得する取得機能と、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出する抽出機能と、
前記抽出された複数のプレーイベントを示すデータを解析する解析機能と
をコンピュータに実現させるためのプログラムが格納された記録媒体。
(13)スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得することと、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出することと、
前記抽出された複数のプレーイベントを示すデータを解析することと
を含む解析方法。
100 センサ装置
110 センサ
120 処理部
200 スマートフォン
210 受信部
220 処理部
300 サーバ
310 受信部
320 処理部
401 メタデータ
403 モーションデータ
405 フィルタ部
407 特徴抽出部
409 判定部
Claims (13)
- スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得する取得機能と、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出する抽出機能と、
前記抽出された複数のプレーイベントを示すデータを解析する解析機能と
を実現するプロセッサを備える解析装置。 - 前記解析機能は、前記抽出された複数のプレーイベントに対応する前記モーションの特徴を解析する、請求項1に記載の解析装置。
- 前記抽出された複数のプレーイベントは、単一のユーザのプレーイベントであり、
前記解析機能は、前記抽出された複数のプレーイベントによって示される前記単一のユーザのプレーを前記特徴に基づいて評価する、請求項2に記載の解析装置。 - 前記解析機能は、前記単一のユーザのプレーの安定度を前記特徴に基づいて評価する、請求項3に記載の解析装置。
- 前記解析機能は、前記単一のユーザのプレーの習熟度を前記安定度に基づいて判定する、請求項4に記載の解析装置。
- 前記解析機能は、前記特徴の時間的な変化を検出する、請求項3に記載の解析装置。
- 前記抽出された複数のプレーイベントは、複数のユーザのプレーイベントを含み、
前記解析機能は、前記抽出された複数のプレーイベントのそれぞれによって示される前記複数のユーザのプレーを前記特徴に基づいて比較する、請求項2に記載の解析装置。 - 前記複数のユーザは、第1のユーザおよび第2のユーザを含み、
前記解析機能は、前記プレーに関して第1のユーザに提供される情報を、前記比較の結果によって前記プレーが前記第1のユーザと類似していることが示される前記第2のユーザにも提供する、請求項7に記載の解析装置。 - 前記プロセッサは、前記解析によって生成された情報をディスプレイに表示させる表示制御機能をさらに実現する、請求項1に記載の解析装置。
- 前記解析機能は、前記抽出された複数のプレーイベントに対応する前記モーションの特徴を解析し、
前記表示制御機能は、前記特徴を示す情報を表示させる、請求項9に記載の解析装置。 - 前記抽出された複数のプレーイベントは、複数のユーザのプレーイベントを含み、
前記表示制御機能は、前記複数のユーザの前記特徴を2次元平面において示す情報を表示させる、請求項10に記載の解析装置。 - スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得する取得機能と、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出する抽出機能と、
前記抽出された複数のプレーイベントを示すデータを解析する解析機能と
をコンピュータに実現させるためのプログラムが格納された記録媒体。 - スポーツをプレーするユーザのモーションに基づいて定義されたプレーイベントを示すデータを取得することと、
前記プレーイベントの中から同じ種類に分類される複数のプレーイベントを抽出することと、
前記抽出された複数のプレーイベントを示すデータを解析することと
を含む解析方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14875753.7A EP3088058A4 (en) | 2013-12-27 | 2014-11-06 | Analysis device, recording medium, and analysis method |
JP2015554651A JP6458739B2 (ja) | 2013-12-27 | 2014-11-06 | 解析装置、記録媒体および解析方法 |
CN201480069802.6A CN105828894A (zh) | 2013-12-27 | 2014-11-06 | 分析装置、记录介质及分析方法 |
US15/033,755 US10558854B2 (en) | 2013-12-27 | 2014-11-06 | Analysis device, recording medium, and analysis method |
US16/730,168 US11042731B2 (en) | 2013-12-27 | 2019-12-30 | Analysis device, recording medium, and analysis method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-273516 | 2013-12-27 | ||
JP2013273516 | 2013-12-27 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/033,755 A-371-Of-International US10558854B2 (en) | 2013-12-27 | 2014-11-06 | Analysis device, recording medium, and analysis method |
US16/730,168 Continuation US11042731B2 (en) | 2013-12-27 | 2019-12-30 | Analysis device, recording medium, and analysis method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015098304A1 true WO2015098304A1 (ja) | 2015-07-02 |
Family
ID=53478182
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2014/079389 WO2015098304A1 (ja) | 2013-12-27 | 2014-11-06 | 解析装置、記録媒体および解析方法 |
Country Status (5)
Country | Link |
---|---|
US (2) | US10558854B2 (ja) |
EP (1) | EP3088058A4 (ja) |
JP (1) | JP6458739B2 (ja) |
CN (1) | CN105828894A (ja) |
WO (1) | WO2015098304A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017189492A (ja) * | 2016-04-15 | 2017-10-19 | セイコーエプソン株式会社 | 表示方法、スイング解析装置、スイング解析システム、スイング解析プログラム、および記録媒体 |
JP2017189490A (ja) * | 2016-04-15 | 2017-10-19 | セイコーエプソン株式会社 | 表示方法、スイング解析装置、スイング解析システム、スイング解析プログラム、および記録媒体 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013069447A1 (ja) | 2011-11-08 | 2013-05-16 | ソニー株式会社 | センサ装置、解析装置および記憶媒体 |
US20150352404A1 (en) * | 2014-06-06 | 2015-12-10 | Head Technology Gmbh | Swing analysis system |
US10706740B2 (en) * | 2014-12-24 | 2020-07-07 | Sony Corporation | System and method for processing sensor data |
JP2018093979A (ja) * | 2016-12-09 | 2018-06-21 | セイコーエプソン株式会社 | 運動診断装置、運動診断システム、プログラム、記録媒体及び運動診断方法 |
JP7385826B2 (ja) * | 2019-11-08 | 2023-11-24 | オムロン株式会社 | 動作分析装置、動作分析方法及び動作分析プログラム |
US12053671B1 (en) * | 2020-10-30 | 2024-08-06 | Diamond Kinetics, Inc. | Method and system for sports swing training |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0581232A (ja) * | 1991-09-19 | 1993-04-02 | Fujitsu Ltd | スウイング動作時の運動解析方式 |
WO2012138605A2 (en) * | 2011-04-05 | 2012-10-11 | Nike International Ltd. | Systems and methods for storing and analyzing golf data, including community and individual golf data collection and storage at a central hub |
US20120277890A1 (en) * | 2011-04-29 | 2012-11-01 | Zheng Han | Method of Ball Game Motion Recognition, Apparatus for the same, and motion assisting device |
JP2012254205A (ja) | 2011-06-09 | 2012-12-27 | Seiko Epson Corp | スイング分析装置、スイング分析プログラム、および記録媒体 |
US20130053190A1 (en) * | 2011-08-29 | 2013-02-28 | Icuemotion, Llc | Racket sport inertial sensor motion tracking and analysis |
JP2013244405A (ja) * | 2012-05-23 | 2013-12-09 | Babolat Vs | テニスラケットを使用するときの再現性指標を計算するための方法 |
Family Cites Families (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060025229A1 (en) * | 2003-12-19 | 2006-02-02 | Satayan Mahajan | Motion tracking and analysis apparatus and method and system implementations thereof |
US9229540B2 (en) * | 2004-01-30 | 2016-01-05 | Electronic Scripting Products, Inc. | Deriving input from six degrees of freedom interfaces |
US20060084516A1 (en) * | 2004-07-28 | 2006-04-20 | Smartswing, Inc. | Method and system for defining and using a reference swing for a sports training system |
US7219033B2 (en) * | 2005-02-15 | 2007-05-15 | Magneto Inertial Sensing Technology, Inc. | Single/multiple axes six degrees of freedom (6 DOF) inertial motion capture system with initial orientation determination capability |
US20070105664A1 (en) * | 2005-11-07 | 2007-05-10 | Fun Racquets, Inc. | Racquet with Entertainment and Performance Feedback |
US20070135225A1 (en) * | 2005-12-12 | 2007-06-14 | Nieminen Heikki V | Sport movement analyzer and training device |
US7536033B2 (en) * | 2006-03-03 | 2009-05-19 | Richard Albert Kirby | Portable swing analyzer |
US20080200287A1 (en) * | 2007-01-10 | 2008-08-21 | Pillar Vision Corporation | Trajectory detection and feedfack system for tennis |
WO2009020886A1 (en) * | 2007-08-03 | 2009-02-12 | Pro Tee Systems, Inc. | Golf gaming systems and methods |
US8622795B2 (en) * | 2008-12-04 | 2014-01-07 | Home Box Office, Inc. | System and method for gathering and analyzing objective motion data |
US20110021280A1 (en) * | 2009-07-27 | 2011-01-27 | Vladimir Boroda | Hitting technique by identifying ball impact points |
US8602922B2 (en) * | 2009-09-25 | 2013-12-10 | Head Technology Gmbh | Methods and apparatuses for enhancing performance in racket sports |
JP2011232445A (ja) * | 2010-04-26 | 2011-11-17 | Sony Corp | 情報処理装置、出題傾向設定方法及びプログラム |
US9626554B2 (en) * | 2010-08-26 | 2017-04-18 | Blast Motion Inc. | Motion capture system that combines sensors with different measurement ranges |
US8941723B2 (en) * | 2010-08-26 | 2015-01-27 | Blast Motion Inc. | Portable wireless mobile device motion capture and analysis system and method |
JP2012228351A (ja) * | 2011-04-26 | 2012-11-22 | Dunlop Sports Co Ltd | 打感の評価方法 |
JP5273214B2 (ja) * | 2011-06-30 | 2013-08-28 | 株式会社デンソー | 近距離無線通信装置 |
US20130018494A1 (en) * | 2011-07-14 | 2013-01-17 | Alexander Andre Amini | System and method for motion analysis and feedback with ongoing dynamic training orientation determination |
US8419560B2 (en) * | 2011-07-14 | 2013-04-16 | Alexander Andre Amini | System and method for adaptive delivery of game balls based on player-specific performance data analysis |
US9327177B2 (en) * | 2011-10-14 | 2016-05-03 | Dunlop Sports Co. Ltd. | Tennis swing analysis method |
US9717967B2 (en) * | 2011-10-14 | 2017-08-01 | Dunlop Sports Co. Ltd. | Method and computer-readable storage medium for fitting tennis racket and analysis device |
EP2592612B1 (en) * | 2011-10-14 | 2018-07-25 | Sumitomo Rubber Industries, Ltd. | Tennis swing analyzing apparatus, analyzing method, and analyzing program |
US9101812B2 (en) * | 2011-10-25 | 2015-08-11 | Aquimo, Llc | Method and system to analyze sports motions using motion sensors of a mobile device |
US9161708B2 (en) * | 2013-02-14 | 2015-10-20 | P3 Analytics, Inc. | Generation of personalized training regimens from motion capture data |
-
2014
- 2014-11-06 EP EP14875753.7A patent/EP3088058A4/en not_active Ceased
- 2014-11-06 CN CN201480069802.6A patent/CN105828894A/zh active Pending
- 2014-11-06 WO PCT/JP2014/079389 patent/WO2015098304A1/ja active Application Filing
- 2014-11-06 US US15/033,755 patent/US10558854B2/en active Active
- 2014-11-06 JP JP2015554651A patent/JP6458739B2/ja active Active
-
2019
- 2019-12-30 US US16/730,168 patent/US11042731B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0581232A (ja) * | 1991-09-19 | 1993-04-02 | Fujitsu Ltd | スウイング動作時の運動解析方式 |
WO2012138605A2 (en) * | 2011-04-05 | 2012-10-11 | Nike International Ltd. | Systems and methods for storing and analyzing golf data, including community and individual golf data collection and storage at a central hub |
US20120277890A1 (en) * | 2011-04-29 | 2012-11-01 | Zheng Han | Method of Ball Game Motion Recognition, Apparatus for the same, and motion assisting device |
JP2012254205A (ja) | 2011-06-09 | 2012-12-27 | Seiko Epson Corp | スイング分析装置、スイング分析プログラム、および記録媒体 |
US20130053190A1 (en) * | 2011-08-29 | 2013-02-28 | Icuemotion, Llc | Racket sport inertial sensor motion tracking and analysis |
JP2013244405A (ja) * | 2012-05-23 | 2013-12-09 | Babolat Vs | テニスラケットを使用するときの再現性指標を計算するための方法 |
Non-Patent Citations (4)
Title |
---|
A. AHMADI ET AL.: "Towards a wearable device for skill assessment and skillacquisition of a tennis player during the first serve", SPORTS TECHNOLOGY, 2010, vol. 2, no. 3-4, 2009, pages 129 - 136, XP055355384 * |
S. KITAHARA ET AL.: "A Proposal on Automatic Analysis Method of Tennis Play Using Movies of Tennis Match", ITE TECHNICAL REPORT, vol. 37, no. 35, August 2013 (2013-08-01), pages 51 - 54, XP008184069 * |
See also references of EP3088058A4 |
ZEPP LAB: "Zepp Tennis User Guide", Retrieved from the Internet <URL:https://zepp.zendesk.com/entries/22649719-Zepp-Tennis-User-Guide> [retrieved on 20150116] * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017189492A (ja) * | 2016-04-15 | 2017-10-19 | セイコーエプソン株式会社 | 表示方法、スイング解析装置、スイング解析システム、スイング解析プログラム、および記録媒体 |
JP2017189490A (ja) * | 2016-04-15 | 2017-10-19 | セイコーエプソン株式会社 | 表示方法、スイング解析装置、スイング解析システム、スイング解析プログラム、および記録媒体 |
Also Published As
Publication number | Publication date |
---|---|
EP3088058A4 (en) | 2017-09-06 |
US10558854B2 (en) | 2020-02-11 |
JP6458739B2 (ja) | 2019-01-30 |
CN105828894A (zh) | 2016-08-03 |
US20160253553A1 (en) | 2016-09-01 |
US11042731B2 (en) | 2021-06-22 |
US20200134303A1 (en) | 2020-04-30 |
EP3088058A1 (en) | 2016-11-02 |
JPWO2015098304A1 (ja) | 2017-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6458739B2 (ja) | 解析装置、記録媒体および解析方法 | |
WO2020233464A1 (zh) | 模型训练方法、装置、存储介质及设备 | |
AU2017331639B2 (en) | A system and method to analyze and improve sports performance using monitoring devices | |
US10769796B2 (en) | Information processing apparatus, information processing method and recording medium | |
CN111698564B (zh) | 信息推荐方法、装置、设备及存储介质 | |
JP2014183931A (ja) | 情報処理装置、センサ装置、情報処理システムおよび記録媒体 | |
WO2012063560A1 (ja) | 画像処理システム、画像処理方法、及び画像処理プログラムを記憶した記憶媒体 | |
US20160346612A1 (en) | Enhancing Exercise Through Augmented Reality | |
US10441847B2 (en) | Framework, devices, and methodologies configured to enable gamification via sensor-based monitoring of physically performed skills, including location-specific gamification | |
JP7116097B2 (ja) | ターゲット画像の検出による拡張現実経験のロック解除 | |
JP2019136493A (ja) | 運動の採点方法、システム及びプログラム | |
JP6432523B2 (ja) | 解析装置、記録媒体および解析方法 | |
JP6459979B2 (ja) | 解析装置、記録媒体および解析方法 | |
JP2016010714A (ja) | 情報処理装置、情報処理システムおよび記録媒体 | |
US20230129718A1 (en) | Biometric feedback captured during viewing of displayed content | |
WO2015098303A1 (ja) | 解析装置、記録媒体および解析方法 | |
JP2022075741A (ja) | アニメーション生成装置、アニメーション生成方法及びプログラム | |
US20130225294A1 (en) | Detecting illegal moves in a game using inertial sensors | |
WO2015098301A1 (ja) | 解析装置、記録媒体および解析方法 | |
CN117752998A (zh) | 一种数据处理方法、装置、设备及介质 | |
Gorla | Active mobile interface for smart health |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14875753 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015554651 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15033755 Country of ref document: US |
|
REEP | Request for entry into the european phase |
Ref document number: 2014875753 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014875753 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |