WO2017209260A1 - Audio training device and audio training method - Google Patents

Audio training device and audio training method Download PDF

Info

Publication number
WO2017209260A1
WO2017209260A1 PCT/JP2017/020514 JP2017020514W WO2017209260A1 WO 2017209260 A1 WO2017209260 A1 WO 2017209260A1 JP 2017020514 W JP2017020514 W JP 2017020514W WO 2017209260 A1 WO2017209260 A1 WO 2017209260A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
voice
sound
guide
posture
Prior art date
Application number
PCT/JP2017/020514
Other languages
French (fr)
Japanese (ja)
Inventor
旭 保彦
満 細尾
Original Assignee
ヤマハ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ヤマハ株式会社 filed Critical ヤマハ株式会社
Publication of WO2017209260A1 publication Critical patent/WO2017209260A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B69/00Training appliances or apparatus for special sports
    • A63B69/36Training appliances or apparatus for special sports for golf
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B71/00Games or sports accessories not covered in groups A63B1/00 - A63B69/00
    • A63B71/06Indicating or scoring devices for games or players, or for other sports activities
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/04Electrically-operated educational appliances with audible presentation of the material to be studied
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10KSOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
    • G10K15/00Acoustics not otherwise provided for
    • G10K15/04Sound-producing devices

Definitions

  • This disclosure relates to a voice learning apparatus and a voice learning method for learning a movement of a user using a voice in real time during a predetermined movement.
  • Patent Document 1 An apparatus for analyzing and evaluating a user's predetermined operation, for example, a golf swing, has been put into practical use (for example, Patent Document 1 and Patent Document 2). These devices analyze the user's swing and display the trajectory and speed after the swing to inform the user.
  • the above-described conventional apparatus does not notify the user in real time during the swing of the quality of the swing, for example, the deviation of the swing trajectory. Further, even if there is a device that displays a trajectory deviation or the like in real time during the swing, the user cannot see the display during the swing.
  • An object of this disclosure is to provide a voice learning device capable of learning a predetermined operation such as a golf swing in real time during the operation (determining the quality of the operation without looking at the display). .
  • the present disclosure includes a mounting unit that is mounted in the vicinity of the user's ear canal and includes a sound emitting unit that emits a guide voice to a user, and a sensor that detects the posture or movement of the user, and the detected user
  • a voice learning device including a guide voice generation unit that generates the guide voice determined in real time based on the posture or motion of the voice.
  • the present disclosure includes a detection step of detecting a posture or movement of a user, a generation step of generating guide voice determined in real time based on the detected posture or movement of the user, and the user during the movement of the user.
  • a sound learning method including a sound emitting step for emitting a guide sound.
  • a predetermined operation such as a golf swing can be learned in real time during the operation using a voice.
  • FIG. 1 is a diagram illustrating a usage pattern of a voice learning device (voice instructor) according to an embodiment of the present disclosure.
  • FIG. 2 is a block diagram of the speech learning apparatus.
  • FIG. 3 is a diagram illustrating an example of changes in the user's swing and head speed.
  • FIG. 4 is a diagram illustrating an example of changes in the user's swing and head speed.
  • FIG. 5 is a diagram showing a guide voice generation rule of the voice learning device.
  • FIG. 6A and FIG. 6B are flowcharts showing a guide voice generation procedure and a reference update procedure of the voice learning device.
  • FIG. 1 is a diagram showing a usage pattern of the audio learning device 1.
  • FIG. 2 is a block diagram of the speech learning apparatus 1.
  • the voice learning apparatus 1 has a mounting portion 2 that is worn on the user's ear, and a small speaker 14 and a motion sensor 15 are built in the mounting portion 2.
  • the motion sensor 15 detects the motion of the user P and the motion of the head. For example, the motion of the head during the swing motion of the golf club C is detected.
  • the voice learning device 1 detects the motion of the head of the user P by the motion sensors 15L and 15R provided in the mounting portions 2L and 2R of both ears, and the control unit 10 determines the difference between the motion and the reference with a small speaker.
  • the user P is notified by voice (guide voice) via 14L and 14R. Thereby, the user P can correct the swing in real time by listening to the guide voice while performing the swing operation.
  • the voice learning device 1 includes a mounting unit 2 (2L, 2R) mounted on both ears of a user and a control box 3 provided near the middle of a cable connecting the left and right mounting units 2L, 2R. .
  • the left and right mounting portions 2L and 2R incorporate small speakers 14L and 14R that emit sound to the user's outer ear, and motion sensors 15L and 15R that detect swing.
  • a 9-axis MEMS sensor is applicable as the motion sensors 15L and 15R.
  • the motion sensors 15L and 15R are also referred to as a left ear sensor 15L and a right ear sensor 15R.
  • a guide voice for instructing a golf swing in real time is emitted. That is, the guide voice for instructing the golf swing is emitted from the small speakers 14L and 14R while the user P is performing the golf swing.
  • the left mounting portion 2L and the control box 3, and the right mounting portion 2R and the control box 3 are connected by a stereo cable.
  • the control box 3 transmits audio (guide audio) to be heard by the user P to the left and right mounting portions 2L and 2R, and acquires detection values from the motion sensors 15L and 15R provided in the left and right mounting portions 2L and 2R.
  • the control box 3 includes a control unit 10 which is a small computer.
  • a left ear sensor 15L, a right ear sensor 15R, a memory 12, and a sound source 11 are connected to the control unit 10.
  • the control box 3 may be dedicated to the audio learning device 1, but may be realized by a multi-function mobile phone (smartphone) and an application program.
  • the mounting unit 2 and the multi-function telephone may communicate with each other via Bluetooth (registered trademark).
  • the control unit 10 estimates and analyzes the movement of the user P based on the detection values input from the left ear sensor 15L and the right ear sensor 15R, and causes the sound source 11 to generate based on the difference between the analysis result and the reference.
  • the mode of the guide voice is determined.
  • the mode of the guide voice includes beats added to the guide voice, volume, left / right balance, and localization.
  • the reference is stored in the memory 12. This guide sound mode determination process is performed, for example, every 10 milliseconds, and is output to the sound source 11 as guide sound control information. Details of determining the mode of the guide voice will be described later (see FIG. 4).
  • the sound source (sound generator) 11 generates guide voice based on the control information input from the control unit 10 and controls the change of the mode.
  • the guide sound is amplified by the left and right amplifiers (drivers) 13L and 13R and emitted from the small speakers 14L and 14R.
  • the amplifiers 13L and 13R are provided in the control box 3, but may be provided in the mounting portions 2L and 2R.
  • FIG. 3 and FIG. 4 are diagrams showing a change in posture and a change in head direction when the user P swings the golf club C.
  • FIG. FIG. 3 shows an example when the user P makes a preferable swing
  • FIG. 4 shows an example when the user P makes an unfavorable swing.
  • the moving direction of the head is the forward direction in the drawing (the hitting ball direction) and the reverse direction is the backward direction.
  • the user P when the user P swings and hits the golf club C, the user P first takes an address posture (A) and then performs a backswing from this posture (B) to create a top posture. (C).
  • A address posture
  • B backswing from this posture
  • C top posture
  • the golf club C is slowly moving (turning) in a direction opposite to the swing direction.
  • the user P starts the swing from the top posture (C) (D), and finishes the swing at the follow-through (F) through the impact (E) while accelerating.
  • FIG. 5 is a diagram illustrating rules for controlling the mode of the guide voice that the voice learning device 1 emits to the user P.
  • the speech learning apparatus 1 of this embodiment detects the movement of the head of the user P, and estimates and analyzes the movement of the whole body. Specifically, the front and back movement and turning of the user P's head are detected based on the detection values of the left and right ear sensors 15L and 15R.
  • the movement of the head of the user P represents the posture of the user P, and based on the movement of the head, it is possible to estimate the movement of the whole body such as whether the body is facing the ball or the body is not open. Is possible.
  • References which are model data relating to the movement of the head of the user P, which is the object of analysis, and the turning of the head, are stored in the memory 12 and read out in synchronization with the swing of the user P for analysis. Compared with content.
  • the reference may be data that records an ideal (teacher's) swing, or may be past data (in a good form).
  • the sound source 11 generates a sound having a predetermined frequency (for example, 440 Hz) as a guide sound.
  • a predetermined frequency for example, 440 Hz
  • a sinusoidal wave may be used, but a sawtooth wave that includes overtones and has an easy-to-understand localization may be used.
  • the control unit 10 compares the head turning angle and the movement amount detected in the address posture, the back swing posture, and the top posture with the reference, and causes the guide voice to produce an amount of beat according to the deviation.
  • the beat is a subtracted waveform of two sounds, and has a frequency corresponding to the frequency difference between the two sounds.
  • amplitude modulation AM
  • the number of beats may be increased as the deviation is larger, and the number of beats may be reduced as the deviation is reduced.
  • the depth of amplitude modulation may be adjusted. That is, the larger the deviation, the larger the beat (modulation deeper), and the smaller the deviation, the smaller the beat (modulation smaller). Since the deviation of the turning angle and the amount of movement of the head with respect to the reference is reflected in the number of times the guide voice beats, the user P may correct the swing posture so as to eliminate the beat while listening to the guide voice.
  • the coefficient is adjusted so that the number of beats is around 10 Hz as in the case of normal instrument tuning. Further, a guide 1 sound and a guide 2 sound whose frequencies are controlled may be generated so that a predetermined number of beats occur, and played simultaneously.
  • the magnitude of the beat includes both or one of the magnitude of the beat frequency and the magnitude (depth of modulation) of the beat.
  • the number of beats and the depth of beat are made to correspond to the deviation of the amount of forward and backward movement and the deviation of the turning angle, respectively.
  • the beat may be generated by amplitude-modulating the guide sound, but the basic guide sound (440 Hz) is emitted to one ear and the frequency is raised or lowered to the other ear according to the deviation of the swing trajectory.
  • the sound of the swing analysis sound (440 ⁇ b hertz) may be emitted to cause a beat in the user P's hearing.
  • the above describes the example in which the user P is informed of the deviation between the turning angle of the head and the reference of the moving amount by using the guide voice.
  • the volume balance or the localization is used.
  • the user P may be notified of the deviation. Examples thereof will be described below.
  • the control unit 10 compares the analyzed front / rear position of the head with the reference, and if there is a deviation, changes the left / right volume balance to notify the user P of the deviation. That is, when the head is shifted behind the reference, the left sound is increased to alert the left direction. If the head is displaced before the reference, the right sound is increased to alert the right. This informs the user P that the position of the head is displaced and alerts the user in the direction to return, so that the user P can reflexively correct the head position. Note that, depending on the type of sound, there is a possibility that attention can be drawn when the volume is reduced. In this case, the direction of head displacement and the direction of volume balance may be reversed.
  • the control unit 10 compares the analyzed turning of the head with the reference, and changes the localization of the guide voice when it is deviated.
  • the guide voice is usually localized on the top of the head or the face.
  • the guide voice is localized to the left rear.
  • interest is generated in the left rear, and an extra right turn can be stopped (returned to the original).
  • the guide voice is localized to the right rear.
  • interest of the user P is generated at the right rear, and an extra left turn can be stopped.
  • the above-described analysis of the detection value (motion) of the motion sensor and the mode control process of the guide voice are executed at short time intervals (for example, 10 milliseconds), and the mode of the guide voice generated by the sound source 11 is controlled in real time. While swinging the golf club C, the user P can correct the position of the address, the posture of the back swing, the posture during the swing, and the like by listening to the guide voice controlled in various modes.
  • FIG. 6A is a flowchart of the swing guide process of the control unit 10.
  • the control unit 10 repeats this process every 10 milliseconds.
  • the control unit 10 acquires the detection values of the left and right ear sensors 15L and 15R (S20), and analyzes the back and forth movement of the head and the left and right turns based on the detection values (S21). Then, the control unit 10 compares the left / right movement of the head with the reference, and determines the number of beats of the guide voice and the left / right volume balance according to the deviation (S23).
  • the control unit 10 compares the turning angle of the head with the reference, and determines the beat depth and localization position of the guide voice based on the deviation (S14). Then, the sound source 11 is instructed to generate the guide voice in the manner determined in S13, S14, S23, and S24.
  • FIG. 6B is a flowchart illustrating an example of a reference update procedure.
  • a procedure for registering a record of a past swing as a reference is shown.
  • the memory 12 has an area for storing a plurality of times of swing analysis data, analyzes the detection values of the motion sensors 15L and R for each swing of the user P (S30), and stores the analysis data of the swing (S30). S31).
  • the swing at S30 may be a practice swing using a reference that has already been registered. After swinging once or a plurality of times, the user P who has determined that there is a good swing performs a predetermined reference update operation on the speech training apparatus 1.
  • control unit 10 advances the process to the history selection process in S33.
  • the control unit 10 accepts a history selection by the user P.
  • the memory 12 can store analysis data for a plurality of past times, the user P selects the one that seems to be good. Then, the control unit 10 transfers the selected analysis data to the reference storage area (S34), and returns to the normal processing for receiving the detected value of the motion sensor by the swing.
  • control unit 10 may automatically register the stored analysis data as a reference. Further, the analysis data of a plurality of times may be stored in the memory 12, and the average value may be used as a reference.
  • the object of analysis is not limited to the address position, back swing posture, swing trajectory, swing speed, head left / right movement, and head turn. Some of these or other elements related to the operation of the user P may be analyzed.
  • the change mode of the guide voice for notifying the deviation between the analysis result and the reference in real time is not limited to the above.
  • a guide word may be generated and emitted.
  • Guide words include, for example, “Gaze remains in the ball!” And “Do not thrust the upper body!”.
  • the present disclosure can be applied to other than the golf swing.
  • a baseball bat swing, a tennis racket swing, and the analysis target are not limited to swings, and anything that can compare the posture / motion of the model and the posture / motion of the analysis target, such as dance.
  • the transmission medium of the sound emitting unit is not limited to air vibration, and any medium that transmits vibration to the human auditory organ may be used.
  • a mounting unit that is mounted near the user's ear canal, and that includes a sound emitting unit that emits a guide voice to the user, and a sensor that detects the posture or movement of the user, and the detected user's
  • a voice learning device including a guide voice generation unit that generates the guide voice determined in real time based on posture or motion.
  • the guide voice generation unit may be at least one of beat, volume, left / right balance, and localization of the guide voice based on the detected posture or action of the user. Control one.
  • the guide sound generation unit generates a pulse sound as the guide sound, and controls at least one of the interval and the volume of the pulse sound based on the detected posture or action of the user.
  • the speech training device according to any one of (1) to (3) further includes a reference memory that stores a reference of the posture or motion of the user, and the guide speech generation unit includes the detected posture of the user Alternatively, the operation is compared with the reference, and the mode of the guide voice is controlled in real time based on the comparison result.
  • the voice training device according to (1) to (4) described above is generated by the musical sound generator that generates musical sounds, the guide voice generated by the guide voice generator, and the musical sound generator.
  • the guide speech generation unit indicates a deviation between the detected user posture or motion and the reference based on the comparison result. Thus, the mode of the guide voice is controlled.
  • the sound emitting unit and the sensor are built in the mounting unit.
  • the mounting unit is an earphone.
  • the voice learning method of the present disclosure includes a detection step of detecting a posture or action of a user, a generation step of generating guide voice determined in real time based on the detected posture or action of the user, And a sound emitting step for emitting the guide sound during the user's operation.
  • the generation step at least one of beat, volume, left / right balance, and localization of the guide voice based on the detected posture or action of the user.
  • a pulse sound is generated as the guide sound, and the interval and volume of the pulse sound are determined based on the detected posture or action of the user. Control at least one.
  • the speech learning method according to (9) to (12) includes a musical sound generating step for generating a musical sound, the guide voice generated in the generating step, and the musical sound generated in the musical sound generating step.
  • the generation step may indicate a deviation between the detected user posture or motion and the reference based on the comparison result. The aspect of the guide voice is controlled.
  • a predetermined operation such as a golf swing can be learned in real time during the operation using the voice.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • General Health & Medical Sciences (AREA)
  • Physical Education & Sports Medicine (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Acoustics & Sound (AREA)
  • Electrically Operated Instructional Devices (AREA)
  • Rehabilitation Tools (AREA)

Abstract

An audio training device (1) is provided with: a wearable unit (2) worn by a user and having a sound emitting unit (14) for emitting guidance audio to the user; a sensor (15) for detecting user posture or movement; and a guidance audio generating unit (10) for generating guidance audio determined in real time on the basis of user posture or movement detected by the sensor (15).

Description

音声教習装置及び音声教習方法Voice learning apparatus and voice learning method
 この開示は、ユーザの所定の動作中にリアルタイムに音声を用いてその動作を教習する音声教習装置及び音声教習方法に関する。 This disclosure relates to a voice learning apparatus and a voice learning method for learning a movement of a user using a voice in real time during a predetermined movement.
 ユーザの所定の動作、たとえばゴルフスイングを解析および評価する装置が実用化されている(たとえば、特許文献1,特許文献2など)。これらの装置は、ユーザのスイングを解析し、スイング後に軌道や速度を表示してユーザに知らせるものである。 An apparatus for analyzing and evaluating a user's predetermined operation, for example, a golf swing, has been put into practical use (for example, Patent Document 1 and Patent Document 2). These devices analyze the user's swing and display the trajectory and speed after the swing to inform the user.
日本国特開2010-025737号公報Japanese Laid-Open Patent Publication No. 2010-025737
日本国特開2012-254205号公報Japanese Unexamined Patent Publication No. 2012-254205
 しかしながら、上記従来の装置は、スイングの良否、たとえばスイング軌道のずれなどをスイング中にリアルタイムにユーザに知らせるものではなかった。また仮に、スイング中にリアルタイムに軌道ずれなどを表示するものがあったとしても、ユーザがスイング中にその表示を見ることは不可能であった。 However, the above-described conventional apparatus does not notify the user in real time during the swing of the quality of the swing, for example, the deviation of the swing trajectory. Further, even if there is a device that displays a trajectory deviation or the like in real time during the swing, the user cannot see the display during the swing.
 この開示の目的は、ゴルフスイングなどの所定の動作を、その動作中にリアルタイムに教習する(表示を見ずに動作の良否の度合を判断する)ことができる音声教習装置を提供することにある。 An object of this disclosure is to provide a voice learning device capable of learning a predetermined operation such as a golf swing in real time during the operation (determining the quality of the operation without looking at the display). .
 本開示は、ユーザにガイド音声を放音する放音部と、前記ユーザの姿勢または動作を検出するセンサと、を有する、前記ユーザの外耳道付近に装着される装着部と、前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定された前記ガイド音声を生成するガイド音声生成部と、を備えた音声教習装置を提供する。 The present disclosure includes a mounting unit that is mounted in the vicinity of the user's ear canal and includes a sound emitting unit that emits a guide voice to a user, and a sensor that detects the posture or movement of the user, and the detected user There is provided a voice learning device including a guide voice generation unit that generates the guide voice determined in real time based on the posture or motion of the voice.
 本開示は、ユーザの姿勢または動作を検出する検出工程と、前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定されたガイド音声を生成する生成工程と、前記ユーザの前記動作中に前記ガイド音声を放音する放音工程と、を備えた音声教習方法を提供する。 The present disclosure includes a detection step of detecting a posture or movement of a user, a generation step of generating guide voice determined in real time based on the detected posture or movement of the user, and the user during the movement of the user There is provided a sound learning method including a sound emitting step for emitting a guide sound.
 本開示によれば、ゴルフスイングなどの所定の動作を、音声を用いて、その動作中にリアルタイムに教習することができる。 According to the present disclosure, a predetermined operation such as a golf swing can be learned in real time during the operation using a voice.
図1は、この開示の実施形態である音声教習装置(音声インストラクター)の使用形態を示す図である。FIG. 1 is a diagram illustrating a usage pattern of a voice learning device (voice instructor) according to an embodiment of the present disclosure. 図2は、音声教習装置のブロック図である。FIG. 2 is a block diagram of the speech learning apparatus. 図3は、ユーザのスイングとヘッドスピードの変化の例を示す図である。FIG. 3 is a diagram illustrating an example of changes in the user's swing and head speed. 図4は、ユーザのスイングとヘッドスピードの変化の例を示す図である。FIG. 4 is a diagram illustrating an example of changes in the user's swing and head speed. 図5は、音声教習装置のガイド音声生成ルールを示す図である。FIG. 5 is a diagram showing a guide voice generation rule of the voice learning device. 図6(A)及び図6(B)は、音声教習装置のガイド音声発生手順およびリファレンス更新手順を示すフローチャートである。FIG. 6A and FIG. 6B are flowcharts showing a guide voice generation procedure and a reference update procedure of the voice learning device.
 図1は、音声教習装置1の使用形態を示す図である。図2は、音声教習装置1のブロック図である。音声教習装置1は、ユーザの耳に装着される装着部2を有し、この装着部2に小型スピーカ14およびモーションセンサ15を内蔵している。モーションセンサ15は、ユーザPの動作、頭部の動作を検出する。例えば、ゴルフクラブCのスイング動作中の頭部の動作を検出する。音声教習装置1は、両耳の装着部2L,2Rに設けられたモーションセンサ15L、15RによってユーザPの頭部の動作を検出し、制御部10において、その動作とリファレンスとの差を小型スピーカ14L、14Rを介してユーザPに音声(ガイド音声)で知らせる。これにより、ユーザPは、スイング動作をしながらガイド音声を聴くことでリアルタイムにスイングを修正することができる。 FIG. 1 is a diagram showing a usage pattern of the audio learning device 1. FIG. 2 is a block diagram of the speech learning apparatus 1. The voice learning apparatus 1 has a mounting portion 2 that is worn on the user's ear, and a small speaker 14 and a motion sensor 15 are built in the mounting portion 2. The motion sensor 15 detects the motion of the user P and the motion of the head. For example, the motion of the head during the swing motion of the golf club C is detected. The voice learning device 1 detects the motion of the head of the user P by the motion sensors 15L and 15R provided in the mounting portions 2L and 2R of both ears, and the control unit 10 determines the difference between the motion and the reference with a small speaker. The user P is notified by voice (guide voice) via 14L and 14R. Thereby, the user P can correct the swing in real time by listening to the guide voice while performing the swing operation.
 図1において、ユーザPは、音声教習装置1を装着し、ゴルフクラブCを持ってアドレスの姿勢をとっている。音声教習装置1は、ユーザの両耳に装着される装着部2(2L,2R)、および、左右の装着部2L,2Rをつなぐケーブルの中間付近に設けられたコントロールボックス3を有している。左右の装着部2L,2Rには、上述したように、ユーザの外耳に音を放音する小型スピーカ14L、14R、および、揺動を検出するモーションセンサ15L、15Rが内蔵されている。モーションセンサ15L、15Rとしては、9軸のMEMSセンサが適用可能である。なお、モーションセンサ15L、15Rを左耳センサ15L、右耳センサ15Rとも呼ぶ。小型スピーカ14L、14Rからは、リアルタイムにゴルフスイングを指導するガイド音声が放音される。つまり、ゴルフスイングを指導するガイド音声が、ユーザPがゴルフスイングを行っている間に小型スピーカ14L、14Rから放音される。左装着部2Lとコントロールボックス3、および、右装着部2Rとコントロールボックス3は、それぞれステレオケーブルで接続されている。コントロールボックス3は、左右装着部2L,2RへユーザPに聴かせる音声(ガイド音声)を送信し、左右装着部2L,2Rに設けられたモーションセンサ15L,15Rから検出値を取得する。 In FIG. 1, the user P wears the voice learning device 1 and takes the address posture with the golf club C. The voice learning device 1 includes a mounting unit 2 (2L, 2R) mounted on both ears of a user and a control box 3 provided near the middle of a cable connecting the left and right mounting units 2L, 2R. . As described above, the left and right mounting portions 2L and 2R incorporate small speakers 14L and 14R that emit sound to the user's outer ear, and motion sensors 15L and 15R that detect swing. A 9-axis MEMS sensor is applicable as the motion sensors 15L and 15R. The motion sensors 15L and 15R are also referred to as a left ear sensor 15L and a right ear sensor 15R. From the small speakers 14L and 14R, a guide voice for instructing a golf swing in real time is emitted. That is, the guide voice for instructing the golf swing is emitted from the small speakers 14L and 14R while the user P is performing the golf swing. The left mounting portion 2L and the control box 3, and the right mounting portion 2R and the control box 3 are connected by a stereo cable. The control box 3 transmits audio (guide audio) to be heard by the user P to the left and right mounting portions 2L and 2R, and acquires detection values from the motion sensors 15L and 15R provided in the left and right mounting portions 2L and 2R.
 コントロールボックス3には、小型コンピュータである制御部10が内蔵されている。制御部10には、左耳センサ15L、右耳センサ15R、メモリ12、および、音源11が接続されている。なお、コントロールボックス3は、音声教習装置1専用のものであってもよいが、多機能携帯電話機(スマートフォン)とアプリケーションプログラムとで実現されてもよい。この場合、装着部2と多機能電話機とはBluetooth(登録商標)で通信してもよい。 The control box 3 includes a control unit 10 which is a small computer. A left ear sensor 15L, a right ear sensor 15R, a memory 12, and a sound source 11 are connected to the control unit 10. Note that the control box 3 may be dedicated to the audio learning device 1, but may be realized by a multi-function mobile phone (smartphone) and an application program. In this case, the mounting unit 2 and the multi-function telephone may communicate with each other via Bluetooth (registered trademark).
 制御部10は、左耳センサ15L、右耳センサ15Rから入力される検出値に基づいてユーザPの動きを推測・解析し、その解析結果とリファレンスとの差に基づいて、音源11に生成させるガイド音声の態様を決定する。ガイド音声の態様とは、ガイド音声に付加されるうなり、音量、左右バランス、定位などである。なお、リファレンスはメモリ12に記憶されている。このガイド音声の態様の決定処理は、たとえば10ミリ秒毎に行われ、音源11に対してガイド音声の制御情報として出力される。ガイド音声の態様の決定の詳細は後述する(図4参照)。音源(サウンド・ジェネレータ)11は、制御部10から入力された制御情報に基づいてガイド音声を生成し、態様を変更制御する。ガイド音声は左右のアンプ(ドライバ)13L,13Rで増幅され、小型スピーカ14L,14Rから放音される。なお、図1ではアンプ13L,13Rがコントロールボックス3に設けられているが、装着部2L,2Rに設けられてもよい。 The control unit 10 estimates and analyzes the movement of the user P based on the detection values input from the left ear sensor 15L and the right ear sensor 15R, and causes the sound source 11 to generate based on the difference between the analysis result and the reference. The mode of the guide voice is determined. The mode of the guide voice includes beats added to the guide voice, volume, left / right balance, and localization. The reference is stored in the memory 12. This guide sound mode determination process is performed, for example, every 10 milliseconds, and is output to the sound source 11 as guide sound control information. Details of determining the mode of the guide voice will be described later (see FIG. 4). The sound source (sound generator) 11 generates guide voice based on the control information input from the control unit 10 and controls the change of the mode. The guide sound is amplified by the left and right amplifiers (drivers) 13L and 13R and emitted from the small speakers 14L and 14R. In FIG. 1, the amplifiers 13L and 13R are provided in the control box 3, but may be provided in the mounting portions 2L and 2R.
 図3、図4は、ユーザPがゴルフクラブCをスイングしたときの姿勢の推移と頭部の向きの遷移を示す図である。図3はユーザPが好ましいスイングをしたときの例を示しており、図4はユーザPが好ましくないスイングをした場合の例を示している。ここで頭部の移動方向は図面右方向(打球方向)が前、その逆方向が後である。 FIG. 3 and FIG. 4 are diagrams showing a change in posture and a change in head direction when the user P swings the golf club C. FIG. FIG. 3 shows an example when the user P makes a preferable swing, and FIG. 4 shows an example when the user P makes an unfavorable swing. Here, the moving direction of the head is the forward direction in the drawing (the hitting ball direction) and the reverse direction is the backward direction.
 図3の好ましい例において、ユーザPがゴルフクラブCをスイングして打球するとき、ユーザPは、まずアドレス姿勢をとり(A)、この姿勢からバックスイングを行って(B)、トップ姿勢を作る(C)。この過程で、モーションセンサ15の検出値から頭部は、小さい角度だけ右に旋回し、頭部の前後位置はほとんど変化していないことがわかる。ゴルフクラブCはスイング方向とは逆の方向にゆっくり移動(旋回)している。ユーザPは、トップの姿勢(C)からスイングを開始し(D)、加速しながらインパクト(E)を経てフォロースルー(F)でスイングを終了する。スイングの開始時に、頭部は前に移動且つ右に旋回し、スイングにともなって、前移動、右旋回が解消されて(逆方向に移動・旋回して)ゆく。そして、インパクトの瞬間に、頭部は、正面を向き(旋回角度=0)、大きく後ろに移動している。こののち、フォロースルーとともに、頭部は、左(打球方向)に大きく旋回し、後ろ移動が解消されて若干の前移動となる。 In the preferred example of FIG. 3, when the user P swings and hits the golf club C, the user P first takes an address posture (A) and then performs a backswing from this posture (B) to create a top posture. (C). In this process, it can be seen from the detection value of the motion sensor 15 that the head turns to the right by a small angle, and the front-rear position of the head hardly changes. The golf club C is slowly moving (turning) in a direction opposite to the swing direction. The user P starts the swing from the top posture (C) (D), and finishes the swing at the follow-through (F) through the impact (E) while accelerating. At the start of the swing, the head moves forward and turns to the right, and the forward movement and right turn are eliminated (moved and turned in the opposite direction) with the swing. Then, at the moment of impact, the head is facing backwards (turning angle = 0) and is moving backward greatly. After that, along with the follow-through, the head largely turns to the left (in the direction of the hit ball), and the backward movement is canceled and the head moves slightly forward.
 図4の好ましくない例において、ユーザPがゴルフクラブCをスイングして打球するとき、ユーザPは、まずアドレス姿勢をとり(A)、この姿勢からバックスイングを行って(B)、トップ姿勢を作る(C)。ここまでの過程は、図3の好ましい例とほぼ同様である。ユーザPは、トップの姿勢(C)からスイングを開始し(D)、加速しながらインパクト(E)を経てフォロースルー(F)でスイングを終了するが、スイングの開始時から頭部は前に移動するとともに、左に旋回し始める。インパクトの瞬間でも、頭部は、前に移動したままで左(打球方向)に旋回している。そして、その姿勢のままフォロースルーとなる。 In the unfavorable example of FIG. 4, when the user P swings and hits the golf club C, the user P first takes an address posture (A), performs a backswing from this posture (B), and changes the top posture. Make (C). The process so far is almost the same as the preferred example of FIG. The user P starts swinging from the top posture (C) (D), passes through the impact (E) while accelerating and ends the swing through the follow-through (F), but the head moves forward from the start of the swing. Move and start turning left. Even at the moment of impact, the head is turning left (hit direction) while moving forward. And it will follow through with that posture.
 図5は、音声教習装置1がユーザPに対して放音するガイド音声の態様を制御するルールを説明する図である。この実施形態の音声教習装置1は、ユーザPの頭部の動作を検出し、全身の動作を推測・解析する。具体的には、左右耳センサ15L,Rの検出値に基づいてユーザPの頭部の前後の移動と旋回を検出する。ユーザPの頭部の動きは、ユーザPの姿勢を表しており、頭部の動きに基づいて、ボールの方向を向いているか、体が開いていないかなどの全身の動きを推定することが可能である。 FIG. 5 is a diagram illustrating rules for controlling the mode of the guide voice that the voice learning device 1 emits to the user P. The speech learning apparatus 1 of this embodiment detects the movement of the head of the user P, and estimates and analyzes the movement of the whole body. Specifically, the front and back movement and turning of the user P's head are detected based on the detection values of the left and right ear sensors 15L and 15R. The movement of the head of the user P represents the posture of the user P, and based on the movement of the head, it is possible to estimate the movement of the whole body such as whether the body is facing the ball or the body is not open. Is possible.
 解析の対象であるユーザPの頭部の前後の移動、および、頭部の旋回に関するお手本データであるリファレンスは、メモリ12に記憶されており、ユーザPのスイングに同期して読み出され、解析内容と比較される。リファレンスは、理想的な(先生の)スイングを記録したデータであってもよく、自分の過去の(よいフォームの)データであってもよい。 References, which are model data relating to the movement of the head of the user P, which is the object of analysis, and the turning of the head, are stored in the memory 12 and read out in synchronization with the swing of the user P for analysis. Compared with content. The reference may be data that records an ideal (teacher's) swing, or may be past data (in a good form).
 音源11は、ガイド音声として所定周波数(たとえば440ヘルツ)の音声を生成する。正弦波でもよいが、倍音を含み定位がわかりやすいノコギリ波を用いてもよい。 The sound source 11 generates a sound having a predetermined frequency (for example, 440 Hz) as a guide sound. A sinusoidal wave may be used, but a sawtooth wave that includes overtones and has an easy-to-understand localization may be used.
 制御部10は、アドレスの姿勢、バックスイングの姿勢、トップの姿勢において検出された頭部の旋回角度および移動量をリファレンスと比較し、そのずれに応じた量のうなりをガイド音声に生じさせる。うなりは、2つの音声の減算波形であり、2つの音声の周波数差に相当する周波数を有する。具体的には、ガイド音声に対して、うなりに相当する低周波信号波形で振幅変調(AM)すればよい。ずれが大きいほどうなりの回数を多くし、ずれが少なくなるにつれてうなりの回数を少なくすればよい。たとえば、以下の式を用いてうなり回数(うなり周波数)を決定することもできる。
  うなり回数= 係数×(1-(解析値/リファレンス値))×ガイド音声周波数
The control unit 10 compares the head turning angle and the movement amount detected in the address posture, the back swing posture, and the top posture with the reference, and causes the guide voice to produce an amount of beat according to the deviation. The beat is a subtracted waveform of two sounds, and has a frequency corresponding to the frequency difference between the two sounds. Specifically, amplitude modulation (AM) may be performed on the guide voice with a low frequency signal waveform corresponding to beat. The number of beats may be increased as the deviation is larger, and the number of beats may be reduced as the deviation is reduced. For example, the number of beats (beat frequency) can be determined using the following equation.
Number of beats = coefficient x (1- (analysis value / reference value)) x guide audio frequency
 また、振幅変調の深さを調整してもよい。すなわち、ずれが大きいほどうなりを大きく(変調を深く)し、ずれが少なくなるにつれてうなりを小さく(変調を小さく)してもよい。リファレンスに対する頭部の旋回角度および移動量のずれがガイド音声のうなりの回数に反映されるため、ユーザPはガイド音声を聴きながらうなりが無くなるようにスイングの姿勢を修正すればよい。うなりの回数は通常の楽器のチューニングのように10Hz前後になるよう、係数が調整されている。また、所定回数のうなりが生じるように、周波数を制御したガイド1音とガイド2音を生成し、同時に鳴らすようにしてもよい。 Also, the depth of amplitude modulation may be adjusted. That is, the larger the deviation, the larger the beat (modulation deeper), and the smaller the deviation, the smaller the beat (modulation smaller). Since the deviation of the turning angle and the amount of movement of the head with respect to the reference is reflected in the number of times the guide voice beats, the user P may correct the swing posture so as to eliminate the beat while listening to the guide voice. The coefficient is adjusted so that the number of beats is around 10 Hz as in the case of normal instrument tuning. Further, a guide 1 sound and a guide 2 sound whose frequencies are controlled may be generated so that a predetermined number of beats occur, and played simultaneously.
 上記のように、うなりの大小は、うなり周波数の高低およびうなりの深さ(変調度)の大小の両方または一方を含む。この実施形態では、うなり回数とうなりの深さをそれぞれ前後移動量のずれおよび旋回角度のずれに対応させている。うなりは、ガイド音声を振幅変調して生じさせてもよいが、一方の耳に基本のガイド音声(440ヘルツ)を放音し、他方の耳にスイング軌道のずれに合わせて周波数を上下させたスイング解析音(440±bヘルツ)の音を放音し、ユーザPの聴覚においてうなりを生じさせるようにしてもよい。 As described above, the magnitude of the beat includes both or one of the magnitude of the beat frequency and the magnitude (depth of modulation) of the beat. In this embodiment, the number of beats and the depth of beat are made to correspond to the deviation of the amount of forward and backward movement and the deviation of the turning angle, respectively. The beat may be generated by amplitude-modulating the guide sound, but the basic guide sound (440 Hz) is emitted to one ear and the frequency is raised or lowered to the other ear according to the deviation of the swing trajectory. The sound of the swing analysis sound (440 ± b hertz) may be emitted to cause a beat in the user P's hearing.
 以上は、頭部の旋回角度と移動量のリファレンスとのずれをガイド音声のうなりでユーザPに知らせる例について説明したが、うなりに代えて、または、うなりに加えて、音量バランスや定位を用いて上記ずれをユーザPに知らせてもよい。以下その例を説明する。 The above describes the example in which the user P is informed of the deviation between the turning angle of the head and the reference of the moving amount by using the guide voice. However, instead of or in addition to the beat, the volume balance or the localization is used. Thus, the user P may be notified of the deviation. Examples thereof will be described below.
 制御部10は、解析された頭部の前後位置をリファレンスと比較し、ずれている場合には左右の音量バランスを変化させてユーザPにずれを知らせる。すなわち、頭部がリファレンスよりも後ろにずれている場合には、左の音を大きくして左方向の注意を喚起する。頭部がリファレンスよりも前にずれている場合には、右の音を大きくして右方向の注意を喚起する。これにより、ユーザPに頭部の位置がずれていることを知らせるとともに、戻すべき方向へ注意を喚起するため、ユーザPは反射的に頭部位置を修正することが可能になる。なお、音の種類によっては小さくなったほうが注意を喚起できる可能性があるため、その場合には頭部のずれ方向と音量バランスの大小の方向を上記と逆にしてもよい。 The control unit 10 compares the analyzed front / rear position of the head with the reference, and if there is a deviation, changes the left / right volume balance to notify the user P of the deviation. That is, when the head is shifted behind the reference, the left sound is increased to alert the left direction. If the head is displaced before the reference, the right sound is increased to alert the right. This informs the user P that the position of the head is displaced and alerts the user in the direction to return, so that the user P can reflexively correct the head position. Note that, depending on the type of sound, there is a possibility that attention can be drawn when the volume is reduced. In this case, the direction of head displacement and the direction of volume balance may be reversed.
 制御部10は、解析された頭部の旋回をリファレンスと比較し、ずれている場合にはガイド音声の定位を変化させる。ガイド音声は、通常は頭頂または顔の前面に定位されている。ユーザPの頭部が、リファレンスよりも右に旋回している場合には、ガイド音声を左後ろに定位させる。これにより、左後ろに関心が生じ、余分な右旋回を止める(元に戻す)ことができる。逆に、頭部がリファレンスよりも左に旋回している場合には、ガイド音声を右後ろに定位させる。これにより、右後ろにユーザPの関心が生じ、余分な左旋回を止めることができる。 The control unit 10 compares the analyzed turning of the head with the reference, and changes the localization of the guide voice when it is deviated. The guide voice is usually localized on the top of the head or the face. When the user P's head is turning to the right of the reference, the guide voice is localized to the left rear. As a result, interest is generated in the left rear, and an extra right turn can be stopped (returned to the original). Conversely, when the head is turning to the left of the reference, the guide voice is localized to the right rear. As a result, interest of the user P is generated at the right rear, and an extra left turn can be stopped.
 以上のモーションセンサの検出値(動作)の解析およびガイド音声の態様制御処理は、短時間(たとえば10ミリ秒)間隔で実行され、音源11が生成するガイド音声の態様がリアルタイムに制御される。ユーザPはゴルフクラブCをスイングしながら、種々の態様に制御されたガイド音声を聴くことでアドレスの位置や、バックスイングの姿勢、スイング中の姿勢などを修正することができる。 The above-described analysis of the detection value (motion) of the motion sensor and the mode control process of the guide voice are executed at short time intervals (for example, 10 milliseconds), and the mode of the guide voice generated by the sound source 11 is controlled in real time. While swinging the golf club C, the user P can correct the position of the address, the posture of the back swing, the posture during the swing, and the like by listening to the guide voice controlled in various modes.
 図6(A)は、制御部10のスイングガイド処理のフローチャートである。制御部10は、この処理を10ミリ秒ごとに繰り返して実行する。制御部10は、左右耳センサ15L,Rの検出値を取得し(S20)、検出値に基づいて頭部の前後の動きおよび左右の旋回を解析する(S21)。そして制御部10は、頭部の左右の移動をリファレンスと比較し、そのずれに応じてガイド音声のうなり回数および左右の音量バランスを決定する(S23)。制御部10は、頭部の旋回角度をリファレンスと比較し、そのずれに基づいてガイド音声のうなり深さおよび定位位置を決定する(S14)。そして、S13、S14、S23およびS24で決定された態様でガイド音声を生成するように音源11に指示する。 FIG. 6A is a flowchart of the swing guide process of the control unit 10. The control unit 10 repeats this process every 10 milliseconds. The control unit 10 acquires the detection values of the left and right ear sensors 15L and 15R (S20), and analyzes the back and forth movement of the head and the left and right turns based on the detection values (S21). Then, the control unit 10 compares the left / right movement of the head with the reference, and determines the number of beats of the guide voice and the left / right volume balance according to the deviation (S23). The control unit 10 compares the turning angle of the head with the reference, and determines the beat depth and localization position of the guide voice based on the deviation (S14). Then, the sound source 11 is instructed to generate the guide voice in the manner determined in S13, S14, S23, and S24.
 図6(B)は、リファレンスの更新手順の一例を示すフローチャートである。この例では、過去のスイングの記録をリファレンスとして登録する手順を示している。メモリ12は複数回のスイング解析データを記憶する領域を有しており、ユーザPのスイング毎に、モーションセンサ15L,Rの検出値を解析し(S30)、そのスイングの解析データを記憶する(S31)。なお、S30のスイングは既に登録されているリファレンスを用いた練習スイングであってよい。1回または複数回スイングしたのち、良いスイングがあったと判断したユーザPは、音声教習装置1に対して所定のリファレンス更新操作を行う。リファレンス更新操作が行われると(S32でYES)、制御部10は、処理をS33の履歴選択処理に進める。S33では、制御部10はユーザPによる履歴の選択を受け付ける。メモリ12が過去複数回分の解析データを記憶可能な場合、ユーザPはそのなかからよかったと思われるものを選択する。そして、制御部10は選択された解析データをリファレンス記憶エリアに転記して(S34)、スイングによるモーションセンサの検出値を受け付ける通常処理にもどる。 FIG. 6B is a flowchart illustrating an example of a reference update procedure. In this example, a procedure for registering a record of a past swing as a reference is shown. The memory 12 has an area for storing a plurality of times of swing analysis data, analyzes the detection values of the motion sensors 15L and R for each swing of the user P (S30), and stores the analysis data of the swing (S30). S31). The swing at S30 may be a practice swing using a reference that has already been registered. After swinging once or a plurality of times, the user P who has determined that there is a good swing performs a predetermined reference update operation on the speech training apparatus 1. When the reference update operation is performed (YES in S32), the control unit 10 advances the process to the history selection process in S33. In S33, the control unit 10 accepts a history selection by the user P. When the memory 12 can store analysis data for a plurality of past times, the user P selects the one that seems to be good. Then, the control unit 10 transfers the selected analysis data to the reference storage area (S34), and returns to the normal processing for receiving the detected value of the motion sensor by the swing.
 なお、メモリ12が過去1回分の解析データを記憶し、リファレンス更新操作が行われると、制御部10は記憶していた解析データを自動的にリファレンスとして登録するようにしてもよい。また、メモリ12に複数回の解析データを記憶し、その平均値をリファレンスとしてもよい。 Note that when the memory 12 stores analysis data for the past one time and a reference update operation is performed, the control unit 10 may automatically register the stored analysis data as a reference. Further, the analysis data of a plurality of times may be stored in the memory 12, and the average value may be used as a reference.
 また、解析の対象は、アドレスの位置、バックスイングの姿勢、スイング軌道、スイング速度、頭部の左右移動および頭部の旋回に限定されない。これらの一部、または、ユーザPの動作に係る他の要素を解析してもよい。 Also, the object of analysis is not limited to the address position, back swing posture, swing trajectory, swing speed, head left / right movement, and head turn. Some of these or other elements related to the operation of the user P may be analyzed.
 また、解析結果とリファレンスとのずれをリアルタイムに告知するためのガイド音声の変化態様も上記に限定されない。また、ガイド音声に代えて、または、ガイド音声とともに、ガイド文言を生成且つ放音してもよい。ガイド文言は、たとえば「視線はボールのまま!」や「上体を突っ込まない!」などがある。 Also, the change mode of the guide voice for notifying the deviation between the analysis result and the reference in real time is not limited to the above. Further, instead of the guide voice or together with the guide voice, a guide word may be generated and emitted. Guide words include, for example, “Gaze remains in the ball!” And “Do not thrust the upper body!”.
 上記の実施形態は、ゴルフクラブCのスイングを解析且つ指導する例について説明したが、本開示はゴルフスイング以外にも適用可能である。たとえば、野球のバットスイング、テニスのラケットのスイング、また、解析対象はスイングに限定されず、舞踊等、お手本の姿勢・動作と解析対象の姿勢・動作が比較できるものなら何でもよい。また、放音部の伝達媒体は空気の振動に限定されず、人間の聴覚器官に振動を伝えるものであれば何でも良い。 Although the above embodiment has described an example of analyzing and instructing the swing of the golf club C, the present disclosure can be applied to other than the golf swing. For example, a baseball bat swing, a tennis racket swing, and the analysis target are not limited to swings, and anything that can compare the posture / motion of the model and the posture / motion of the analysis target, such as dance. Further, the transmission medium of the sound emitting unit is not limited to air vibration, and any medium that transmits vibration to the human auditory organ may be used.
 ここで、本開示の実施形態を以下のようにまとめる。
(1)ユーザにガイド音声を放音する放音部と、前記ユーザの姿勢または動作を検出するセンサと、を有する、前記ユーザの外耳道付近に装着される装着部と、前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定された前記ガイド音声を生成するガイド音声生成部と、を備えた音声教習装置を提供する。
(2)上記(1)に記載の音声教習装置において、前記ガイド音声生成部は、前記検出されたユーザの姿勢または動作に基づいて、前記ガイド音声のうなり、音量、左右バランスおよび定位のうち少なくとも一つを制御する。
(3)上記(1)に記載の音声教習装置において、
 前記ガイド音声生成部は、前記ガイド音声としてパルス音を生成し、前記検出されたユーザの姿勢または動作に基づいて前記パルス音の間隔および音量の少なくとも一方を制御する。
(4)上記(1)乃至(3)に記載の音声教習装置は、前記ユーザの姿勢または動作のリファレンスを記憶するリファレンスメモリをさらに備え、前記ガイド音声生成部は、前記検出されたユーザの姿勢または動作を前記リファレンスと比較し、比較結果に基づいて前記ガイド音声の態様をリアルタイムに制御する。
(5)上記(1)乃至(4)に記載の音声教習装置は、楽音を発生する楽音発生部と、前記ガイド音声生成部によって生成されたガイド音声、および、前記楽音発生部によって発生された楽音を、ミキシングし、ミキシングされた音声を前記放音部に出力するミキサと、をさらに備える。
(6)上記(4)乃至(5)に記載の音声教習装置において、前記ガイド音声生成部は、前記比較結果に基づいて、前記検出されたユーザの姿勢または動作と前記リファレンスとのずれを示すように前記ガイド音声の態様を制御する。
(7)上記(1)乃至(6)に記載の音声教習装置において、前記放音部及び前記センサが、前記装着部に内蔵されている。
(8)上記(1)乃至(7)に記載の音声教習装置において、前記装着部はイヤホンである。
(9)本開示の音声教習方法は、ユーザの姿勢または動作を検出する検出工程と、前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定されたガイド音声を生成する生成工程と、前記ユーザの前記動作中に前記ガイド音声を放音する放音工程と、を備える。
(10)上記(9)に記載の音声教習方法において、前記生成工程において、前記検出されたユーザの姿勢または動作に基づいて、前記ガイド音声のうなり、音量、左右バランスおよび定位のうち少なくとも一つを制御する。
(11)上記(9)に記載の音声教習方法において、前記生成工程において、前記ガイド音声としてパルス音を生成し、前記検出されたユーザの姿勢または動作に基づいて前記パルス音の間隔および音量の少なくとも一方を制御する。
(12)上記(9)乃至(11)に記載の音声教習方法において、前記ユーザの姿勢または動作のリファレンスがリファレンスメモリに記憶されており、前記生成工程において、前記検出されたユーザの姿勢または動作を前記リファレンスと比較し、比較結果に基づいて前記ガイド音声の態様をリアルタイムに制御する。
(13)上記(9)乃至(12)に記載の音声教習方法は、楽音を発生する楽音発生工程と、前記生成工程において生成された前記ガイド音声と、前記楽音発生工程において発生された前記楽音とをミキシングするミキシング工程と、ミキシングされた音声を放音する放音工程と、をさらに備える。
(14)上記(12)乃至(13)に記載の音声教習方法において、前記生成工程において、前記比較結果に基づいて、前記検出されたユーザの姿勢または動作と前記リファレンスとのずれを示すように前記ガイド音声の態様を制御する。
Here, the embodiments of the present disclosure are summarized as follows.
(1) A mounting unit that is mounted near the user's ear canal, and that includes a sound emitting unit that emits a guide voice to the user, and a sensor that detects the posture or movement of the user, and the detected user's Provided is a voice learning device including a guide voice generation unit that generates the guide voice determined in real time based on posture or motion.
(2) In the voice training device according to (1), the guide voice generation unit may be at least one of beat, volume, left / right balance, and localization of the guide voice based on the detected posture or action of the user. Control one.
(3) In the speech training apparatus according to (1) above,
The guide sound generation unit generates a pulse sound as the guide sound, and controls at least one of the interval and the volume of the pulse sound based on the detected posture or action of the user.
(4) The speech training device according to any one of (1) to (3) further includes a reference memory that stores a reference of the posture or motion of the user, and the guide speech generation unit includes the detected posture of the user Alternatively, the operation is compared with the reference, and the mode of the guide voice is controlled in real time based on the comparison result.
(5) The voice training device according to (1) to (4) described above is generated by the musical sound generator that generates musical sounds, the guide voice generated by the guide voice generator, and the musical sound generator. A mixer that mixes the musical sound and outputs the mixed sound to the sound emitting unit;
(6) In the speech training apparatus according to (4) to (5), the guide speech generation unit indicates a deviation between the detected user posture or motion and the reference based on the comparison result. Thus, the mode of the guide voice is controlled.
(7) In the voice training device according to (1) to (6), the sound emitting unit and the sensor are built in the mounting unit.
(8) In the audio learning device according to (1) to (7), the mounting unit is an earphone.
(9) The voice learning method of the present disclosure includes a detection step of detecting a posture or action of a user, a generation step of generating guide voice determined in real time based on the detected posture or action of the user, And a sound emitting step for emitting the guide sound during the user's operation.
(10) In the audio learning method according to (9), in the generation step, at least one of beat, volume, left / right balance, and localization of the guide voice based on the detected posture or action of the user. To control.
(11) In the audio learning method according to (9), in the generation step, a pulse sound is generated as the guide sound, and the interval and volume of the pulse sound are determined based on the detected posture or action of the user. Control at least one.
(12) In the audio learning method according to (9) to (11) above, a reference of the posture or motion of the user is stored in a reference memory, and the posture or motion of the user detected in the generation step Is compared with the reference, and the mode of the guide voice is controlled in real time based on the comparison result.
(13) The speech learning method according to (9) to (12) includes a musical sound generating step for generating a musical sound, the guide voice generated in the generating step, and the musical sound generated in the musical sound generating step. A mixing step for mixing the sound and a sound emission step for emitting the mixed sound.
(14) In the audio learning method according to (12) to (13), the generation step may indicate a deviation between the detected user posture or motion and the reference based on the comparison result. The aspect of the guide voice is controlled.
 本出願は、2016年6月1日に出願された日本特許出願(特願2016-109687)に基づくものであり、ここに参照として取り込まれる。 This application is based on a Japanese patent application (Japanese Patent Application No. 2016-109687) filed on June 1, 2016, and is incorporated herein by reference.
 本開示によれば、ゴルフスイングなどの所定の動作を、音声を用いて、その動作中にリアルタイムに教習することができるため有用である。 According to the present disclosure, it is useful because a predetermined operation such as a golf swing can be learned in real time during the operation using the voice.
1 音声教習装置
2(2L、2R) 装着部
15L,15R モーションセンサ
 
1 Voice learning device 2 (2L, 2R) Mounting part 15L, 15R Motion sensor

Claims (14)

  1.  ユーザにガイド音声を放音する放音部と、前記ユーザの姿勢または動作を検出するセンサと、を有する、前記ユーザの外耳道付近に装着される装着部と、
     前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定された前記ガイド音声を生成するガイド音声生成部と、
     を備えた音声教習装置。
    A mounting unit that is mounted near the ear canal of the user, and includes a sound emitting unit that emits a guide voice to the user, and a sensor that detects the posture or operation of the user;
    A guide voice generating unit that generates the guide voice determined in real time based on the detected posture or motion of the user;
    Voice learning device with
  2.  前記ガイド音声生成部は、前記検出されたユーザの姿勢または動作に基づいて、前記ガイド音声のうなり、音量、左右バランスおよび定位のうち少なくとも一つを制御する請求項1に記載の音声教習装置。 The voice training device according to claim 1, wherein the guide voice generation unit controls at least one of beat, volume, left / right balance, and localization of the guide voice based on the detected posture or action of the user.
  3.  前記ガイド音声生成部は、前記ガイド音声としてパルス音を生成し、前記検出されたユーザの姿勢または動作に基づいて前記パルス音の間隔および音量の少なくとも一方を制御する請求項1に記載の音声教習装置。 The voice lesson according to claim 1, wherein the guide voice generation unit generates a pulse sound as the guide voice and controls at least one of an interval and a volume of the pulse sound based on the detected posture or action of the user. apparatus.
  4.  前記ユーザの姿勢または動作のリファレンスを記憶するリファレンスメモリをさらに備え、
     前記ガイド音声生成部は、前記検出されたユーザの姿勢または動作を前記リファレンスと比較し、比較結果に基づいて前記ガイド音声の態様をリアルタイムに制御する請求項1乃至3のいずれか1項に記載の音声教習装置。
    A reference memory for storing a reference of the posture or movement of the user;
    The said guide audio | voice production | generation part compares the detected user's attitude | position or operation | movement with the said reference, and controls the aspect of the said guide audio | voice in real time based on a comparison result. Voice learning device.
  5.  楽音を発生する楽音発生部と、
     前記ガイド音声生成部によって生成されたガイド音声、および、前記楽音発生部によって発生された楽音を、ミキシングし、ミキシングされた音声を前記放音部に出力するミキサと、
     をさらに備えた請求項1乃至4のいずれか1項に記載の音声教習装置。
    A musical sound generator for generating musical sounds;
    A mixer that mixes the guide sound generated by the guide sound generation unit and the musical sound generated by the musical sound generation unit and outputs the mixed sound to the sound emission unit;
    The speech learning apparatus according to claim 1, further comprising:
  6.  前記ガイド音声生成部は、前記比較結果に基づいて、前記検出されたユーザの姿勢または動作と前記リファレンスとのずれを示すように前記ガイド音声の態様を制御する請求項4乃至5のいずれか1項に記載の音声教習装置。 6. The guide voice generation unit according to any one of claims 4 to 5, wherein the guide voice generation unit controls an aspect of the guide voice so as to indicate a deviation between the detected posture or action of the user and the reference based on the comparison result. The audio training device according to item.
  7.  前記放音部及び前記センサが、前記装着部に内蔵されている請求項1乃至6のいずれか1項に記載の音声教習装置。 The voice learning device according to any one of claims 1 to 6, wherein the sound emitting unit and the sensor are built in the mounting unit.
  8.  前記装着部はイヤホンである請求項1乃至7のいずれか1項に記載の音声教習装置。 The voice learning device according to any one of claims 1 to 7, wherein the mounting portion is an earphone.
  9.  ユーザの姿勢または動作を検出する検出工程と、
     前記検出されたユーザの姿勢または動作に基づいてリアルタイムに決定されたガイド音声を生成する生成工程と、
     前記ユーザの前記動作中に前記ガイド音声を放音する放音工程と、
     を備えた音声教習方法。
    A detection process for detecting the posture or movement of the user;
    Generating a guide voice determined in real time based on the detected posture or motion of the user;
    A sound emission step of emitting the guide voice during the operation of the user;
    Voice learning method with
  10.  前記生成工程において、前記検出されたユーザの姿勢または動作に基づいて、前記ガイド音声のうなり、音量、左右バランスおよび定位のうち少なくとも一つを制御する請求項9に記載の音声教習方法。 10. The speech learning method according to claim 9, wherein in the generation step, at least one of beat, volume, left / right balance, and localization of the guide voice is controlled based on the detected posture or action of the user.
  11.  前記生成工程において、前記ガイド音声としてパルス音を生成し、前記検出されたユーザの姿勢または動作に基づいて前記パルス音の間隔および音量の少なくとも一方を制御する請求項9に記載の音声教習方法。 10. The speech learning method according to claim 9, wherein in the generation step, a pulse sound is generated as the guide sound, and at least one of the interval and the volume of the pulse sound is controlled based on the detected posture or action of the user.
  12.  前記ユーザの姿勢または動作のリファレンスがリファレンスメモリに記憶されており、
     前記生成工程において、前記検出されたユーザの姿勢または動作を前記リファレンスと比較し、比較結果に基づいて前記ガイド音声の態様をリアルタイムに制御する請求項9乃至11のいずれか1項に記載の音声教習方法。
    A reference of the user's posture or motion is stored in a reference memory;
    The voice according to any one of claims 9 to 11, wherein in the generation step, the detected user's posture or motion is compared with the reference, and the mode of the guide voice is controlled in real time based on the comparison result. Teaching method.
  13.  楽音を発生する楽音発生工程と、
     前記生成工程において生成された前記ガイド音声と、前記楽音発生工程において発生された前記楽音とをミキシングするミキシング工程と、
     ミキシングされた音声を放音する放音工程と、
     をさらに備えた請求項9乃至12のいずれか1項に記載の音声教習方法。
    A musical sound generation process for generating musical sounds;
    A mixing step of mixing the guide voice generated in the generation step and the musical sound generated in the musical sound generation step;
    A sound emission process for emitting the mixed sound;
    The speech learning method according to claim 9, further comprising:
  14.  前記生成工程において、前記比較結果に基づいて、前記検出されたユーザの姿勢または動作と前記リファレンスとのずれを示すように前記ガイド音声の態様を制御する請求項12乃至13のいずれか1項に記載の音声教習方法。 The method according to any one of claims 12 to 13, wherein in the generation step, a mode of the guide voice is controlled so as to indicate a deviation between the detected user posture or motion and the reference based on the comparison result. The audio teaching method described.
PCT/JP2017/020514 2016-06-01 2017-06-01 Audio training device and audio training method WO2017209260A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2016-109687 2016-06-01
JP2016109687A JP2017213240A (en) 2016-06-01 2016-06-01 Voice instruction device

Publications (1)

Publication Number Publication Date
WO2017209260A1 true WO2017209260A1 (en) 2017-12-07

Family

ID=60477614

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/020514 WO2017209260A1 (en) 2016-06-01 2017-06-01 Audio training device and audio training method

Country Status (2)

Country Link
JP (1) JP2017213240A (en)
WO (1) WO2017209260A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019024550A (en) * 2017-07-25 2019-02-21 株式会社クオンタム Detection device, detection system, processing device, detection method and detection program

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04309383A (en) * 1991-04-08 1992-10-30 Kubota Corp Ball hitting action training machine
JPH05115585A (en) * 1991-10-28 1993-05-14 Sugino Mach Ltd Golf swing analyzing device
JP2009125507A (en) * 2007-11-27 2009-06-11 Panasonic Electric Works Co Ltd Golf improvement support device
JP2009233092A (en) * 2008-03-27 2009-10-15 Yamaha Corp Exercise support system and program
US20110021318A1 (en) * 2009-07-20 2011-01-27 Joanna Lumsden Audio feedback for motor control training
JP2015150134A (en) * 2014-02-13 2015-08-24 株式会社ユピテル Sway detector, sway detection system, and sway detection program

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04309383A (en) * 1991-04-08 1992-10-30 Kubota Corp Ball hitting action training machine
JPH05115585A (en) * 1991-10-28 1993-05-14 Sugino Mach Ltd Golf swing analyzing device
JP2009125507A (en) * 2007-11-27 2009-06-11 Panasonic Electric Works Co Ltd Golf improvement support device
JP2009233092A (en) * 2008-03-27 2009-10-15 Yamaha Corp Exercise support system and program
US20110021318A1 (en) * 2009-07-20 2011-01-27 Joanna Lumsden Audio feedback for motor control training
JP2015150134A (en) * 2014-02-13 2015-08-24 株式会社ユピテル Sway detector, sway detection system, and sway detection program

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Ongaku wa Mochiron Workout no Kiroku Nado nimo Tsukaeru Wireless Earphone", THE DASH '', GIGAZINE, 20 March 2014 (2014-03-20), Retrieved from the Internet <URL:http://gigazine.net/news/20140320-the-dash> [retrieved on 20170726] *
"Sony, PGS ya Shinpakusu Sensor o Tosai shita Bosui Earphone Ittaigata Walkman o CES ni Shutten", GIZMODO· JAPAN, 12 January 2015 (2015-01-12), XP055445004, Retrieved from the Internet <URL:http://www.gizmodo.jp/2015/01/sony_smart_b_trainer.html> [retrieved on 20170726] *
KOJI SHIROTA: "Improvement of the Golf Swing Using Biofeedback", HOSEI UNIVERSITY GRADUATE SCHOOL KOGAKU KENKYUKA KIYO, vol. 55, 24 March 2014 (2014-03-24), pages 1 - 4, XP055444999 *
TAKANOBU OZAWA: "CES]CES Unveiled - Parrot, Sports Muke Koseino Earphone 'Zik Sport'/ Headphone ya Wearable Device ga Tasu Shutten", PHILE WEB, 5 January 2015 (2015-01-05), XP055445001, Retrieved from the Internet <URL:http://www.phileweb.com/news/d-av/201501/05/36151.html> [retrieved on 20170726] *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019024550A (en) * 2017-07-25 2019-02-21 株式会社クオンタム Detection device, detection system, processing device, detection method and detection program
JP7011416B2 (en) 2017-07-25 2022-01-26 株式会社クオンタム Detection device, detection system, processing device and detection program

Also Published As

Publication number Publication date
JP2017213240A (en) 2017-12-07

Similar Documents

Publication Publication Date Title
JP6322830B2 (en) Information processing apparatus, information processing program, information processing system, and information processing method
JP4626087B2 (en) Musical sound control system and musical sound control device
JP3867515B2 (en) Musical sound control system and musical sound control device
JP6461850B2 (en) Simulation system and program
KR101415944B1 (en) Virtual golf simulation device and method for providing stereophonic sound for whether condition
US10878718B2 (en) System and method for synchronizing audio, movement, and patterns
JP6419932B1 (en) Program for supporting performance of musical instrument in virtual space, method executed by computer to support selection of musical instrument, and information processing apparatus
JP2005103241A (en) Input device, game system, program, and information storage medium
US8656307B2 (en) Information storage medium, computer terminal, and change method
US6224386B1 (en) Sound field simulation method and sound field simulation apparatus
JPWO2019163260A1 (en) Information processing equipment, information processing methods, and programs
US11120780B2 (en) Emulation of at least one sound of a drum-type percussion instrument
WO2017209260A1 (en) Audio training device and audio training method
JP2005034195A (en) Lesson support system and method
JP2021119993A (en) Sway detector and sway detection program
JP2019200257A (en) Learning support system and program
JP4731168B2 (en) Program, information storage medium, and game system
JP7476930B2 (en) Vibration Sensor
US20210319715A1 (en) Information processing apparatus, information processing method, and program
JP2017136308A (en) Vocal training device
JP2018064216A (en) Force sense data development apparatus, electronic apparatus, force sense data development method and control program
JP4420713B2 (en) PROGRAM, INFORMATION STORAGE MEDIUM, AND GAME DEVICE
KR101471712B1 (en) Volume control apparatus
KR101875140B1 (en) A system of decrease for impact sound by golf ball
TWI816120B (en) Tempo practice module

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17806810

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17806810

Country of ref document: EP

Kind code of ref document: A1