WO2021137534A1 - Method and system for learning korean pronunciation via voice analysis - Google Patents

Method and system for learning korean pronunciation via voice analysis Download PDF

Info

Publication number
WO2021137534A1
WO2021137534A1 PCT/KR2020/019110 KR2020019110W WO2021137534A1 WO 2021137534 A1 WO2021137534 A1 WO 2021137534A1 KR 2020019110 W KR2020019110 W KR 2020019110W WO 2021137534 A1 WO2021137534 A1 WO 2021137534A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
pronunciation
evaluation
korean
learning
Prior art date
Application number
PCT/KR2020/019110
Other languages
French (fr)
Korean (ko)
Inventor
송진주
Original Assignee
(주)헤이스타즈
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by (주)헤이스타즈 filed Critical (주)헤이스타즈
Publication of WO2021137534A1 publication Critical patent/WO2021137534A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/20Education
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/06Foreign languages
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers
    • G09B7/02Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student
    • G09B7/04Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student characterised by modifying the teaching programme in response to a wrong answer, e.g. repeating the question, supplying a further explanation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/005Language recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present invention relates to a method and system for learning Korean pronunciation, and more particularly, based on data such as the learner's voice, nationality, gender, age, etc., allowing foreigners to intensively learn words or sentences that are easy to make mistakes when learning Korean. It relates to a Korean pronunciation learning system through voice analysis that can maximize the learning effect.
  • the human hearing is adjusted to better hear sounds that are often heard, and not to hear sounds that are not usually heard well.
  • the 'middle ear' of the human body structure blocks sounds that are not normally heard in order to protect the ears. It becomes blunt to sound in a frequency band that you don't normally hear.
  • 'Patent Document 1' analyzes the pronunciation of the words or sentences learned by the learner, and envelops the stress and rhythm. By analyzing the voice waveform with the voice waveform and displaying the result, it is possible to objectively analyze pronunciation, stress, rhythm, and mouth shape.
  • Patent Document 1 is merely to check the learning performance by analyzing the frequency characteristics of the words or sentences learned by the learner, and a method to improve the learning efficiency by improving the listening ability for a specific language before learning is not disclosed for
  • the audio signal of the first frequency band is transitioned to the audio signal of the second frequency band, and the user is familiar with a specific frequency band. It allows people to better hear foreign languages that mainly use different frequency bands.
  • Patent Document 2 while the frequency band of the foreign language is shifted, the tone or tone of the sentence heard by the learner is different from the original sentence. In this case, there is a problem that the learner's listening ability may be lowered even though the foreign language sentence can be heard well at that moment.
  • the present invention is based on the learner's voice, nationality, gender, age, etc., so that foreigners can intensively learn words or sentences that are easy to make mistakes when learning Korean based on data such as voice analysis to maximize the learning effect. It's about learning systems.
  • the user uploads personal information to the system
  • the system provides Korean video content to the user
  • the system presents the pronunciation evaluation problem to the user and then requests feedback
  • user voice data is collected in response to the feedback.
  • the user's voice data and the evaluation score are stored in a user terminal
  • the stored evaluation score is collected in a server with a predetermined period
  • the server is the collected information It provides a method for learning Korean pronunciation through voice analysis, including correcting a pronunciation evaluation problem provided to a user using
  • the user's pronunciation evaluation is performed by STT (Speak to text) technology.
  • the pronunciation evaluation is determined by whether the STT server can accurately recognize how many letters in Korean when a foreigner's pronunciation is input.
  • the evaluation score is collected anonymously according to the user's gender, residence area, nationality, and age.
  • the system classifies the user level after presenting the pronunciation evaluation problem for the level test to the user, the level corresponds to the same level to the divided user, and at least one of gender, residence area, nationality, and age Two or more items present a pronunciation evaluation problem with a low score for the same users.
  • the pronunciation evaluation score data of the user is copied and aggregated in the main server and collected anonymously according to the user's gender, residence area, nationality, age, etc., and then frequently incorrect words, sentences, expressions according to country, age, gender, etc. and so on, and it enables big data analysis by quantifying how many users pronounce incorrectly among the total number of pronunciation attempts.
  • 1 is a conceptual diagram of a system for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
  • FIG. 2 is a flowchart of a method for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
  • FIG. 3 is a conceptual diagram illustrating an embodiment of content that is customized to a learner through big data analysis.
  • FIG. 4 is a conceptual diagram illustrating an embodiment of a pronunciation pattern analysis for providing customized content.
  • the user uploads personal information to the system
  • the system provides Korean video content to the user
  • the system presents the pronunciation evaluation problem to the user and then requests feedback
  • user voice data is collected in response to the feedback.
  • the user's voice data and the evaluation score are stored in a user terminal
  • the stored evaluation score is collected in a server with a predetermined period
  • the server is the collected information It provides a method for learning Korean pronunciation through voice analysis, including correcting a pronunciation evaluation problem provided to a user using
  • 1 is a conceptual diagram of a system for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
  • Learners may be provided with language learning content through the user terminal.
  • User terminals include mobile phones, smart phones, laptop computers, digital broadcasting terminals, personal digital assistants (PDA), portable multimedia players (PMPs), navigation systems, slate PCs, and tablet PCs.
  • PDA personal digital assistants
  • PMPs portable multimedia players
  • PC tablet PCs.
  • ultrabooks wearable devices, for example, watch-type terminals (smartwatch), glass-type terminals (smart glass), mobile terminals such as HMD (head mounted display), digital TV, desktop A fixed terminal such as a computer, digital signage, and the like may be included.
  • the system 100 may include a control module 110 , a data collection module 120 , a database 130 , a sentence generation module 140 , an analysis module 150 , and the like. .
  • control module 110 is formed to include functions of other modules.
  • system 100 may be a specific server including the modules or a user terminal including the modules.
  • content may be delivered to a user terminal, or analysis and content execution may be performed in a user terminal in which an application program (application) is installed.
  • application application
  • control module can improve the Korean listening/speaking ability by providing the read request sentence generated through the analysis module, the sentence generation module, etc. to the learner.
  • the data collection module 120 may collect data for sentences exposed online and supply it to the user terminal.
  • the database 130 stores various data for driving the system 100 .
  • a plurality of voice data corresponding to words may be stored in the database 130 .
  • the database 130 may store information on a frequency band used for each language, information on a frequency band used in a specific fingerprint, and the like.
  • evaluation point data 131 evaluated by users according to the pronunciation feedback provided by the system and the voice data 131 stored in the user terminal by the users according to the feedback of the system may be stored in the database.
  • the sentence generation module 140 generates a read request sentence provided to the learner based on the user's personal information or the big data analysis result stored in the database.
  • FIG. 2 is a flowchart of a method for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
  • the method for learning Korean pronunciation includes the steps of providing Korean video content and requesting pronunciation feedback from the user (S100), evaluating the user's pronunciation through the STT system (S200), and sending the user to the user terminal. Storing pronunciation and evaluation data of (S300), collecting data of users in the main server (S400), analyzing the collected big data, and then providing customized for each user (S500), etc. can do.
  • Learners can enter personal information such as their nationality, gender, and age before starting learning.
  • the system provides customized learning content to individual learners by analyzing the learner's personal information and learning-related big data stored in the server (or database).
  • the Korean video content collected through the data collection module 120 or the Korean video content stored in the user terminal through the application is provided to the user. is provided
  • the system After the learner watches the video, the system requests the learner for pronunciation feedback on the sentence generated by the sentence generating module 140 .
  • step (S200) of evaluating the user's pronunciation through the STT system foreign users who use the application using STT (Speak to Text) technology provided by global pronunciation evaluation systems (Google, MINDs Lab, etc.) Evaluate how accurately you pronounce
  • Pronunciation evaluation criteria is when the foreigner's pronunciation input through the microphone of the user's terminal is transmitted to the server through the STT API and then the foreigner's pronunciation is inputted through the STT server after the read request sentence presented by the application using the STT system.
  • a method of numerically evaluating how many characters can be accurately recognized as Hangul is used.
  • the application stores the voice data pronounced by the foreign user in the user's mobile device.
  • the system creates a voice data storage space that supports the user to listen to their pronunciation again and a pronunciation evaluation score data storage space that stores pronunciation evaluation scores, and stores the original sound data every time a foreign user repeatedly trains pronunciation. compacted in space.
  • the evaluation score is also stored in the history type at the same time in the score management space according to the pronunciation point and sentence. This data is used to listen again and correct pronunciation according to the history.
  • the evaluation score data stored in the user terminal is transmitted to the server.
  • evaluation score data use of anonymous data to improve application functions and sentence recommendation service
  • the user's Pronunciation score data is replicated and aggregated in the main server.
  • This evaluation data may be collected anonymously according to the user's gender, residence area, nationality, age, and the like.
  • the system presents to the learner a sentence containing the pronunciation (ex. 'fate') that Vietnamese women in their 20s often make wrong through the data analysis collected in step S300. (See Figure 3).
  • the system can determine if the selected word is followed by another word, if the selected word is followed by a proposition, if the selected word comes at the end of a sentence, if the selected word is placed in an interrogative sentence, if the selected word is placed in an exclamation sentence, etc.
  • Various types of sentences can be presented to learners.
  • the system it is possible for the system to provide learning content by analyzing the pronunciation pattern of a language mainly used by the learner.
  • FIG. 4 is a conceptual diagram illustrating an embodiment of pronunciation pattern analysis for providing customized content.
  • ⁇ t means the main period of the bullish pattern.
  • the system analyzes the main frequency band and stress pattern of the language used by the user as a native language, and it is possible to reinforce the learning of sentence components that are different from those of Korean.
  • Koreans usually use sounds of 500 to 2200 Hz, sounds between 800 and 3500 Hz used in English (American) between 2200 and 3500 Hz that do not overlap are not accurately heard or difficult to distinguish.
  • the system can enhance the learning of words or sentences outside the frequency band mainly used in the learner's native language.
  • the type of language is a syllable language, in which syllables are simply arranged without a rhyme repeating at regular intervals (there is no stress cycle in a sentence), and a rhyme with a certain pattern (the stress of a certain period is repeated in a sentence). appear) can be divided into existing stress language.
  • the length of a syllable is different depending on the situation, but the interval at which stress (stress) is given is constant.
  • the time interval between syllables giving stress is similar, and the syllables in between are compressed and pronounced to match the time interval that used to be included.
  • the analysis module including a language type analysis module and a stress pattern analysis module, analyzes which country the sentence corresponds to, whether the language corresponds to a stress language, and how the stress pattern is in the sentence.
  • the analysis module determines what kind of language the sentence is made in based on data such as the type of language used by foreigners, words, and word order, and the language belongs to the sealable language based on the collected data about the language. You can determine whether or not you belong to a stress language.
  • the analysis module analyzes the voice output data of the sentence and analyzes the position where the speaker cuts and reads the sentence, the output dB value, tone or intonation, etc. to determine the stress pattern of the sentence the learner wants to listen to.
  • the period of the stress pattern may be slightly different for each user or sentence, but the system can identify the period that occurs with the average or the highest frequency of the periods of the collected foreign language sentences as the main period.
  • the system may arrange important words in positions corresponding to these main cycles and provide them to learners so that the learners can learn Korean sentences accustomed to them.
  • the pronunciation evaluation score data of the user is copied and aggregated in the main server, collected anonymously according to the user's gender, residence area, nationality, age, etc., and then according to country, age, gender, etc. It can be divided into frequently incorrect words, sentences, and expressions, and it enables big data analysis by quantifying how many users pronounce incorrectly among the total number of pronunciation attempts by users.
  • An effect different from the prior art can be expected, such as being able to intensively train pronunciation, sentences and expressions that are difficult in the user environment.

Abstract

The present invention provides a method for learning Korean pronunciation via voice analysis, the method comprising the steps of: uploading, by a user, personal information to a system; providing, by the system, Korean image content to the user, presenting pronunciation evaluation questions, and requesting feedback; collecting user voice data in response to the feedback and then evaluating the pronunciation of the user; storing, in a user terminal, the user voice data and scores of the evaluation; collecting, in a server, the stored scores of the evaluation during a certain period; and correcting, by the server, the pronunciation evaluation questions provided to the user, by using the collected information.

Description

음성 분석을 통한 한국어 발음 학습 방법 및 시스템Method and system for learning Korean pronunciation through voice analysis
본 발명은 한국어 발음 학습 방법 및 시스템에 관한 것으로서, 보다 자세하게는 학습자의 음성, 국적, 성별, 나이 등의 데이터를 토대로 외국인들이 한국어 학습을 할 때 틀리기 쉬운 단어 또는 문장을 집중적으로 학습할 수 있게 하여 학습 효과를 극대화 할 수 있는 음성 분석을 통한 한국어 발음 학습 시스템에 관한 것이다.The present invention relates to a method and system for learning Korean pronunciation, and more particularly, based on data such as the learner's voice, nationality, gender, age, etc., allowing foreigners to intensively learn words or sentences that are easy to make mistakes when learning Korean. It relates to a Korean pronunciation learning system through voice analysis that can maximize the learning effect.
세계에는 다양한 언어가 존재하며 언어마다 강세, 억양, 성조 등 서로 다른 특징을 갖기 때문에 그 사람이 살아온 환경에 따라 잘 듣고 말할 수 있는 단어(문장)와 잘 듣지 못하고 말하기도 어려운 단어(문장)이 존재한다. There are various languages in the world, and each language has different characteristics such as stress, intonation, tone, etc., so there are words (sentences) that you can hear and speak well and words that are difficult to hear (sentences) depending on the environment you live in. do.
이는 인종에 따른 귀의 해부학적인 구조 뿐 아니라 후천적인 상황에도 많은 영향을 받는 것으로 언어 사용 환경에 따라 사람들은 특정 발음을 더 잘 듣거나 말할 수 있게 될 수 있다. This is influenced not only by the anatomical structure of the ear according to race but also by acquired circumstances, and depending on the language environment, people may be able to hear or speak certain pronunciations better.
후천적인 변화는 사람의 생활 환경에 기인한다. 즉, 사람의 청각은 자주 듣는 소리는 더 잘 듣도록 하고, 평소 잘 듣지 않는 소리는 잘 들리지 않도록 조절한다. 예를 들어, 사람의 신체구조 중 '중이'는 평소 잘 듣지 않는 소리는 귀를 보호하기 위해 차단한다. 평소 잘 듣지 않는 주파수 대역의 소리에는 무뎌지게 되는 것이다. Acquired changes are caused by people's living environment. In other words, the human hearing is adjusted to better hear sounds that are often heard, and not to hear sounds that are not usually heard well. For example, the 'middle ear' of the human body structure blocks sounds that are not normally heard in order to protect the ears. It becomes blunt to sound in a frequency band that you don't normally hear.
언어마다 또는 사람마다 말을 할 때 발음이나 억양이 서로 다르기 때문에 상기와 같은 특성을 이용하여 듣기 능력을 향상시키는 것이 가능하다. Since the pronunciation or intonation is different for each language or for each person speaking, it is possible to improve the listening ability by using the above characteristics.
대한민국 등록특허 제10-0405061호 '언어 학습장치 및 그것의 언어 분석방법'(이하 '특허문헌 1'이라 함)에서는 학습자가 학습한 단어 또는 문장에 대한 발음을 분석하며, 강세와 리듬을 포락선을 갖는 보이스 파형으로 분석하여 그 결과를 표시해줌으로써, 발음과 강세, 리듬 및 입 모양 등을 객관적으로 분석할 수 있게 한다. Republic of Korea Patent No. 10-0405061 'Language learning apparatus and its language analysis method' (hereinafter referred to as 'Patent Document 1') analyzes the pronunciation of the words or sentences learned by the learner, and envelops the stress and rhythm. By analyzing the voice waveform with the voice waveform and displaying the result, it is possible to objectively analyze pronunciation, stress, rhythm, and mouth shape.
하지만, 특허문헌 1에 개시된 내용은 학습자가 학습한 단어 또는 문장의 주파수 특성을 분석하여 학습 성과를 확인하는 것에 불과할 뿐, 학습 전 특정 언어에 대한 듣기 능력을 향상시켜 학습 효율을 향상시킬 수 있는 방안에 대해서는 개시되어 있지 않다. However, the content disclosed in Patent Document 1 is merely to check the learning performance by analyzing the frequency characteristics of the words or sentences learned by the learner, and a method to improve the learning efficiency by improving the listening ability for a specific language before learning is not disclosed for
또한, 대한민국 등록특허 제10-1983772호 '듣기 학습 시스템'(이하, '특허문헌 2'라 함)에서는 제1 주파수 대역의 오디오 신호를 제2 주파수 대역의 오디오 신호로 천이 시켜 특정 주파수 대역에 익숙한 사람들이 다른 주파수 대역을 주로 사용하는 외국어를 잘 들을 수 있게 한다. In addition, in Republic of Korea Patent Registration No. 10-1983772 'Listening Learning System' (hereinafter referred to as 'Patent Document 2'), the audio signal of the first frequency band is transitioned to the audio signal of the second frequency band, and the user is familiar with a specific frequency band. It allows people to better hear foreign languages that mainly use different frequency bands.
하지만, 특허문헌 2에 따르면 외국어의 주파수 대역을 천이 시키면서 학습자가 듣는 문장의 어감이나 어조가 원래의 문장과는 달라져 버리게 된다. 이 경우 그 순간 외국어 문장을 잘 들을 수는 있어도 학습자의 듣기 능력이 더 낮아질 수 있다는 문제가 있다.However, according to Patent Document 2, while the frequency band of the foreign language is shifted, the tone or tone of the sentence heard by the learner is different from the original sentence. In this case, there is a problem that the learner's listening ability may be lowered even though the foreign language sentence can be heard well at that moment.
본 발명은 학습자의 음성, 국적, 성별, 나이 등의 데이터를 토대로 외국인들이 한국어 학습을 할 때 틀리기 쉬운 단어 또는 문장을 집중적으로 학습할 수 있게 하여 학습 효과를 극대화 할 수 있는 음성 분석을 통한 한국어 발음 학습 시스템에 관한 것이다. The present invention is based on the learner's voice, nationality, gender, age, etc., so that foreigners can intensively learn words or sentences that are easy to make mistakes when learning Korean based on data such as voice analysis to maximize the learning effect. It's about learning systems.
본 발명은 사용자가 개인정보를 시스템에 업로드하는 단계와, 시스템이 사용자에게 한국 영상 컨텐츠를 제공한 후 발음 평가 문제를 제시한 후 피드백을 요청하는 단계와, 상기 피드백에 대응하여 사용자 음성 데이터가 수집된 후 사용자의 발음이 평가되는 단계와, 사용자 단말기에 상기 사용자 음성 데이터 및 상기 평가 점수가 저장되는 단계와, 일정 주기를 가지고 상기 저장된 평가 점수가 서버에 취합되는 단계와, 서버가 상기 취합된 정보를 이용하여 사용자에게 제공하는 발음 평가 문제를 보정하는 단계를 포함하는 음성 분석을 통한 한국어 발음 학습 방법을 제공한다. In the present invention, the user uploads personal information to the system, the system provides Korean video content to the user, and the system presents the pronunciation evaluation problem to the user and then requests feedback, and user voice data is collected in response to the feedback. After the user's pronunciation is evaluated, the user's voice data and the evaluation score are stored in a user terminal, the stored evaluation score is collected in a server with a predetermined period, and the server is the collected information It provides a method for learning Korean pronunciation through voice analysis, including correcting a pronunciation evaluation problem provided to a user using
본 발명의 일 실시예에 따르면, 상기 사용자 발음 평가는 STT(Speak to text) 기술에 의하여 이루어진다. According to an embodiment of the present invention, the user's pronunciation evaluation is performed by STT (Speak to text) technology.
본 발명의 일 실시예에 따르면, 상기 발음 평가는 STT 서버가 외국인의 발음 입력 시 몇 글자를 정확하게 한글로 인식할 수 있는지 여부에 의하여 결정된다. According to an embodiment of the present invention, the pronunciation evaluation is determined by whether the STT server can accurately recognize how many letters in Korean when a foreigner's pronunciation is input.
본 발명의 일 실시예에 따르면, 상기 평가 점수는 사용자의 성별, 거주지역, 국적, 연령에 따라 무기명으로 취합된다. According to an embodiment of the present invention, the evaluation score is collected anonymously according to the user's gender, residence area, nationality, and age.
본 발명의 일 실시예에 따르면, 시스템은 사용자에게 레벨 테스트용 발음 평가 문제를 제시한 후 사용자 레벨을 구분하며, 레벨이 구분된 사용자에게 동일 레벨에 해당하며 성별, 거주지역, 국적, 연령 중 적어도 2이상의 항목이 동일한 사용자들에게 낮은 점수를 받은 발음 평가 문제를 제시한다.According to an embodiment of the present invention, the system classifies the user level after presenting the pronunciation evaluation problem for the level test to the user, the level corresponds to the same level to the divided user, and at least one of gender, residence area, nationality, and age Two or more items present a pronunciation evaluation problem with a low score for the same users.
본 발명에 따르면 사용자의 발음평가 점수 데이터가 메인서버에 복제되고 집계되어 사용자의 성별, 거주지역, 국적, 연령 등에 따라 무기명으로 취합된 후 국가별, 연령별, 성별 등에 따라 자주 틀리는 단어, 문장, 표현 등으로 구분될 수 있으며 전체 이용자들의 발음 시도 회수 중 몇 명의 사용자가 틀리게 발음하는지를 수치화하여 빅데이터 분석을 가능하게 한다. According to the present invention, the pronunciation evaluation score data of the user is copied and aggregated in the main server and collected anonymously according to the user's gender, residence area, nationality, age, etc., and then frequently incorrect words, sentences, expressions according to country, age, gender, etc. and so on, and it enables big data analysis by quantifying how many users pronounce incorrectly among the total number of pronunciation attempts.
또한, 본 발명은 사용자의 국적, 성별, 나이 등을 정확하게 입력하면 비슷한 사용자 이용 환경에서 어려워하는 발음, 문장과 표현 등을 집중 훈련할 수 있는 효과가 있다. In addition, according to the present invention, if the user's nationality, gender, age, and the like are accurately input, it is possible to intensively train pronunciation, sentences and expressions that are difficult in a similar user environment.
도 1은 본 발명의 일 실시예에 따른 음성 분석을 통한 한국어 발음 학습 시스템의 개념도.1 is a conceptual diagram of a system for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
도 2는 본 발명의 일 실시예에 따른 음성 분석을 통한 한국어 발음 학습 방법의 순서도. 2 is a flowchart of a method for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
도 3은 빅데이터 분석을 통해 학습자에게 맞춤형으로 제공되는 컨텐츠의 일 실시예를 나타내는 개념도. 3 is a conceptual diagram illustrating an embodiment of content that is customized to a learner through big data analysis.
도 4는 맞춤형 컨텐츠 제공을 위한 발음 패턴 분석의 일 실시예를 나타내는 개념도. 4 is a conceptual diagram illustrating an embodiment of a pronunciation pattern analysis for providing customized content.
본 발명은 사용자가 개인정보를 시스템에 업로드하는 단계와, 시스템이 사용자에게 한국 영상 컨텐츠를 제공한 후 발음 평가 문제를 제시한 후 피드백을 요청하는 단계와, 상기 피드백에 대응하여 사용자 음성 데이터가 수집된 후 사용자의 발음이 평가되는 단계와, 사용자 단말기에 상기 사용자 음성 데이터 및 상기 평가 점수가 저장되는 단계와, 일정 주기를 가지고 상기 저장된 평가 점수가 서버에 취합되는 단계와, 서버가 상기 취합된 정보를 이용하여 사용자에게 제공하는 발음 평가 문제를 보정하는 단계를 포함하는 음성 분석을 통한 한국어 발음 학습 방법을 제공한다.In the present invention, the user uploads personal information to the system, the system provides Korean video content to the user, and the system presents the pronunciation evaluation problem to the user and then requests feedback, and user voice data is collected in response to the feedback. After the user's pronunciation is evaluated, the user's voice data and the evaluation score are stored in a user terminal, the stored evaluation score is collected in a server with a predetermined period, and the server is the collected information It provides a method for learning Korean pronunciation through voice analysis, including correcting a pronunciation evaluation problem provided to a user using
이하, 본 발명에 대하여 도면을 참조하여 보다 상세하게 설명한다. 본 명세서에서는 서로 다른 실시예라도 동일·유사한 구성에 대해서는 동일·유사한 참조번호를 부여하고, 그 설명은 처음 설명으로 갈음한다. 본 명세서에서 사용되는 단수의 표현은 문맥상 명백하게 다르게 뜻하지 않는 한, 복수의 표현을 포함한다.Hereinafter, the present invention will be described in more detail with reference to the drawings. In the present specification, the same and similar reference numerals are assigned to the same and similar components in different embodiments, and the description is replaced with the first description. As used herein, the singular expression includes the plural expression unless the context clearly dictates otherwise.
또한, 이하의 설명에서 사용되는 구성요소에 대한 접미사 "모듈" 및 "부"는 명세서 작성의 용이함만이 고려되어 부여되거나 혼용되는 것으로서, 그 자체로 서로 구별되는 의미 또는 역할을 갖는 것은 아니다.In addition, the suffixes "module" and "part" for the components used in the following description are given or mixed in consideration of only the ease of writing the specification, and do not have a meaning or role distinct from each other by themselves.
도 1은 본 발명의 일 실시예에 따른 음성 분석을 통한 한국어 발음 학습 시스템의 개념도이다. 1 is a conceptual diagram of a system for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
본 발명의 시스템에 따르면 학습자에게 맞춤형 발음 학습 컨텐츠를 제공하여 외국인 학습자가 한국어에 대한 듣기/말하기 능력을 향상시킬 수 있게 한다. According to the system of the present invention, customized pronunciation learning contents are provided to learners so that foreign learners can improve their listening/speaking skills in Korean.
학습자들은 사용자 단말기를 통해 언어 학습 컨텐츠를 제공받을 수 있다. 사용자 단말기에는 휴대폰, 스마트 폰(smart phone), 노트북 컴퓨터(laptop computer), 디지털방송용 단말기, PDA(personal digital assistants), PMP(portable multimedia player), 네비게이션, 슬레이트 PC(slate PC), 태블릿 PC(tablet PC), 울트라북(ultrabook), 웨어러블 디바이스(wearable device, 예를 들어, 워치형 단말기 (smartwatch), 글래스형 단말기 (smart glass), HMD(head mounted display)) 등의 이동 단말기나 디지털 TV, 데스크탑 컴퓨터, 디지털 사이니지 등과 같은 고정 단말기가 포함될 수 있다. Learners may be provided with language learning content through the user terminal. User terminals include mobile phones, smart phones, laptop computers, digital broadcasting terminals, personal digital assistants (PDA), portable multimedia players (PMPs), navigation systems, slate PCs, and tablet PCs. PC), ultrabooks, wearable devices, for example, watch-type terminals (smartwatch), glass-type terminals (smart glass), mobile terminals such as HMD (head mounted display), digital TV, desktop A fixed terminal such as a computer, digital signage, and the like may be included.
도 1을 참조하면, 시스템(100)은 제어모듈(110)과, 데이터수집모듈(120)과, 데이터베이스(130)와, 문장 생성 모듈(140), 분석모듈(150) 등을 포함할 수 있다. Referring to FIG. 1 , the system 100 may include a control module 110 , a data collection module 120 , a database 130 , a sentence generation module 140 , an analysis module 150 , and the like. .
이하에서는 설명의 편의를 위해 각 모듈을 구분하여 설명하고 있지만 실제 제어모듈(110)이 다른 모듈들을 기능을 포함하도록 형성되는 것이 가능하다. Hereinafter, each module is separately described for convenience of description, but it is possible that the actual control module 110 is formed to include functions of other modules.
도 1에서는 시스템이 서버, 사용자 단말기와 분리된 것으로 도시되어 있지만, 본 발명에서 지칭하는 시스템(100)은 상기 모듈들을 포함하는 특정 서버 또는 상기 모듈들을 포함하는 사용자 단말기일 수 있다. Although the system is illustrated as being separated from the server and the user terminal in FIG. 1 , the system 100 referred to in the present invention may be a specific server including the modules or a user terminal including the modules.
예를 들어, 중앙 서버를 통해 분석이 이루어진 후 사용자 단말기로 컨텐츠가 전달되거나, 응용프로그램(어플리케이션)이 설치된 사용자 단말기에서 분석 및 컨텐츠 실행이 이루어질 수도 있다.For example, after analysis is performed through a central server, content may be delivered to a user terminal, or analysis and content execution may be performed in a user terminal in which an application program (application) is installed.
본 발명에서는 제어모듈이 분석모듈, 문장 생성 모듈 등을 통해 생성된 읽기 요청 문장을 학습자에게 제공하여 한국어 듣기/말하기 능력을 향상시킬 수 있다. In the present invention, the control module can improve the Korean listening/speaking ability by providing the read request sentence generated through the analysis module, the sentence generation module, etc. to the learner.
데이터수집모듈(120)은 온라인 상에 노출되는 문장에 대한 데이터를 수집하여 사용자 단말기에 공급할 수 있다. The data collection module 120 may collect data for sentences exposed online and supply it to the user terminal.
데이터베이스(130)에는 시스템(100)을 구동시키기 위한 각종 데이터들이 저장된다. 예를 들어, 데이터베이스(130)에는 단어들에 대응되는 복수의 음성 데이터들이 저장될 수 있다. 또한, 데이터베이스(130)에는 언어별 사용 주파수 대역에 대한 정보와, 특정 지문에서 사용되는 주파수 대역에 대한 정보 등이 저장될 수 있다. The database 130 stores various data for driving the system 100 . For example, a plurality of voice data corresponding to words may be stored in the database 130 . Also, the database 130 may store information on a frequency band used for each language, information on a frequency band used in a specific fingerprint, and the like.
또한, 데이터베이스에는 사용자들이 시스템이 제공하는 발음 피드백에 따라 평가 받은 평가 점수 데이터(131)와, 사용자들이 시스템의 피드백에 따라 사용자 단말기에 저장해 두었던 음성 데이터(131)도 저장될 수 있다. In addition, the evaluation point data 131 evaluated by users according to the pronunciation feedback provided by the system and the voice data 131 stored in the user terminal by the users according to the feedback of the system may be stored in the database.
문장 생성 모듈(140)은 사용자 개인 정보 또는 데이터베이스에 보관된 빅데이터 분석 결과를 토대로 학습자에게 제공하는 따라 읽기 요청 문장을 생성한다. The sentence generation module 140 generates a read request sentence provided to the learner based on the user's personal information or the big data analysis result stored in the database.
도 2는 본 발명의 일 실시예에 따른 음성 분석을 통한 한국어 발음 학습 방법의 순서도이다. 2 is a flowchart of a method for learning Korean pronunciation through voice analysis according to an embodiment of the present invention.
도 2를 참조하면, 한국어 발음 학습 방법은 한국 영상 컨텐츠를 제공하고 사용자에게 발음 피드백을 요청하는 단계(S100)와, STT 시스템을 통해 사용자의 발음을 평가하는 단계(S200)와, 사용자 단말기에 사용자의 발음과 평가 데이터를 저장하는 단계(S300)와, 메인서버에 사용자들의 데이터들을 취합하는 단계(S400)와, 취합된 빅데이터를 분석한 후 사용자별로 커스터마이징 하여 제공하는 단계(S500) 등을 포함할 수 있다. Referring to FIG. 2 , the method for learning Korean pronunciation includes the steps of providing Korean video content and requesting pronunciation feedback from the user (S100), evaluating the user's pronunciation through the STT system (S200), and sending the user to the user terminal. Storing pronunciation and evaluation data of (S300), collecting data of users in the main server (S400), analyzing the collected big data, and then providing customized for each user (S500), etc. can do.
학습자는 학습을 시작하기 전 본인의 국적, 성별, 나이 등의 개인 정보를 입력할 수 있다. Learners can enter personal information such as their nationality, gender, and age before starting learning.
시스템은 학습자 개인정보와 서버(또는 데이터베이스)에 저장된 학습 관련 빅데이터를 분석하여 학습자 개인에게 커스터마이징 된 학습 컨텐츠를 제공한다. The system provides customized learning content to individual learners by analyzing the learner's personal information and learning-related big data stored in the server (or database).
한국어 발음 학습 방법은 한국 영상 컨텐츠를 제공하고 사용자에게 발음 피드백을 요청하는 단계(S100)에서는 데이터수집모듈(120)을 통해 수집된 한국어 영상 컨텐츠 또는 어플리케이션을 통해 사용자 단말기에 저장된 한국어 영상 컨텐츠가 사용자에게 제공된다. In the Korean pronunciation learning method, in the step (S100) of providing Korean video content and requesting pronunciation feedback from the user, the Korean video content collected through the data collection module 120 or the Korean video content stored in the user terminal through the application is provided to the user. is provided
학습자가 영상을 시청한 후 시스템은 학습자에게 문장 생성 모듈(140)이 생성한 문장에 대한 발음 피드백을 요청한다. After the learner watches the video, the system requests the learner for pronunciation feedback on the sentence generated by the sentence generating module 140 .
STT 시스템을 통해 사용자의 발음을 평가하는 단계(S200)에서는 글로벌 발음 평가 시스템 (구글, 마인즈랩 등 전문 AI 개발사) 들이 제공하는 STT(Speak to Text) 기술을 이용하여 어플리케이션을 사용하는 외국인 사용자들이 한국어를 얼마나 정확하게 발음하는지를 평가한다. In the step (S200) of evaluating the user's pronunciation through the STT system, foreign users who use the application using STT (Speak to Text) technology provided by global pronunciation evaluation systems (Google, MINDs Lab, etc.) Evaluate how accurately you pronounce
발음 평가 기준은 사용자 단말기의 마이크를 통하여 입력된 외국인의 발음을 STT 시스템을 이용하여 어플리케이션에서 제시한 따라 읽기 요청 문장을 STT API 를 통해 서버로 전송한 후 STT 서버를 통해 외국인의 발음을 입력하였을 때 몇 글자를 정확하게 한글로 인식할 수 있는지를 수치로 평가하는 방법을 이용한다.Pronunciation evaluation criteria is when the foreigner's pronunciation input through the microphone of the user's terminal is transmitted to the server through the STT API and then the foreigner's pronunciation is inputted through the STT server after the read request sentence presented by the application using the STT system. A method of numerically evaluating how many characters can be accurately recognized as Hangul is used.
사용자 단말기에 사용자의 발음과 평가 데이터를 저장하는 단계(S300)에서 어플리케이션은 외국인 사용자가 발음한 음성 데이터를 사용자의 모바일 기기 안에 보관한다. In the step (S300) of storing the user's pronunciation and evaluation data in the user terminal, the application stores the voice data pronounced by the foreign user in the user's mobile device.
시스템은 사용자가 본인 발음 다시 듣기를 지원하는 음성 데이터 저장 공간 및 발음 평가 점수를 저장하는 발음 평가 점수 데이터 저장 공간을 생성하고, 외국인 사용자가 발음을 반복하여 훈련할 때마다 원음 음성 데이터를 음성 데이터 저장 공간에 압축 보관한다. The system creates a voice data storage space that supports the user to listen to their pronunciation again and a pronunciation evaluation score data storage space that stores pronunciation evaluation scores, and stores the original sound data every time a foreign user repeatedly trains pronunciation. compacted in space.
평가 점수 역시 발음 시점 및 문장에 맞춰 점수 관리 공간에 동시에 히스토리형으로 저장된다. 이 데이터는 히스토리에 맞춰 본인이 다시 들어보고 발음을 교정하는데 사용된다. The evaluation score is also stored in the history type at the same time in the score management space according to the pronunciation point and sentence. This data is used to listen again and correct pronunciation according to the history.
메인서버에 사용자들의 데이터들을 취합하는 단계(S400)에서는 사용자 단말기에 저장된 평가 점수 데이터가 서버로 전송된다. In the step of collecting data of users in the main server (S400), the evaluation score data stored in the user terminal is transmitted to the server.
어플리케이션 설치 시 평가 점수 데이터의 통계적 사용(어플리케이션 기능 및 문장 추천 서비스 개선을 위한 무기명 데이터의 활용) 에 대한 사용자의 동의를 사전에 받게 되며, 특정 시점(Ex. Wi-Fi 환경 접속 시)에 사용자의 발음평가 점수 데이터가 메인서버에 복제되고 집계된다. 이 평가 데이터는 사용자의 성별, 거주지역, 국적, 연령 등에 따라 무기명으로 취합될 수 있다. Upon application installation, the user's consent for statistical use of evaluation score data (use of anonymous data to improve application functions and sentence recommendation service) is obtained in advance, and the user's Pronunciation score data is replicated and aggregated in the main server. This evaluation data may be collected anonymously according to the user's gender, residence area, nationality, age, and the like.
취합된 빅데이터를 분석한 후 사용자별로 커스터마이징 하여 제공하는 단계(S500)에서는 사용자의 개인정보에 따라 최적화된 컨텐츠를 제공한다. After analyzing the collected big data, customized for each user and provided (S500) provides optimized content according to the user's personal information.
예를 들어, 학습자가 베트남 국적의 20대 여성이라면 시스템은 S300 단계에서 취합된 데이터 분석을 통해 베트남 국적의 20대 여성들이 자주 틀리는 발음(ex.'운명')을 포함하는 문장을 학습자에게 제시한다(도 3참조). For example, if the learner is a woman in her 20s of Vietnamese nationality, the system presents to the learner a sentence containing the pronunciation (ex. 'fate') that Vietnamese women in their 20s often make wrong through the data analysis collected in step S300. (See Figure 3).
또한, 본 발명의 일 실시예에 따르면 자주 틀리는 단어가 다른 형태로 들어가 있는 문장들을 순차적으로 제시하여 발음의 변화 방식을 체득할 수 있게 할 수 있다. 예를 들어, 시스템은 선택된 단어의 뒤에 다른 단어가 오는 경우, 선택된 단어 뒤에 조사가 오는 경우, 선택된 단어가 문장의 마지막에 오는 경우, 선택된 단어가 의문문에 들어간 경우, 선택된 단어가 감탄문에 들어간 경우 등 다양한 형태의 문장을 학습자에게 제시할 수 있다. In addition, according to an embodiment of the present invention, it is possible to sequentially present sentences in which frequently incorrect words are included in different forms to acquire a pronunciation change method. For example, the system can determine if the selected word is followed by another word, if the selected word is followed by a proposition, if the selected word comes at the end of a sentence, if the selected word is placed in an interrogative sentence, if the selected word is placed in an exclamation sentence, etc. Various types of sentences can be presented to learners.
또한, 본 발명의 다른 실시예에 따르면 시스템은 학습자가 주로 사용하는 언어의 발음 패턴을 분석하여 학습 컨텐츠를 제공하는 것이 가능하다. In addition, according to another embodiment of the present invention, it is possible for the system to provide learning content by analyzing the pronunciation pattern of a language mainly used by the learner.
도 4는 맞춤형 컨텐츠 제공을 위한 발음 패턴 분석의 일 실시예를 나타내는 개념도이다. 4 is a conceptual diagram illustrating an embodiment of pronunciation pattern analysis for providing customized content.
도 4를 참조하면, 주기적으로 강세를 가지는 언어(영어 등)를 사용하는 사용자에게 익숙한 발음 패턴을 확인할 수 있다. 여기서 Δt는 강세 패턴의 메인 주기를 의미한다. Referring to FIG. 4 , a pronunciation pattern familiar to a user who periodically uses a language (eg, English) with an accent may be identified. Here, Δt means the main period of the bullish pattern.
시스템은 사용자가 모국어로 사용하는 언어의 주된 주파수 대역 및 강세 패턴을 분석하여 한국어의 그것과 차이가 나는 문장 성분에 대한 학습을 강화하는 것이 가능하다. The system analyzes the main frequency band and stress pattern of the language used by the user as a native language, and it is possible to reinforce the learning of sentence components that are different from those of Korean.
주파수대역 사운드를 이용하여 한국어 듣기/말하기 능력 향상시키는 원리는 다음과 같다. The principle of improving Korean listening/speaking skills using frequency band sound is as follows.
각 나라의 언어는 언어별로 사용하는 주파수가 상이하다.Languages of each country use different frequencies for each language.
예를 들어, 한국어에서는 500 내지 2200 Hz 대역의 소리를 사용하여 의사소통을 하는 반면, 영어(미국식)에서는 800 내지 3500 Hz 대역의 소리를 사용하여 의사소통을 한다. For example, in Korean, a sound in a range of 500 to 2200 Hz is used for communication, whereas in English (American style) a sound in a band of 800 to 3500 Hz is used for communication.
각국의 사람들은 자기가 사용하는 언어의 주파수 대역에 익숙해지며, 이 주파수 대역을 벗어나는 소리는 잘 인지하지 못하게 된다. 이로 인해 같은 소리도 사람마다 다르게 들릴 수 있다. 소리를 듣지 못하면 그 소리를 내는 것도 불가능하기 때문에 위와 같이 잘 들을 수 있는 주파수 대역의 차이는 모국어가 아닌 언어를 배우는데 장애가 될 수 있다. People in each country become accustomed to the frequency band of the language they speak, and it is difficult to recognize sounds outside this frequency band. Because of this, the same sound can be heard differently by different people. Since it is impossible to make a sound if you cannot hear it, the difference in frequency bands that you can hear well as above can be an obstacle to learning a language other than your native language.
그런데 배우려는 언어가 사용하는 주파수가 주로 사용하는 언어의 주파수와 다른 경우 주로 사용하는 언어의 주파수에서 벗어나는 주파수 부분에 대해서는 정확히 듣지 못하게 되며, 그로 인해 발음하기도 어려워진다. However, if the frequency used by the language to be learned is different from the frequency of the main language, it is difficult to hear the frequency part that is out of the frequency of the mainly used language, and thus it is difficult to pronounce.
예를 들어, 한국 사람은 평상시에 500 내지 2200 Hz의 소리를 사용하기 때문에 영어(미국식)에서 사용하는 800 내지 3500 Hz 중 겹치지 않는 2200 내지 3500 Hz 사이의 소리가 정확하게 들리지 않거나 구분이 어렵게 된다. For example, since Koreans usually use sounds of 500 to 2200 Hz, sounds between 800 and 3500 Hz used in English (American) between 2200 and 3500 Hz that do not overlap are not accurately heard or difficult to distinguish.
이러한 원리를 활용하여 시스템은 학습자의 모국어에서 주로 사용하는 주파수 대역에서 벗어난 단어 또는 문장에 대한 학습을 강화할 수 있다. Utilizing this principle, the system can enhance the learning of words or sentences outside the frequency band mainly used in the learner's native language.
또한, 언어의 종류는 일정 간격으로 반복되는 운율(문장에서 강세의 주기가 없음)이 없이 단순히 음절이 나열되는 실러블 랭귀지(syllable language)와 일정한 패턴의 운율(문장에서 일정한 주기의 강세가 반복되어 나타남)이 존재하는 스트레스 랭귀지(stress language)로 나뉠 수 있다. In addition, the type of language is a syllable language, in which syllables are simply arranged without a rhyme repeating at regular intervals (there is no stress cycle in a sentence), and a rhyme with a certain pattern (the stress of a certain period is repeated in a sentence). appear) can be divided into existing stress language.
스트레스 랭귀지에서는 상황에 따라 음절의 길이는 상이하지만 스트레스(강세)를 주는 간격이 일정하다. 즉, 스트레스를 주는 음절끼리의 시간 간격이 비슷하고 그 사이 음절들은 몇 개가 들어가던 시간 간격을 맞추기 위해 압축되어 발음된다. In a stress language, the length of a syllable is different depending on the situation, but the interval at which stress (stress) is given is constant. In other words, the time interval between syllables giving stress is similar, and the syllables in between are compressed and pronounced to match the time interval that used to be included.
분석모듈은 언어종류 분석모듈과 강세 패턴 분석모듈을 포함하여 문장이 어느 국가의 언어에 해당하는지, 상기 언어가 스트레스 랭귀지에 해당하는지, 그리고 문장에서 강세 패턴이 어떻게 되는지를 분석한다. The analysis module, including a language type analysis module and a stress pattern analysis module, analyzes which country the sentence corresponds to, whether the language corresponds to a stress language, and how the stress pattern is in the sentence.
구체적으로, 분석모듈은 외국인이 사용하는 언어의 종류, 단어, 어순 등의 데이터에 기반하여 문장이 어떤 종류의 언어로 이루어졌는지와 해당 언어에 대한 수집 데이터를 기반으로 해당 언어가 실러블 랭귀지에 속하는지 스트레스 랭귀지에 속하는지를 판단할 수 있다. Specifically, the analysis module determines what kind of language the sentence is made in based on data such as the type of language used by foreigners, words, and word order, and the language belongs to the sealable language based on the collected data about the language. You can determine whether or not you belong to a stress language.
분석모듈은 문장의 음성출력 데이터를 분석하여 화자가 문장을 끊어 읽는 위치, 출력되는 dB 값, 성조나 억양 등을 분석하여 학습자가 청취하고자 하는 문장의 강세 패턴을 판단할 수 있다. The analysis module analyzes the voice output data of the sentence and analyzes the position where the speaker cuts and reads the sentence, the output dB value, tone or intonation, etc. to determine the stress pattern of the sentence the learner wants to listen to.
강세 패턴의 주기는 사용자 또는 문장마다 다소 차이가 있을 수 있지만 시스템은 수집된 해당 외국어 문장들의 주기의 평균 또는 가장 높은 빈도로 발생하는 주기를 메인 주기로 파악하는 것이 가능하다. The period of the stress pattern may be slightly different for each user or sentence, but the system can identify the period that occurs with the average or the highest frequency of the periods of the collected foreign language sentences as the main period.
시스템은 이러한 메인 주기에 해당하는 위치에 중요 단어를 배치하여 학습자에게 제공하여 학습자가 익숙하게 한국어 문장을 학습할 수 있게 할 수도 있다. The system may arrange important words in positions corresponding to these main cycles and provide them to learners so that the learners can learn Korean sentences accustomed to them.
위에서 살펴본 본 발명의 실시예들에 따르면, 사용자의 발음평가 점수 데이터가 메인서버에 복제되고 집계되어 사용자의 성별, 거주지역, 국적, 연령 등에 따라 무기명으로 취합된 후 국가별, 연령별, 성별 등에 따라 자주 틀리는 단어, 문장, 표현 등으로 구분될 수 있으며 전체 이용자들의 발음 시도 회수 중 몇 명의 사용자가 틀리게 발음하는지를 수치화하여 빅데이터 분석을 가능하게 하며, 사용자의 국적, 성별, 나이 등을 정확하게 입력하면 비슷한 사용자 이용 환경에서 어려워하는 발음, 문장과 표현 등을 집중 훈련할 수 있는 등 종래 기술과는 차별되는 효과를 기대할 수 있다. According to the embodiments of the present invention discussed above, the pronunciation evaluation score data of the user is copied and aggregated in the main server, collected anonymously according to the user's gender, residence area, nationality, age, etc., and then according to country, age, gender, etc. It can be divided into frequently incorrect words, sentences, and expressions, and it enables big data analysis by quantifying how many users pronounce incorrectly among the total number of pronunciation attempts by users. An effect different from the prior art can be expected, such as being able to intensively train pronunciation, sentences and expressions that are difficult in the user environment.
이상에서 설명한 본 발명은 위에서 설명된 실시예들의 구성과 방법에 한정되는 것이 아니라, 상기 실시예들은 다양한 변형이 이루어질 수 있도록 각 실시예들의 전부 또는 일부가 선택적으로 조합되어 구성될 수도 있다.The present invention described above is not limited to the configuration and method of the above-described embodiments, but all or some of the embodiments may be selectively combined so that various modifications may be made.

Claims (5)

  1. 사용자가 개인정보를 시스템에 업로드하는 단계; uploading personal information to the system by the user;
    시스템이 사용자에게 한국 영상 컨텐츠를 제공한 후 발음 평가 문제를 제시한 후 피드백을 요청하는 단계; After the system provides Korean video content to the user, presenting a pronunciation evaluation problem, and then requesting feedback;
    상기 피드백에 대응하여 사용자 음성 데이터가 수집된 후 사용자의 발음이 평가되는 단계; evaluating the user's pronunciation after collecting user voice data in response to the feedback;
    사용자 단말기에 상기 사용자 음성 데이터 및 상기 평가 점수가 저장되는 단계; storing the user voice data and the evaluation score in a user terminal;
    일정 주기를 가지고 상기 저장된 평가 점수가 서버에 취합되는 단계; 및collecting the stored evaluation scores in a server with a predetermined period; and
    서버가 상기 취합된 정보를 이용하여 사용자에게 제공하는 발음 평가 문제를 보정하는 단계를 포함하는 음성 분석을 통한 한국어 발음 학습 방법.and correcting a pronunciation evaluation problem provided to a user by a server using the collected information.
  2. 제1항에 있어서, According to claim 1,
    상기 사용자 발음 평가는 STT(Speak to text) 기술에 의하여 이루어지는 것을 특징으로 하는 음성 분석을 통한 한국어 발음 학습 방법.The method for learning Korean pronunciation through voice analysis, characterized in that the evaluation of the user's pronunciation is performed by STT (Speak to text) technology.
  3. 제2항에 있어서, 3. The method of claim 2,
    상기 발음 평가는 STT 서버가 외국인의 발음 입력 시 몇 글자를 정확하게 한글로 인식할 수 있는지 여부에 의하여 결정되는 것을 특징으로 하는 음성 분석을 통한 한국어 발음 학습 방법.The pronunciation evaluation is a Korean pronunciation learning method through voice analysis, characterized in that it is determined by whether the STT server can accurately recognize how many letters in Korean when a foreigner's pronunciation is input.
  4. 제3항에 있어서, 4. The method of claim 3,
    상기 평가 점수는 사용자의 성별, 거주지역, 국적, 연령에 따라 무기명으로 취합되는 것을 특징으로 하는 음성 분석을 통한 한국어 발음 학습 방법.The evaluation score is a method for learning Korean pronunciation through voice analysis, characterized in that it is collected anonymously according to the user's gender, residential area, nationality, and age.
  5. 제4항에 있어서, 5. The method of claim 4,
    시스템은 사용자에게 레벨 테스트용 발음 평가 문제를 제시한 후 사용자 레벨을 구분하며, 레벨이 구분된 사용자에게 동일 레벨에 해당하며 성별, 거주지역, 국적, 연령 중 적어도 2이상의 항목이 동일한 사용자들에게 낮은 점수를 받은 발음 평가 문제를 제시하는 것을 특징으로 하는 음성 분석을 통한 한국어 발음 학습 방법.The system classifies the user level after presenting the pronunciation evaluation problem for the level test to the user, and the level corresponds to the same level to the divided user, and at least two items among gender, residence area, nationality, and age are lower for the same users. A method for learning Korean pronunciation through speech analysis, characterized in that it presents a pronunciation evaluation problem that has been scored.
PCT/KR2020/019110 2019-12-31 2020-12-24 Method and system for learning korean pronunciation via voice analysis WO2021137534A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020190179952A KR102396833B1 (en) 2019-12-31 2019-12-31 System and method for studying korean pronunciation using voice analysis
KR10-2019-0179952 2019-12-31

Publications (1)

Publication Number Publication Date
WO2021137534A1 true WO2021137534A1 (en) 2021-07-08

Family

ID=76686944

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2020/019110 WO2021137534A1 (en) 2019-12-31 2020-12-24 Method and system for learning korean pronunciation via voice analysis

Country Status (2)

Country Link
KR (2) KR102396833B1 (en)
WO (1) WO2021137534A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20230018095A (en) 2021-07-29 2023-02-07 케이랩 주식회사 Method of Korean Postposition Study for Non-native Learners
KR102631382B1 (en) * 2023-06-16 2024-01-31 주식회사 미카 System for providing korean education service for korean native speaker

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR19990047118A (en) * 1997-12-02 1999-07-05 정선종 Korean language education system and control method
JP2003066818A (en) * 2001-08-27 2003-03-05 Advanced Telecommunication Research Institute International Foreign language learning device
KR20050074298A (en) * 2004-01-08 2005-07-18 정보통신연구진흥원 Pronunciation test system and method of foreign language
KR20110068490A (en) * 2009-12-16 2011-06-22 포항공과대학교 산학협력단 Apparatus for foreign language learning and method for providing foreign language learning service
KR20160008949A (en) * 2014-07-15 2016-01-25 한국전자통신연구원 Apparatus and method for foreign language learning based on spoken dialogue

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100405061B1 (en) 2000-03-10 2003-11-10 문창호 Apparatus for training language and Method for analyzing language thereof
KR101048214B1 (en) * 2008-12-30 2011-07-08 주식회사 케이티 Pronunciation Correction Service Device Using Social Learning and Semantic Technology
KR101071392B1 (en) * 2009-04-16 2011-10-12 전일호 Learner centered foreign language Education system and its teaching method
KR101983772B1 (en) 2016-02-26 2019-05-29 소리노리닷컴(주) Learning system for listening
KR20190041772A (en) * 2017-10-13 2019-04-23 주식회사 하얀마인드 Apparatus and method for evaluating linguistic performance based on silence interval using comparison with other users

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR19990047118A (en) * 1997-12-02 1999-07-05 정선종 Korean language education system and control method
JP2003066818A (en) * 2001-08-27 2003-03-05 Advanced Telecommunication Research Institute International Foreign language learning device
KR20050074298A (en) * 2004-01-08 2005-07-18 정보통신연구진흥원 Pronunciation test system and method of foreign language
KR20110068490A (en) * 2009-12-16 2011-06-22 포항공과대학교 산학협력단 Apparatus for foreign language learning and method for providing foreign language learning service
KR20160008949A (en) * 2014-07-15 2016-01-25 한국전자통신연구원 Apparatus and method for foreign language learning based on spoken dialogue

Also Published As

Publication number Publication date
KR20220039679A (en) 2022-03-29
KR102396833B1 (en) 2022-05-13
KR102542602B1 (en) 2023-06-14
KR20210086182A (en) 2021-07-08

Similar Documents

Publication Publication Date Title
CN105792752B (en) Computing techniques for diagnosing and treating language-related disorders
Nowrouzi et al. Iranian EFL students' listening comprehension problems
WO2021137534A1 (en) Method and system for learning korean pronunciation via voice analysis
Bashori et al. ‘Look, I can speak correctly’: learning vocabulary and pronunciation through websites equipped with automatic speech recognition technology
WO2017061753A1 (en) Language learning system using text visualization and learner corpus
JPWO2014136534A1 (en) Understanding support system, understanding support server, understanding support method, and program
CN110853422A (en) Immersive language learning system and learning method thereof
WO2011074772A2 (en) Grammatical error simulation device and method
Yan An examination of foreign language classroom anxiety: Its sources and effects in a college English program in China
WO2009119991A2 (en) Method and system for learning language based on sound analysis on the internet
JP2002258729A (en) Foreign language learning system, information processing terminal for the same and server
Rees et al. Effects of english Cued Speech on speech perception, phonological awareness and literacy: a case study of a 9-year-old deaf boy using a cochlear implant
Wang et al. “Oh my gosh! The expression is too Chinese”: Attitudes of university teachers and students towards China English
Huang et al. An exploration of listening strategy use and proficiency in China
Yunina ARTIFICIAL INTELLIGENCE TOOLS IN FOREIGN LANGUAGE TEACHING IN HIGHER EDUCATION INSTITUTIONS
Ekiz The Effect of Human Relations and Communication Lesson on Eloquent Speaking Skill
Chua et al. Perceived difficulties in learning of mandarin among foreign-language learners and strategies to mitigate them
KR102236861B1 (en) Language acquisition assistance system using frequency bands by language
Zhang et al. Chinese EFL learners’ use of mobile dictionaries in reading comprehension tasks
KR20160086152A (en) English trainning method and system based on sound classification in internet
Podder et al. Sentence Learning System for Children with Dyslexia to Learn English As Second Language
KR20190122399A (en) Method for providing foreign language education service learning grammar using puzzle game
KR102550406B1 (en) System for providing online interactive realtime english speaking class platform service
Lan et al. A Speech Accent Detection Metrics (L2AM) for Processing Accented Speech Signal
FREITAG THE DEVELOPMENT OF SOCIOLINGUISTIC AWARENESS AND SUCCESS IN READING PERFORMANCE

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20910141

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20910141

Country of ref document: EP

Kind code of ref document: A1