WO2019132092A1 - Plush doll robot with voice recognition function - Google Patents

Plush doll robot with voice recognition function Download PDF

Info

Publication number
WO2019132092A1
WO2019132092A1 PCT/KR2018/000173 KR2018000173W WO2019132092A1 WO 2019132092 A1 WO2019132092 A1 WO 2019132092A1 KR 2018000173 W KR2018000173 W KR 2018000173W WO 2019132092 A1 WO2019132092 A1 WO 2019132092A1
Authority
WO
WIPO (PCT)
Prior art keywords
voice
file
robot
unit
input
Prior art date
Application number
PCT/KR2018/000173
Other languages
French (fr)
Korean (ko)
Inventor
이성종
Original Assignee
수상에스티주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 수상에스티주식회사 filed Critical 수상에스티주식회사
Priority to KR1020207023831A priority Critical patent/KR20200119821A/en
Publication of WO2019132092A1 publication Critical patent/WO2019132092A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H3/00Dolls
    • A63H3/28Arrangements of sound-producing means in dolls; Means in dolls for producing sounds
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H3/00Dolls
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H3/00Dolls
    • A63H3/003Dolls specially adapted for a particular function not connected with dolls
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H3/00Dolls
    • A63H3/02Dolls made of fabrics or stuffed
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H30/00Remote-control arrangements specially adapted for toys, e.g. for toy vehicles
    • A63H30/02Electrical arrangements
    • A63H30/04Electrical arrangements using wireless transmission
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H2200/00Computerized interactive toys, e.g. dolls

Definitions

  • the present invention relates to a sewing robot having a voice recognition function, and more particularly to a plush toy having a voice recognition function for outputting a voice corresponding to a user's voice by connecting to a voice recognition server through an external terminal using wireless communication Robot.
  • the present invention provides a plush dolphin robot having a voice recognition function that is connected to a server through wireless communication to analyze a voice of a user and output a voice corresponding thereto.
  • the present invention provides a plush dolphin robot having a speech recognition function capable of providing a variety of contents to a user by converting a file including a text into a voice file and correcting an error occurring during the conversion.
  • the present invention provides a stuffed toy robot having a voice recognition function that can be of interest to a child user by providing mobile contents composed of a stuffed toy character connected to an external terminal.
  • a robot for performing a sewing robot having a voice recognition function includes a controller for performing an action corresponding to a command input of a user, a controller for storing a user's voice as an input voice file, A voice recognition unit for transmitting the input voice file to the control unit and transmitting the input voice file to the voice recognition server if the voice of the user is not included, A touch sensor, and a pulse sensor for receiving the answer voice file from the speech recognition server and outputting the answer voice file as a sound; a sensor unit for detecting a body input by providing at least one sensor, A wireless communication unit for wirelessly communicating with an external terminal using at least one of Wi-Fi, Bluetooth, and NFC; A motor unit for controlling the movement of the body by driving a number of motors and an LED lighting signal among signals received from the sensor unit and the voice recognition server, And an LED (light emitting diode).
  • the voice providing unit may allow a user to synthesize a voice to select a base voice or a base voice with a predetermined voice.
  • the speech recognition server may include a TTS conversion unit that receives a scan file including text and converts the text included in the scan file into a speech file.
  • the TTS conversion unit provides basic information of the scan file to a user registered in a predetermined space on a network, and first requests an operation of correcting an error of the voice file in accordance with the basic information
  • the original of the scan file and the voice file is transmitted to the correction applicant, the correction applicant receives the corrected voice file to verify the corrected voice file, and when the verification of the corrected voice file is completed, A predetermined portion of the sales revenue of one voice file can be provided to the correction applicant.
  • the TTS conversion unit provides the correction applicant with a (100-X)% (here, X denotes the contribution of the correction applicant) of the net profit to the predetermined sales quantity of the voice file ((100-X) - (sales quantity - predetermined sales quantity))% of the net profit when the sales quantity exceeds the preset sales quantity.
  • a sewing robot having a voice recognition function connected to a server through wireless communication to analyze a voice of a user and output a voice corresponding thereto.
  • an operation of the robot due to the motor and an LED (LED) blink to provide a sewing robot having a voice recognition function that can be of interest to the user .
  • a plush dolls robot having a speech recognition function capable of providing a variety of contents to a user by converting a file including a text into a voice file and correcting an error occurring during the conversion do.
  • a stuffed toy robot having a voice recognition function that can be of interest to a child user by providing mobile contents composed of a stuffed toy character connected to an external terminal.
  • FIG. 1 is a block diagram of a stuffed toy robot having a speech recognition function according to an embodiment of the present invention.
  • FIG. 2 is a diagram for explaining a process of converting a scan file including text into an audio file according to an embodiment of the present invention.
  • FIG. 3 is a diagram showing a voice file sold according to an embodiment of the present invention.
  • FIG. 4 is a diagram illustrating a change in expression of an application character according to a user voice according to an embodiment of the present invention.
  • FIG. 1 is a block diagram of a sewing robot 100 having a speech recognition function according to an embodiment of the present invention.
  • a sewing robot 100 having a voice recognition function includes a control unit 110, a voice recognition unit 120, a voice data providing unit 130, a wireless communication unit 140, a sensor unit 150, An LED 170, a voice recognition server 200, and an external terminal 300.
  • the voice recognition server 200 and the voice recognition server 200 are connected to each other via a network.
  • the controller 110 may perform an action corresponding to a command input by the user.
  • the voice recognition unit 120 stores a voice of a user as an input voice file and transmits the input voice file to the control unit 110 when a predetermined command among the input voice is included, If the preset command is not included, the input voice file can be transferred to the voice recognition server 200 through the external terminal 300.
  • the voice recognition unit 120 recognizes a predetermined 'LED lighting' command among the input voice file, And the controller 110 turns on the LED 170 provided in the stuffed animal robot 100 or the LED provided in the outside. In addition, the control unit 110 analyzes the input voice file to start the LED lighting. And transmits the answer voice file to the stuffed animal robot 100.
  • the voice data providing server 130 receives the answer voice file corresponding to the input voice file from the voice recognition server 200, and outputs the answer voice file transmitted through the voice.
  • the voice data providing unit 130 may select a base voice by a user to synthesize a voice, or may select a base voice with a predetermined voice. For example, the user may input his /
  • the wireless communication unit 140 may communicate with an external terminal 300 using at least one of Wi-Fi, Bluetooth, and NFC,
  • Wireless communication can be performed.
  • the NFC communication is NFC-communicated with the NFC reader attached to the sewing robot 100 and the NFC reader installed in the external terminal 300, the NFC communication is performed through the application executed in the external terminal 300, (100) can be controlled.
  • the stuffed animal robot 100 and the external terminal 300 are in a state in which the NFC communication is connected through the NFC reader.
  • the application of the external terminal 300 is executed and the 'fairy tale' menu is clicked.
  • the moving picture is clicked, an answer voice file in which the text is changed to speech in advance is selected, and the answer voice file is transmitted to the voice data providing unit 130 of the stuffed animal robot 100, 130) outputs the answer voice file as a sound.
  • the sensor unit 150 may detect at least one of a touch sensor and a pulse sensor to detect a body input.
  • the touch sensor is attached to the head portion of the stuffed animal robot 100, and when the user's touch is recognized, the touch sensor transmits a signal to the controller 110.
  • the sound supplied from the sound providing unit may be output to the signal transmitted to the controller 110, or the arm may be operated by driving the motor.
  • the pulse sensor is attached to an arm portion of the sewing robot 100 and is executed when a voice command is given to the user's voice by the voice recognition unit 120.
  • the voice recognition unit 120 When the user holds the arm of the stuffed animal robot 100, the pulse is measured, and the measurement signal is transmitted to the controller 110.
  • the controller 110 controls the controller 110 to generate a voice corresponding to the measurement signal, A sound is output from the voice remover 130 and the user can measure the pulse.
  • the motor of the motor unit 160 is attached to the shoulder of the sewing robot 100.
  • the controller 110 receives a signal, the motor unit 160 drives the motor based on the received signal.
  • the LED 170 is transmitted to the controller 110 when an LED lighting signal is generated in the sensor unit 150 and the voice recognition server 200.
  • the LED lighting signal is transmitted to the LED lighting signal received from the controller 110 And can be turned on in a corresponding manner.
  • the input voice file stored in the voice recognition server 200 can be analyzed and stored through the voice recognition program 210 included in the voice recognition server 200.
  • an answer voice file is generated based on the analyzed information through the speech recognition program 210, and the generated answer voice file is transmitted to the voice data providing unit 130 of the stuffed animal robot 100,
  • the answerer 130 can output a voice response file by voice.
  • the wireless communication unit 140 and the external terminal 300 of the stuffed animal robot 100 are connected to Wi-Fi, Bluetooth, and NFC communication,
  • the plush doll robot 100 tells the voice 'what is the weather today', the plush doll
  • the voice recognition unit 120 of the robot 100 stores the voice as an input voice file
  • the input speech file transmitted from the speech recognition server 200 is analyzed through the speech recognition program 210, and the 'Today's weather is minus 4 degrees.
  • a voice file is generated and transmitted to the voice data providing unit 130 of the stuffed animal robot 100.
  • the voice data providing unit 130 sounds an answer voice file to output voice data, .
  • FIG. 2 is a diagram for explaining a process of converting a scan file including text into an audio file according to an embodiment of the present invention.
  • the voice recognition server 200 may include a TTS converter 220 for receiving a scan file including text and converting the text included in the scan file to a voice file.
  • the TTS conversion unit 220 provides the basic information of the scan file to a user registered in a predetermined space on the network and transmits the first correction applicant who applied for the correction of the error of the voice file corresponding to the basic information And a verification unit that verifies the corrected voice file by receiving the voice file whose correction is completed by the correction applicant, and when the verification of the corrected voice file is completed, Of the sales revenue of the applicant.
  • FIG. 3 is a diagram showing a voice file sold according to an embodiment of the present invention.
  • the scan file including the text is converted into a voice file, and the voice file is sold to explain the occurrence of profit.
  • the distribution of the profit will be described in detail.
  • the TTS conversion unit provides the correction applicant with a (100-X)% (where X represents the contribution of the correction applicant) of the net profit to the predetermined sales quantity of the audio file, (100-X) - (Sales Quantity - Pre-established Sales Quantity))% of the net profit.
  • the applicant can increase the participation in the application for correction, and the manager can increase the profit more as the sales quantity increases, thereby effectively distributing the profit of the applicant and the manager.
  • a user registered in a predetermined space on the network is notified of a scan file of a book called 'great pussy', basic information (e.g., title of a book, number of pages of a book, etc.)
  • the error% e.g., 30%, in this case, the error included in the voice file means the contribution when the correction is completed
  • the user registered in the predetermined space provides the voice file 30% of the speech file and send the original of the speech file containing the 'great shit' scan file and 30% error to the first applicant who applied for the correction of the 30% error of the voice file do.
  • the original is transmitted to the correction applicant, and when the correction applicant completes correcting the error of the voice file including the 30% error, the voice file having corrected the error is transmitted to the voice recognition server 200.
  • the voice recognition server (200) notifies the voice recognition server (200) of the voice file that the voice recognition server (200) corrects the error and informs another user registered in a predetermined space on the network
  • the voice recognition server (200) When the user purchases and reproduces an audio file, the sound is output from the sewing robot 100 and the external terminal 300.
  • the correction applicant who corrected the 30% error provides 70 (100-30%) of the net profit up to 10 of the sales quantity of the error-corrected voice file, and ((100 -30) - (sales quantity-10))%.
  • the error compensated user provides 55 ((100-30) - (25-10))% of the net profit.
  • the speech recognition server 200 may extract the word when a preset word is included in the input speech file received from the speech recognition unit 120, generate information based on the extracted word, And generate an answer voice file by analyzing based on the generated information.
  • the stuffed animal robot 100 For example, if the user speaks a voice 'promise tomorrow at 12 o'clock', the stuffed animal robot 100 generates an input voice file based on the voice and transmits it to the voice recognition server 200, 'Tomorrow', '12 o'clock', and 'appointment' of the input voice file are extracted and stored in the server 200.
  • the speech recognition server 200 When the user speaks a voice 'Tell me a schedule tomorrow', the speech recognition server 200 generates an answer voice file, reflects the extracted and stored information, and the 'tomorrow schedule has an appointment at 12 o'clock.' And sends the answer voice file to the voice data providing unit 130 of the stuffed animal robot 100 and outputs the answer voice to the user so that the user can recognize the schedule for tomorrow.
  • the external terminal 300 receives a signal of the stuffed animal robot 100 from the wireless communication unit 140 of the stuffed animal robot 100,
  • the mobile communication terminal 200 provides the mobile contents using the received signal
  • the doll robot 100 can relay the communication.
  • FIG. 4 is a diagram illustrating a change in expression of an application character according to a user voice according to an embodiment of the present invention.
  • the mobile contents include a function of expressing the representation of the mobile robot 100 in the form of animation through an application, a function of moving the mouth of the robot character to the sewing robot, outputting a voice file,
  • the stuffed toy robot 100 can perform at least one of the functions of outputting sound, driving the motor and flashing LEDs, and providing a learning content by executing the application .
  • an input voice file based on the voice is analyzed by the voice recognition server 200 to select an emotion, and the selected emotion is transmitted to the application of the external terminal 300 And displays a facial expression matching the emotion on the external terminal 300.
  • the present invention can provide a plush dolphin robot having a speech recognition function, which is connected to a server through wireless communication to analyze a user's voice and output a corresponding voice.
  • a plush dolphin robot having a speech recognition function that allows a user to provide various contents by converting a file including a text into a voice file and correcting an error occurring during the conversion.
  • the control method of the sewing robot having the speech recognition function may be recorded in a computer-readable medium including program instructions for performing various computer-implemented operations.
  • the computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination.
  • the media may be program instructions that are specially designed and constructed for the present invention or may be available to those skilled in the art of computer software.
  • Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Magneto-optical media, and hardware devices specifically configured to store and execute program instructions such as ROM, RAM, flash memory, and the like.
  • Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Toys (AREA)

Abstract

Disclosed is a plush doll robot with a voice recognition function. A plush doll robot with a voice recognition function, according to one embodiment of the present invention, comprises: a control unit for performing an action corresponding to a command input by a user; a voice recognition unit for storing a user's voice as an input voice file, transmitting the input voice file to the control unit when a predetermined command is included in the input voice, and transmitting the input voice file to a voice recognition server when a predetermined command is not included in the input voice; a voice providing unit for receiving a voice response file corresponding to the input voice file from the voice recognition server and outputting the voice response file as a voice; a sensor unit, provided with at least one sensor of a touch sensor and a pulse sensor, for sensing a body input; a wireless communication unit for wirelessly communicating with an external terminal by using at least one of Wi-Fi, Bluetooth, and NFC; a motor unit for driving a plurality of motors to control movement of the body of the plush doll robot; and an LED for, when an LED lighting signal is included in signals received from the sensor unit and the voice recognition server, illuminating light in response to the received lighting signal.

Description

음성인식 기능을 가진 봉제 인형 로봇Plush doll robot with speech recognition function
본 발명은 음성인식 기능을 가진 봉제 인형 로봇에 관한 것으로 특히, 무선통신을 이용하여 외부 단말을 통해 음성인식서버와 연결하여 사용자의 말에 대응하는 답변을 음성으로 출력하는 음성인식 기능을 가진 봉제 인형 로봇에 관한 것이다.BACKGROUND OF THE INVENTION 1. Field of the Invention [0001] The present invention relates to a sewing robot having a voice recognition function, and more particularly to a plush toy having a voice recognition function for outputting a voice corresponding to a user's voice by connecting to a voice recognition server through an external terminal using wireless communication Robot.
종래에는 인형에 인터넷을 결합하여 인행을 제어하는 시스템이 개발 되고 있다. 인형은 유아나 어린이가 가지고 놀면서 신체 운동 발달 및 기능을 숙달하고, 상상력이나 창의력 개발을 통해 지능 발달 등 교육적으로 중요한 역할을 하기 때문에 스마트한 인형 기술 개발이 크게 관심을 받고 있다. 다만, 인형은 제한된 소리를 출력하거나, 동작이 없으므로 사용자로 하여금 지속적으로 새로운 관심과 흥미를 끌어내기 어려운 한계가 있었다.따라서, 인형은 사용자의 소리를 인식하여 대답을 하거나, 인형의 모터를 연결하여 기설정된 상황에 동작을 할 수 있게 하고, 외부 단말과 연결되어 다양한 콘텐츠를 제공함으로써 사용자로 하여금 흥미를 잃지 않게 하는 음성인식 기능을 가진 봉제 인형 로봇에 관한 연구가 필요하다.Conventionally, a system for controlling an envoy by combining the Internet with a doll has been developed. Dolls are very interested in the development of smart doll technology because they play an important role in infant and child playing and develop the physical exercise and function, and play an educational role by developing imagination and creativity through intelligence development. However, since the doll has a limited sound output or no action, it is difficult for the user to continuously attract new interest and interest. Therefore, the doll recognizes the user's voice and answers, or connects the doll motor It is necessary to study a sewing doll robot having a voice recognition function which enables a user to operate in a preset situation and to provide a variety of contents by being connected to an external terminal so that the user does not lose interest.
본 발명은 무선통신을 통해 서버에 접속되어 사용자의 음성을 분석하여 그에 상응하는 음성을 출력하는 음성인식 기능을 가진 봉제 인형 로봇을 제공한다.The present invention provides a plush dolphin robot having a voice recognition function that is connected to a server through wireless communication to analyze a voice of a user and output a voice corresponding thereto.
본 발명은 사용자의 신체 입력이 감지되면 모터로 인한 로봇의 동작 및 엘이디(LED)가 점멸함으로써, 사용자로 하여금 흥미를 가질 수 있는 음성인식When the user's body input is sensed, the operation of the robot and the LED (LED) blink due to the motor, so that the user can recognize the voice recognition
기능을 가진 봉제 인형 로봇을 제공한다.To provide a plush doll robot with a function.
본 발명은 텍스트가 포함된 파일을 음성파일로 변환하고, 상기 변환 중 생기는 오류를 수정함으로써, 사용자로 하여금 다양한 콘텐츠를 제공할 수 있는 음성인식 기능을 가진 봉제 인형 로봇을 제공한다.The present invention provides a plush dolphin robot having a speech recognition function capable of providing a variety of contents to a user by converting a file including a text into a voice file and correcting an error occurring during the conversion.
본 발명은 외부 단말과 연결되어, 봉제 인형 로봇 캐릭터로 이루어진 모바일 콘텐츠를 제공함으로써, 어린이 사용자로 하여금 흥미를 가질 수 있는 음성인식 기능을 가진 봉제 인형 로봇을 제공한다.The present invention provides a stuffed toy robot having a voice recognition function that can be of interest to a child user by providing mobile contents composed of a stuffed toy character connected to an external terminal.
본 발명의 일실시예에 따른 음성인식 기능을 가진 봉제 인형 로봇은 사용자의 명령 입력에 대응하는 액션을 수행하는 제어부, 사용자의 음성을 입력음성파일로 저장하고, 상기 입력된 음성 중 기설정된 명령이 포함되어 있는 경우, 상기 입력음성파일을 상기 제어부로 전송하고, 상기 사용자의 음성 중 기설정된 명령이 포함되어 있지 않으면 상기 입력음성파일을 음성인식서버로 전달하는 음성인식부, 상기 입력음성파일에 대응하는 답변음성파일을 음성인식서버로부터 전달받고, 상기 전달받은 답변음성파일을 소리로 출력하는 음성제공부, 터치센서 및 맥박센서 중 적어도 하나의 센서를 마련하여 신체 입력을 감지하는 센서부, 와이파이(Wi-Fi), 블루투스(Bluetooth) 및 NFC 중 적어도 하나의 통신을 사용하여 외부 단말과 무선통신을 하는 무선통신부, 다수의 모터를 구동하여 신체의 움직임을 제어하는 모터부 및 상기 센서부 및 음성인식서버에서 전달받은 신호 중 엘이디(LED) 점등신호가 포함되어 있는 경우, 상기 전달받은 점등신호에 대응하는 방식으로 점등하는 엘이디(LED)를 포함한다.A robot for performing a sewing robot having a voice recognition function according to an embodiment of the present invention includes a controller for performing an action corresponding to a command input of a user, a controller for storing a user's voice as an input voice file, A voice recognition unit for transmitting the input voice file to the control unit and transmitting the input voice file to the voice recognition server if the voice of the user is not included, A touch sensor, and a pulse sensor for receiving the answer voice file from the speech recognition server and outputting the answer voice file as a sound; a sensor unit for detecting a body input by providing at least one sensor, A wireless communication unit for wirelessly communicating with an external terminal using at least one of Wi-Fi, Bluetooth, and NFC; A motor unit for controlling the movement of the body by driving a number of motors and an LED lighting signal among signals received from the sensor unit and the voice recognition server, And an LED (light emitting diode).
본 발명의 일측면에 따르면, 상기 음성제공부는 사용자가 목소리를 합성하여 베이스 목소리를 선택하거나 기설정된 목소리로 베이스 목소리를 선택할 수 있다.According to an aspect of the present invention, the voice providing unit may allow a user to synthesize a voice to select a base voice or a base voice with a predetermined voice.
본 발명의 일측면에 따르면, 상기 음성인식서버는 텍스트가 포함된 스캔파일을 전송 받고, 상기 스캔파일에 포함된 텍스트를 음성파일로 변환하는 TTS 변환부를 포함할 수 있다.According to an aspect of the present invention, the speech recognition server may include a TTS conversion unit that receives a scan file including text and converts the text included in the scan file into a speech file.
본 발명의 일측면에 따르면, 상기 TTS 변환부는 네트워크 상의 소정공간에 등록된 사용자에게 상기 스캔파일의 기초정보를 제공하고, 상기 기초정보에 대응하여 상기 음성파일의 오류를 보정하는 작업을 최초로 신청한 보정신청자에게 상기 스캔파일 및 상기 음성파일의 원본을 전송하며, 상기 보정신청자가 보정을 완료한 음성파일을 수신하여 상기 보정한 음성파일을 검증하고, 상기 보정한 음성파일의 검증이 완료되면 상기 보정한 음성파일의 판매 수익의 기설정된 부분을 상기 보정신청자에게 제공할 수 있다.According to an aspect of the present invention, the TTS conversion unit provides basic information of the scan file to a user registered in a predetermined space on a network, and first requests an operation of correcting an error of the voice file in accordance with the basic information The original of the scan file and the voice file is transmitted to the correction applicant, the correction applicant receives the corrected voice file to verify the corrected voice file, and when the verification of the corrected voice file is completed, A predetermined portion of the sales revenue of one voice file can be provided to the correction applicant.
본 발명의 일측면에 따르면, 상기 TTS 변환부는 상기 음성파일의 기설정된 판매 수량에 대해서는 순수익의 (100-X)%(여기서, X는 상기 보정신청자의 기여도를 의미함)를 상기 보정신청자에게 제공하고, 기설정된 판매 수량을 초과하는 경우, 순수익의 ((100-X)-(판매 수량-기설정된 판매 수량))%를 제공할 수 있다.According to one aspect of the present invention, the TTS conversion unit provides the correction applicant with a (100-X)% (here, X denotes the contribution of the correction applicant) of the net profit to the predetermined sales quantity of the voice file ((100-X) - (sales quantity - predetermined sales quantity))% of the net profit when the sales quantity exceeds the preset sales quantity.
본 발명의 일실시예에 따르면, 무선통신을 통해 서버에 접속되어 사용자의 음성을 분석하여 그에 상응하는 음성을 출력하는 음성인식 기능을 가진 봉제 인형 로봇이 제공된다.According to an embodiment of the present invention, there is provided a sewing robot having a voice recognition function connected to a server through wireless communication to analyze a voice of a user and output a voice corresponding thereto.
본 발명의 일실시예에 따르면, 사용자의 신체 입력이 감지되면 모터로 인한 로봇의 동작 및 엘이디(LED)가 점멸함으로써, 사용자로 하여금 흥미를 가질 수 있는 음성인식 기능을 가진 봉제 인형 로봇이 제공된다.According to an embodiment of the present invention, when a user's body input is sensed, an operation of the robot due to the motor and an LED (LED) blink to provide a sewing robot having a voice recognition function that can be of interest to the user .
본 발명의 일실시예에 따르면, 텍스트가 포함된 파일을 음성파일로 변환하고, 상기 변환 중 생기는 오류를 수정함으로써, 사용자로 하여금 다양한 콘텐츠를 제공할 수 있는 음성인식 기능을 가진 봉제 인형 로봇이 제공된다.According to one embodiment of the present invention, a plush dolls robot having a speech recognition function capable of providing a variety of contents to a user by converting a file including a text into a voice file and correcting an error occurring during the conversion do.
본 발명의 일실시예에 따르면, 외부 단말과 연결되어, 봉제 인형 로봇 캐릭터로 이루어진 모바일 콘텐츠를 제공함으로써, 어린이 사용자로 하여금 흥미를 가질 수 있는 음성인식 기능을 가진 봉제 인형 로봇이 제공된다.According to an embodiment of the present invention, there is provided a stuffed toy robot having a voice recognition function that can be of interest to a child user by providing mobile contents composed of a stuffed toy character connected to an external terminal.
도 1은 본 발명의 실시예에 따른 음성인식 기능을 가진 봉제 인형로봇의 블록도이다.1 is a block diagram of a stuffed toy robot having a speech recognition function according to an embodiment of the present invention.
도 2는 본 발명의 실시예에 따라 텍스트가 포함된 스캔파일을 음성파일로 변환하는 과정을 설명하기 위한 도면이다.2 is a diagram for explaining a process of converting a scan file including text into an audio file according to an embodiment of the present invention.
도 3은 본 발명의 실시예에 따라 판매되는 음성파일을 도시한 도면이다.3 is a diagram showing a voice file sold according to an embodiment of the present invention.
도 4는 본 발명의 실시예에 따라 사용자 음성 따른 어플리케이션 캐릭터의 표정 변화를 도시한 도면이다.4 is a diagram illustrating a change in expression of an application character according to a user voice according to an embodiment of the present invention.
이하, 첨부된 도면들에 기재된 내용들을 참조하여 본 발명의 실시 예들을 상세하게 설명한다. 다만, 본 발명이 실시예들에 의해 제한되거나 한정되는 것은 아니다. 각 도면에 제시된 동일한 참조부호는 동일한 부재를 나타낸다.Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings. However, the present invention is not limited to or limited by the embodiments. Like reference symbols in the drawings denote like elements.
도 1은 본 발명의 실시예에 따른 음성인식 기능을 가진 봉제 인형로봇(100)의 블록도이다.1 is a block diagram of a sewing robot 100 having a speech recognition function according to an embodiment of the present invention.
도 1을 참고하면, 음성인식 기능을 가진 봉제 인형 로봇(100)은 제어부(110), 음성인식부(120), 음성제공부(130), 무선통신부(140), 센서부(150), 모터부(160), 엘이디(LED)(170), 음성인식서버(200) 및 외부 단말(300)을 포함할 수 있다.1, a sewing robot 100 having a voice recognition function includes a control unit 110, a voice recognition unit 120, a voice data providing unit 130, a wireless communication unit 140, a sensor unit 150, An LED 170, a voice recognition server 200, and an external terminal 300. The voice recognition server 200 and the voice recognition server 200 are connected to each other via a network.
상기 제어부(110)는 사용자의 명령 입력에 대응하는 액션을 수행할 수 있다.The controller 110 may perform an action corresponding to a command input by the user.
상기 음성인식부(120)는 사용자의 음성을 입력음성파일로 저장하고, 상기 입력된 음성 중 기설정된 명령이 포함되어 있는 경우, 상기 입력음성파일을 제어부(110)로 전송하고, 상기 사용자의 음성 중 기설정된 명령이 포함되어 있지않으면 상기 입력음성파일을 외부 단말(300)을 통해 음성인식서버(200)로 전달할 수 있다.The voice recognition unit 120 stores a voice of a user as an input voice file and transmits the input voice file to the control unit 110 when a predetermined command among the input voice is included, If the preset command is not included, the input voice file can be transferred to the voice recognition server 200 through the external terminal 300.
일예로, 상기 사용자는 'LED 점등해줘' 라는 음성을 말하면, 입력음성파일로 저장되고, 음성인식부(120)에서는 상기 입력음성파일 중 기설정된 'LED 점등' 이라는 명령을 인식하고 제어부(110)에 전송되고, 상기 제어부(110)에서는 상기 전송된 명령을 봉제 인형 로봇(100)에 마련되어 있는 LED(170) 또는 외부에 마련되어 있는 LED를 점등한다. 또한 상기 입력음성파일은 제어부(110)에서는 상기 입력음성파일을 분석하여 'LED 점등 시작하겠습니다.' 라는 답변음성파일을 봉제 인형 로봇(100)에게 전달한다. 다시 도 1을 참고하면, 상기 음성제공부(130)는 상기 입력음성파일에 대응하는 답변음성파일을 음성인식서버(200)에서 전달받고, 상기 전달받은 답변 음성파일을 소리로 출력할 수 있다.For example, when the user speaks 'LED lit', the user is stored as an input voice file. The voice recognition unit 120 recognizes a predetermined 'LED lighting' command among the input voice file, And the controller 110 turns on the LED 170 provided in the stuffed animal robot 100 or the LED provided in the outside. In addition, the control unit 110 analyzes the input voice file to start the LED lighting. And transmits the answer voice file to the stuffed animal robot 100. Referring again to FIG. 1, the voice data providing server 130 receives the answer voice file corresponding to the input voice file from the voice recognition server 200, and outputs the answer voice file transmitted through the voice.
일예로, 사용자가 '곰 세마리 틀어줘' 라는 음성을 말하면, 상기 음성을 음성인식부(120)에서 입력음성파일로 저장되며, 상기 단어에는 기설정된 명령이 없는 상태이다. 따라서, 상기 입력음성파일은 외부 단말(300)을 통해 음성인식서버(200)로 전송되고, 상기 음성인식서버(200)에서는 입력음성파일을 분석하여, '곰 세마리 재생하겠습니다.'라는 말과 '곰 세마리' 노래를 답변음성파일로 생성한다. 상기 생성된 답변음성파일은 음성제공부(130)에 전달되며, 상기 음성제공부(130)에서는 상기 답변음성파일을 소리로 출력한다. 또한 상기 음성제공부(130)는 사용자가 목소리를 합성하여 베이스목소리를 선택하거나 기설정된 목소리로 베이스 목소리를 선택할 수 있다. 일예로, 사용자는 자신의 목소리를 입력하여 특정 단어에 있어서 사For example, if the user speaks a voice of 'bear three bears', the voice is stored in the voice recognition unit 120 as an input voice file, and the word has no preset command. Therefore, the input voice file is transmitted to the voice recognition server 200 through the external terminal 300, and the voice recognition server 200 analyzes the input voice file and reads 'three bears will be played' Create three "Bear" songs as answer voice files. The generated answer voice file is transmitted to the voice data providing unit 130, and the voice data providing unit 130 outputs the answer voice file by voice. In addition, the voice data providing unit 130 may select a base voice by a user to synthesize a voice, or may select a base voice with a predetermined voice. For example, the user may input his /
용자 자신의 목소리로 출력될 수 있고, 여러 개의 목소리를 합성하여 사용자가 희It can be output to the user's own voice, composing multiple voices,
망하는 목소리를 생성하여 상기 목소리를 봉제 인형 로봇(100)에 베이스 목소리로And the voice is transmitted to the sewing doll robot 100 as a base voice
등록하여 상기 등록된 베이스 목소리로 답변음성파일의 소리를 출력할 수 있다.And the voice of the answer voice file can be outputted by the registered base voice.
다시 도 1을 참고하면, 상기 무선통신부(140)는 와이파이(Wi-Fi),블루투스(Bluetooth) 및 NFC 중 적어도 하나의 통신을 사용하여 외부 단말(300)과1, the wireless communication unit 140 may communicate with an external terminal 300 using at least one of Wi-Fi, Bluetooth, and NFC,
무선통신을 할 수 있다.Wireless communication can be performed.
상기 NFC 통신은 봉제 인형 로봇(100)에 부착되어 있는 NFC 리더와 외부 단말(300)에 설치되어 있는 NFC 리더와 NFC 통신을 하게 되면, 외부 단말(300)에 실행되어 있는 어플리케이션을 통해 봉제 인형 로봇(100)의 제어할 수 있다.When the NFC communication is NFC-communicated with the NFC reader attached to the sewing robot 100 and the NFC reader installed in the external terminal 300, the NFC communication is performed through the application executed in the external terminal 300, (100) can be controlled.
일예로, 봉제 인형 로봇(100)과 외부 단말(300)은 NFC 리더를 통해 NFC 통신이 연결되어 있는 상태이며, 상기 외부 단말(300)의 어플리케이션을 실행하여 '동화' 메뉴를 클릭하여 사용자가 원하는 동화를 클릭하면 미리 생성된 상기동화에 텍스트를 음성으로 바꾼 답변음성파일이 선택되고, 상기 봉제 인형 로봇(100)의 음성제공부(130)에게 상기 답변음성파일이 전달되어, 상기 음성제공부(130)에서 답변음성파일을 소리로 출력한다.For example, the stuffed animal robot 100 and the external terminal 300 are in a state in which the NFC communication is connected through the NFC reader. When the application of the external terminal 300 is executed and the 'fairy tale' menu is clicked, When the moving picture is clicked, an answer voice file in which the text is changed to speech in advance is selected, and the answer voice file is transmitted to the voice data providing unit 130 of the stuffed animal robot 100, 130) outputs the answer voice file as a sound.
다시 도 1을 참고하면, 상기 센서부(150)는 터치센서 및 맥박센서 중 적어도 하나의 센서를 마련하여 신체 입력을 감지할 수 있다.Referring again to FIG. 1, the sensor unit 150 may detect at least one of a touch sensor and a pulse sensor to detect a body input.
상기 터치센서는 봉제 인형 로봇(100) 머리부분의 부착되어, 사용자의 터치가 인식이 되면, 제어부(110)에 신호를 전달한다.The touch sensor is attached to the head portion of the stuffed animal robot 100, and when the user's touch is recognized, the touch sensor transmits a signal to the controller 110.
상기 제어부(110)에 전달된 신호에 음성제공 부에서 기설정된 소리를 출력하거나 모터의 구동으로 인한 팔이 동작할 수 있다.The sound supplied from the sound providing unit may be output to the signal transmitted to the controller 110, or the arm may be operated by driving the motor.
상기 맥박센서는 봉제 인형 로봇(100) 팔 부분에 부착되어, 음성인식부(120)에서 사용자 음성에 기설정된 명령이 있는 경우 실행된다. 상기 맥박센서가 실행되면, 사용자는 상기 봉제 인형 로봇(100)의 팔을 잡으면 맥박이 측정이 되어, 상기 측정 신호가 제어부(110)로 전달되고, 상기 제어부(110)에서 측정 신호에 상응하는 음성파일을 생성하여 음성제공부(130)에서 소리로 출력되어 사용자로 하여금 맥박을 측정할 수 있다.The pulse sensor is attached to an arm portion of the sewing robot 100 and is executed when a voice command is given to the user's voice by the voice recognition unit 120. [ When the user holds the arm of the stuffed animal robot 100, the pulse is measured, and the measurement signal is transmitted to the controller 110. The controller 110 controls the controller 110 to generate a voice corresponding to the measurement signal, A sound is output from the voice remover 130 and the user can measure the pulse.
다시 도 1을 참고하면, 상기 모터부(160)의 모터는 봉제 인형 로봇(100)의 어깨 안에 부착되어, 제어부(110)에서 신호를 입력 받으면, 상기 입력 받은 신호를 기반으로 모터를 구동할 수 있다.상기 엘이디(LED)(170)는 센서부(150) 및 음성인식서버(200)에서 LED 점등 신호가 생성되면 제어부(110)에 전달되고, 상기 제어부(110)에서 전달받은 LED 점등 신호에 대응하는 방식으로 점등될 수 있다.1, the motor of the motor unit 160 is attached to the shoulder of the sewing robot 100. When the controller 110 receives a signal, the motor unit 160 drives the motor based on the received signal. The LED 170 is transmitted to the controller 110 when an LED lighting signal is generated in the sensor unit 150 and the voice recognition server 200. The LED lighting signal is transmitted to the LED lighting signal received from the controller 110 And can be turned on in a corresponding manner.
상기 음성인식서버(200)는 봉제 인형 로봇(100)의 무선통신부와 (140) 외부 단말(300)이 와이파이(Wi-Fi), 블루투스(Bluetooth) 및 NFC 통신에 접속되어 있는 경우, 상기 음성인식부(120)에 저장된 입력음성파일을 전송 받고, 상기 전송 받은 입력음성파일을 상기 음성인식서버(200)에 포함되어 있는 음성인식프로그램(210)을 통해 분석 및 저장할 수 있다.When the wireless communication unit of the stuffed animal robot 100 and the external terminal 300 of the robot 140 are connected to Wi-Fi, Bluetooth and NFC communication, The input voice file stored in the voice recognition server 200 can be analyzed and stored through the voice recognition program 210 included in the voice recognition server 200. [
또한 상기 음성인식 프로그램(210)을 통해 분석된 정보를 기반으로하여 답변음성파일을 생성하고, 상기 생성된 답변음성파일을 상기 봉제 인형 로봇(100)의 음성제공부(130)로 전달하여 상기 음성제공부(130)에서 답변음성파일을 소리로 출력할 수 있다.Also, an answer voice file is generated based on the analyzed information through the speech recognition program 210, and the generated answer voice file is transmitted to the voice data providing unit 130 of the stuffed animal robot 100, The answerer 130 can output a voice response file by voice.
일예로, 봉제 인형 로봇(100)의 무선통신부(140)와 외부 단말(300)이 와이파이(Wi-Fi), 블루투스(Bluetooth) 및 NFC 통신에 접속되어 있고, 사용자가For example, when the wireless communication unit 140 and the external terminal 300 of the stuffed animal robot 100 are connected to Wi-Fi, Bluetooth, and NFC communication,
봉제 인형 로봇(100)에게 '오늘 날씨 어때' 라는 음성을 말하면, 상기 봉제 인형If the plush doll robot 100 tells the voice 'what is the weather today', the plush doll
로봇(100)의 음성인식부(120)에서 상기 음성을 입력음성파일로 저장하고, 외부 단The voice recognition unit 120 of the robot 100 stores the voice as an input voice file,
말(300)을 통해 음성인식서버(200)로 전달된다.And transmitted to the speech recognition server 200 through the speech 300.
상기 음성인식서버(200)에서 전달된 입력음성파일을 음성인식 프로그램(210)을 통해 분석하여 '오늘 날씨는 영하 4도 입니다. 외출 하실 때 옷을 따뜻하게 입으세요' 라는 답변음성파일을 생성하여 상기 봉제 인형 로봇(100)의 음성제공부(130)로 전달하고, 상기 음성제공부(130)에서 답변음성파일을 소리 출력하여 사용자에게 음성으로 전달한다.The input speech file transmitted from the speech recognition server 200 is analyzed through the speech recognition program 210, and the 'Today's weather is minus 4 degrees. A voice file is generated and transmitted to the voice data providing unit 130 of the stuffed animal robot 100. The voice data providing unit 130 sounds an answer voice file to output voice data, .
도 2는 본 발명의 실시예에 따라 텍스트가 포함된 스캔파일을 음성파일로 변환하는 과정을 설명하기 위한 도면이다.2 is a diagram for explaining a process of converting a scan file including text into an audio file according to an embodiment of the present invention.
도 2를 참고하면, 상기 음성인식서버(200)는 텍스트가 포함된 스캔파일을 전송 받고, 상기 스캔파일에 포함된 텍스트를 음성파일로 변환하는 TTS 변환부(220)를 포함할 수 있다.Referring to FIG. 2, the voice recognition server 200 may include a TTS converter 220 for receiving a scan file including text and converting the text included in the scan file to a voice file.
상기 TTS 변환부(220)는 네트워크 상의 소정 공간에 등록된 사용자에게 상기 스캔파일의 기초정보를 제공하고, 상기 기초정보에 대응하여 상기 음성파일의 오류를 보정하는 작업을 신청한 최초로 신청한 보정신청자에게 스캔파일 및 상기 음성파일의 원본을 전송하며, 상기 보정신청자가 보정을 완료한 음성파일을 수신하여 상기 보정한 음성파일을 검증하고, 상기 보정한 음성파일의 검증이 완료되면 상기 보정한 음성파일의 판매 수익의 기설정된 부분을 보정신청자에게 제공할 수 있다.The TTS conversion unit 220 provides the basic information of the scan file to a user registered in a predetermined space on the network and transmits the first correction applicant who applied for the correction of the error of the voice file corresponding to the basic information And a verification unit that verifies the corrected voice file by receiving the voice file whose correction is completed by the correction applicant, and when the verification of the corrected voice file is completed, Of the sales revenue of the applicant.
도 3은 본 발명의 실시예에 따라 판매되는 음성파일을 도시한 도면이다.3 is a diagram showing a voice file sold according to an embodiment of the present invention.
도 3을 참고하면 텍스트를 포함한 스캔파일을 음성파일로 변환하여 상기 음성파일을 판매하여 수익의 발생하는 것을 설명할 수 있고, 하기에서는 수익에 대한 분배에 대해서 자세히 설명한다.Referring to FIG. 3, the scan file including the text is converted into a voice file, and the voice file is sold to explain the occurrence of profit. In the following, the distribution of the profit will be described in detail.
상기 TTS 변환부는, 상기 음성파일의 기설정된 판매 수량에 대해서는 순수익의 (100-X)%(여기서, X는 상기 보정신청자의 기여도를 의미함)를 보정신청자에게 제공하고, 기설정된 판매 수량을 초과하게 되는 경우, 순수익의 ((100-X)-(판매 수량-기설정된 판매 수량))%를 제공할 수 있다.Wherein the TTS conversion unit provides the correction applicant with a (100-X)% (where X represents the contribution of the correction applicant) of the net profit to the predetermined sales quantity of the audio file, (100-X) - (Sales Quantity - Pre-established Sales Quantity))% of the net profit.
여기서 상기 ((100-X)-(판매 수량-기설정된 판매 수량))%가 0%이하가 되는 경우 보정신청자에게는 수익을 제공하지 않는다.In this case, if the percentage ((100-X) - (sales quantity - pre-set sales quantity)) becomes 0% or less, no compensation is provided to the applicant for correction.
따라서, 보정신청자로 하여금 초기 수익을 높임으로써, 보정신청의 참여를 높일 수 있고, 관리자로 하여금 판매 수량이 많아질수록 수익을 높임으로써, 보정신청자와 관리자의 수익을 효율적으로 분배할 수 있다.Thus, by increasing the initial profit, the applicant can increase the participation in the application for correction, and the manager can increase the profit more as the sales quantity increases, thereby effectively distributing the profit of the applicant and the manager.
일예로, 네트워크 상의 소정 공간에 등록된 사용자에게 '위대한 똥말' 이라는 책의 스캔파일을 공지하면, 상기 '위대한 똥 말' 스캔파일의 기초정보(ex. 책의 제목, 책의 페이지 수 등)와 음성파일에 포함된 오류 %(ex. 30%, 여기서 음성파일에 포함된 오류는 보정 완료 시 기여도를 의미함)를 제공하고, 상기 소정공간에 등록된 사용자는 상기 기초정보를 기반으로 하여 음성파일의 30% 오류를 보정하는 작업을 신청하고, 상기 음성파일의 30% 오류를 보정하는 작업을 최초로 신청한 신청자에게 '위대한 똥 말' 스캔파일 및 30%의 오류가 포함된 음성파일의 원본을 전송한다.For example, if a user registered in a predetermined space on the network is notified of a scan file of a book called 'great pussy', basic information (e.g., title of a book, number of pages of a book, etc.) The error% (e.g., 30%, in this case, the error included in the voice file means the contribution when the correction is completed) included in the voice file is provided, and the user registered in the predetermined space provides the voice file 30% of the speech file and send the original of the speech file containing the 'great shit' scan file and 30% error to the first applicant who applied for the correction of the 30% error of the voice file do.
상기 원본이 보정신청자에게 전달되어, 상기 보정신청자가 상기 30%의 오류가 포함된 음성파일의 오류를 보정 완료하면, 오류를 보정한 음성파일을 음성인식서버(200)에 전달한다.The original is transmitted to the correction applicant, and when the correction applicant completes correcting the error of the voice file including the 30% error, the voice file having corrected the error is transmitted to the voice recognition server 200.
상기 음성인식서버(200)에서 상기 오류를 보정한 음성파일을 검증하고, 상기 검증이 완료되면, 상기 음성파일을 음성인식서버(200)에 공지하여 네트워크 상의 소정 공간에 등록된 다른 사용자에게 공지가 되어, 상기 사용자가 음성파일을 구매하여 재생하면 봉제 인형 로봇(100) 및 외부 단말(300)에서 소리로 출력된다.When the verification is completed, the voice recognition server (200) notifies the voice recognition server (200) of the voice file that the voice recognition server (200) corrects the error and informs another user registered in a predetermined space on the network When the user purchases and reproduces an audio file, the sound is output from the sewing robot 100 and the external terminal 300.
상기 오류가 보정된 음성파일의 판매과정에서 생긴 순수익의 기설정된 부분을 상기 보정신청자에 제공한다.And provides the correction applicant with a predetermined portion of the net profit generated in the sales process of the error-corrected voice file.
상기 30%의 오류를 보정한 보정신청자에게는 상기 오류가 보정된 음성파일의 판매수량의 10개까지는 순수익의 70(100-30)%를 제공하고, 10개를 초과한 이후부터 순수익의 ((100-30)-(판매 수량-10))%를 제공한다.The correction applicant who corrected the 30% error provides 70 (100-30%) of the net profit up to 10 of the sales quantity of the error-corrected voice file, and ((100 -30) - (sales quantity-10))%.
또 다른 일예로, 상기 30%의 오류를 보정한 음성파일의 판매 수량이 25개인 경우 오류 보정한 사용자에게는 순수익의 55((100-30)-(25-10))%를 제공한다.As another example, if the sales volume of the voice file compensated for the 30% error is 25, the error compensated user provides 55 ((100-30) - (25-10))% of the net profit.
여기서, 판매 수량이 125개를 초과하여 0%가 되는 경우 보정신청자에게는 수익을 제공하지 않는다.Here, if the sales quantity is more than 125 and becomes 0%, the compensation applicant does not provide profit.
또한, 상기 음성인식서버(200)는 상기 음성인식부(120)로부터 전달받은 입력음성파일 중 기설정된 단어가 포함되어 있는 경우 상기 단어를 추출하고, 상기 추출된 단어를 기반으로 정보를 생성하고, 상기 생성된 정보를 기반으로 분석하여 답변음성파일을 생성할 수 있다.In addition, the speech recognition server 200 may extract the word when a preset word is included in the input speech file received from the speech recognition unit 120, generate information based on the extracted word, And generate an answer voice file by analyzing based on the generated information.
일예로, 사용자가 '내일 12시에 약속 있어' 라는 음성을 말하면, 상기 봉제 인형 로봇(100)은 상기 음성을 기반으로 입력음성파일을 생성하여 음성인식서버(200)에 전달하고, 상기 음성인식서버(200)에 상기 입력음성파일 중 '내일', '12시', '약속' 이라는 단어를 추출하여 저장한다.For example, if the user speaks a voice 'promise tomorrow at 12 o'clock', the stuffed animal robot 100 generates an input voice file based on the voice and transmits it to the voice recognition server 200, 'Tomorrow', '12 o'clock', and 'appointment' of the input voice file are extracted and stored in the server 200.
사용자가 '내일 스케줄 알려줘' 라는 음성을 말하면, 상기 음성인식서버(200)에서는 답변음성파일을 생성할 때, 상기 추출하여 저장된 정보를 반영하여 '내일 스케줄은 12시에 약속이 있습니다.' 라는 답변음성파일을 생성하고, 상기답변음성파일을 봉제 인형 로봇(100)의 음성 제공부(130)로 전달하여 소리로 출력하여 상기 사용자에게 내일 스케줄을 인식하게 한다.When the user speaks a voice 'Tell me a schedule tomorrow', the speech recognition server 200 generates an answer voice file, reflects the extracted and stored information, and the 'tomorrow schedule has an appointment at 12 o'clock.' And sends the answer voice file to the voice data providing unit 130 of the stuffed animal robot 100 and outputs the answer voice to the user so that the user can recognize the schedule for tomorrow.
다시 도 1을 참고하면, 상기 외부 단말(300)은 봉제 인형 로봇(100)의 무선통신부(140)로부터 봉제 인형 로봇(100)의 신호를 전송 받고, 상기 전송1, the external terminal 300 receives a signal of the stuffed animal robot 100 from the wireless communication unit 140 of the stuffed animal robot 100,
받은 신호를 이용하여, 모바일 콘텐츠를 제공 및 상기 음성인식서버(200)와 봉제The mobile communication terminal 200 provides the mobile contents using the received signal,
인형 로봇(100)을 통신을 중계할 수 있다.The doll robot 100 can relay the communication.
도 4는 본 발명의 실시예에 따라 사용자 음성 따른 어플리케이션 캐릭터의 표정 변화를 도시한 도면이다.4 is a diagram illustrating a change in expression of an application character according to a user voice according to an embodiment of the present invention.
도 4를 참고하면, 상기 모바일 콘텐츠는 봉제 인형 로봇(100)에 표현을 어플리케이션을 통해 애니메이션으로 표현하는 기능, 상기 봉제 인형 로봇 캐릭터의 입 모양이 움직이며, 음성파일을 출력하는 기능, 어플리케이션의 봉제 인형로봇 캐릭터에게 터치를 하면, 봉제 인형 로봇(100)이 소리의 출력, 모터 구동 및 엘이디(LED) 점멸하는 기능, 어플리케이션을 실행하여 학습용 콘텐츠를 제공하는 기능 중 적어도 하나를 기능을 수행할 수 있다.Referring to FIG. 4, the mobile contents include a function of expressing the representation of the mobile robot 100 in the form of animation through an application, a function of moving the mouth of the robot character to the sewing robot, outputting a voice file, When the doll robot character is touched, the stuffed toy robot 100 can perform at least one of the functions of outputting sound, driving the motor and flashing LEDs, and providing a learning content by executing the application .
일예로, 사용자가 음성을 말하면, 상기 음성을 기반으로 한 입력음성파일을 음성인식서버(200)에서 분석하여 감정을 선택하고, 상기 선택된 감정을 외부 단말(300)의 어플리케이션에 캐릭터가 상기 분석된 감정과 일치하는 표정을 외부 단말(300)에 디스플레이한다.For example, when a user speaks a voice, an input voice file based on the voice is analyzed by the voice recognition server 200 to select an emotion, and the selected emotion is transmitted to the application of the external terminal 300 And displays a facial expression matching the emotion on the external terminal 300.
상기와 같이, 본 발명의 일실시예에서는 본 발명은 무선통신을 통해 서버에 접속되어 사용자의 음성을 분석하여 그에 상응하는 음성을 출력하는 음성인식 기능을 가진 봉제 인형 로봇을 제공할 수 있다.As described above, according to the embodiment of the present invention, the present invention can provide a plush dolphin robot having a speech recognition function, which is connected to a server through wireless communication to analyze a user's voice and output a corresponding voice.
또한, 사용자의 신체 입력이 감지되면 모터로 인한 로봇의 동작 및 엘이디(LED)가 점멸함으로써, 사용자로 하여금 흥미를 가질 수 있는 음성인식 기능을 가진 봉제 인형 로봇을 제공할 수 있다.Further, when the user's body input is detected, the operation of the robot due to the motor and the LED blink, thereby providing a sewing doll robot having a voice recognition function that can be of interest to the user.
또한, 텍스트가 포함된 파일을 음성파일로 변환하고, 상기 변환 중 생기는 오류를 수정함으로써, 사용자로 하여금 다양한 콘텐츠를 제공할 수 있는 음성인식 기능을 가진 봉제 인형 로봇을 제공할 수 있다.In addition, it is possible to provide a plush dolphin robot having a speech recognition function that allows a user to provide various contents by converting a file including a text into a voice file and correcting an error occurring during the conversion.
또한, 외부 단말과 연결되어, 봉제 인형 로봇 캐릭터로 이루어진 모바일 콘텐츠를 제공함으로써, 어린이 사용자로 하여금 흥미를 가질 수 있는 음성인식 기능을 가진 봉제 인형 로봇을 제공할 수 있다.In addition, by providing mobile contents composed of a stuffed toy robot character connected to an external terminal, it is possible to provide a stuffed toy robot having a voice recognition function, which can be of interest to a child user.
또한 본 발명의 일실시예에 따른, 음성인식 기능을 가진 봉제인형로봇의 제어방법은 다양한 컴퓨터로 구현되는 동작을 수행하기 위한 프로그램 명령을 포함하는 컴퓨터 판독 가능 매체에 기록될 수 있다. 상기 컴퓨터 판독 가능 매체는 프로그램 명령, 데이터 파일, 데이터 구조 등을 단독으로 또는 조합하여 포함할 수 있다. 상기 매체는 프로그램 명령은 본 발명을 위하여 특별히 설계되고 구성된 것들이거나 컴퓨터 소프트웨어 당업자에게 공지되어 사용 가능한 것일 수도 있다. 컴퓨터 판독 가능 기록 매체의 예에는 하드 디스크, 플로피 디스크 및 자기 테이프와 같은 자기 매체(magnetic media), CD-ROM, DVD와 같은 광기록 매체(optical media), 플롭티컬 디스크(floptical disk)와 같은 자기-광 매체(magneto-optical media), 및 롬(ROM), 램(RAM), 플래시 메모리 등과 같은 프로그램 명령을 저장하고 수행하도록 특별히 구성된 하드웨어 장치가 포함된다. 프로그램 명령의 예에는 컴파일러에 의해 만들어지는 것과 같은 기계어 코드뿐만 아니라 인터프리터 등을 사용해서 컴퓨터에 의해서 실행될 수 있는 고급 언어 코드를 포함한다.The control method of the sewing robot having the speech recognition function according to an embodiment of the present invention may be recorded in a computer-readable medium including program instructions for performing various computer-implemented operations. The computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination. The media may be program instructions that are specially designed and constructed for the present invention or may be available to those skilled in the art of computer software. Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Magneto-optical media, and hardware devices specifically configured to store and execute program instructions such as ROM, RAM, flash memory, and the like. Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like.
이상과 같이 본 발명의 일실시예는 비록 한정된 실시예와 도면에 의해 설명되었으나, 본 발명의 일실시예는 상기 설명된 실시예에 한정되는 것은 아니며, 이는 본 발명이 속하는 분야에서 통상의 지식을 가진 자라면 이러한 기재로부터 다양한 수정 및 변형이 가능하다. 따라서, 본 발명의 일실시예는 아래에 기재된 특허청구범위에 의해서만 파악되어야 하고, 이의 균등 또는 등가적 변형 모두는 본 발명 사상의 범주에 속한다고 할 것이다.While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is clearly understood that the same is by way of illustration and example only and is not to be construed as limiting the scope of the invention as defined by the appended claims. Various modifications and variations are possible in light of the above teachings. Accordingly, it is to be understood that one embodiment of the present invention should be understood only by the appended claims, and all equivalent or equivalent variations thereof are included in the scope of the present invention.
100 : 봉제 인형 로봇100: Plush doll robot
110 : 제어부110:
120: 음성인식부120:
130 : 음성제공부130:
140 : 무선통신부140:
150 : 센서부150:
160 : 모터부160:
170 : 엘이디(LED)170: LED (LED)
200 : 음성인식서버200: voice recognition server
210 : 음성인식 프로그램210: Speech recognition program
220 : TTS 변환부220: TTS conversion section
300 : 사용자 단말기300: User terminal

Claims (5)

  1. 사용자의 명령 입력에 대응하는 액션을 수행하는 제어부;A controller for performing an action corresponding to a command input of a user;
    사용자의 음성을 입력음성파일로 저장하고, 상기 입력된 음성 중 기설정된Storing the voice of the user as an input voice file,
    명령이 포함되어 있는 경우, 상기 입력음성파일을 상기 제어부로 전송하고, 상기The control unit transmits the input voice file to the control unit,
    사용자의 음성 중 기설정된 명령이 포함되어 있지 않으면 상기 입력음성파일을 음If the preset voice command is not included in the voice of the user,
    성인식서버로 전달하는 음성인식부;A speech recognition unit for delivering the speech to a ceremonial server;
    상기 입력음성파일에 대응하는 답변음성파일을 음성인식서버로부터The answer voice file corresponding to the input voice file is transmitted from the voice recognition server
    전달받고, 상기 전달받은 답변음성파일을 소리로 출력하는 음성제공부;And outputting the answer voice file as a sound;
    터치센서 및 맥박센서 중 적어도 하나의 센서를 마련하여 신체 입력을 감지At least one of a touch sensor and a pulse sensor is provided to detect a body input
    하는 센서부;;
    와이파이(Wi-Fi), 블루투스(Bluetooth) 및 NFC 중 적어도 하나의 통신을 사Wi-Fi, Bluetooth, and NFC.
    용하여 외부 단말과 무선통신을 하는 무선통신부;A wireless communication unit for performing wireless communication with an external terminal using the wireless communication unit;
    다수의 모터를 구동하여 신체의 움직임을 제어하는 모터부; 및A motor unit for driving a plurality of motors to control movement of the body; And
    상기 센서부 및 음성인식서버에서 전달받은 신호 중 엘이디(LED) 점등신호가And an LED (Light Emitting Diode) lighting signal among signals received from the sensor unit and the voice recognition server
    포함되어 있는 경우, 상기 전달받은 점등신호에 대응하는 방식으로 점등하는 엘이If it is determined that the LED is lit,
    디(LED); 를 포함하는 음성인식 기능을 가진 봉제 인형 로봇.Di (LED); A robot having a speech recognition function including a robot.
  2. 제1항에 있어서,The method according to claim 1,
    상기 음성제공부는,Wherein the voice providing unit comprises:
    사용자가 목소리를 합성하여 베이스 목소리를 선택하거나 기설정된 목소리로Users can combine voices to select a bass voice or a preset voice
    베이스 목소리를 선택할 수 있는 것을 특징으로 하는 음성인식 기능을 가진 봉제A sewing machine having a voice recognition function which can select a base voice
    인형 로봇.Doll robots.
  3. 제1항에 있어서,The method according to claim 1,
    상기 음성인식서버는,The voice recognition server comprises:
    텍스트가 포함된 스캔파일을 전송 받고, 상기 스캔파일에 포함된 텍스트를A scan file including text is received, and a text included in the scan file is transmitted
    음성파일로 변환하는 TTS 변환부를 포함하는 것을 특징으로 하는 음성인식 기능을And a TTS conversion unit for converting the voice signal into a voice file.
    가진 봉제 인형 로봇.Robot with a stuffed doll.
  4. 제3항에 있어서,The method of claim 3,
    상기 TTS 변환부는,Wherein the TTS conversion unit comprises:
    네트워크 상의 소정 공간에 등록된 사용자에게 상기 스캔파일의 기초정보를The basic information of the scan file is transmitted to a user registered in a predetermined space on the network
    제공하고, 상기 기초정보에 대응하여 상기 음성파일의 오류를 보정하는 작업을 최And an operation of correcting the error of the voice file in accordance with the basic information is defined as Choi
    초로 신청한 보정신청자에게 상기 스캔파일 및 상기 음성파일의 원본을 전송하며,The original of the scan file and the voice file is transmitted to the correction applicant who has applied for the second time,
    상기 보정신청자가 보정을 완료한 음성파일을 수신하여 상기 보정한 음성파일을 검증하고, 상기 보정한 음성파일의 검증이 완료되면 상기 보정한 음성파일의 판매 수And a verification unit configured to verify the corrected voice file by receiving the voice file whose correction has been completed by the correction applicant and verify the corrected voice file,
    익의 기설정된 부분을 상기 보정신청자에게 제공하는 것을 특징으로 하는 음성인식And provides the predetermined portion of the voice to the correction applicant.
    기능을 가진 봉제 인형 로봇.Plush dolls robot with function.
  5. 제4항에 있어서,5. The method of claim 4,
    상기 TTS 변환부는,Wherein the TTS conversion unit comprises:
    상기 음성파일의 기설정된 판매 수량에 대해서는 순수익의 (100-X)%(여기서,(100-X)% of the net profit (here,
    X는 상기 보정신청자의 기여도를 의미함) 를 상기 보정신청자에게 제공하고, 기설X represents the contribution of the correction applicant) to the correction applicant,
    정된 판매 수량을 초과하는 경우, 순수익의 ((100-X)-(판매 수량-기설정된 판매 수((100-X) - (sales quantity - the number of sales already set)
    량))%를 제공하는 것을 특징으로 하는 음성인식 기능을 가진 봉제 인형 로봇.) Of the robot is provided with a speech recognition function.
PCT/KR2018/000173 2017-12-29 2018-01-04 Plush doll robot with voice recognition function WO2019132092A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020207023831A KR20200119821A (en) 2017-12-29 2018-01-04 Plush toy robot with voice recognition function

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR20170184127 2017-12-29
KR10-2017-0184127 2017-12-29

Publications (1)

Publication Number Publication Date
WO2019132092A1 true WO2019132092A1 (en) 2019-07-04

Family

ID=67063917

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/000173 WO2019132092A1 (en) 2017-12-29 2018-01-04 Plush doll robot with voice recognition function

Country Status (2)

Country Link
KR (1) KR20200119821A (en)
WO (1) WO2019132092A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20010007842A (en) * 2000-10-06 2001-02-05 남호원 The system and method of a dialogue form voice and multi-sense recognition for a toy
US20100041304A1 (en) * 2008-02-13 2010-02-18 Eisenson Henry L Interactive toy system
JP2013099823A (en) * 2011-11-09 2013-05-23 Panasonic Corp Robot device, robot control method, robot control program and robot system
KR20170027705A (en) * 2014-04-17 2017-03-10 소프트뱅크 로보틱스 유럽 Methods and systems of handling a dialog with a robot
KR20170096502A (en) * 2016-02-16 2017-08-24 최진양 Talking doll, circuit module of talking doll and voice service system based on the same

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20010007842A (en) * 2000-10-06 2001-02-05 남호원 The system and method of a dialogue form voice and multi-sense recognition for a toy
US20100041304A1 (en) * 2008-02-13 2010-02-18 Eisenson Henry L Interactive toy system
JP2013099823A (en) * 2011-11-09 2013-05-23 Panasonic Corp Robot device, robot control method, robot control program and robot system
KR20170027705A (en) * 2014-04-17 2017-03-10 소프트뱅크 로보틱스 유럽 Methods and systems of handling a dialog with a robot
KR20170096502A (en) * 2016-02-16 2017-08-24 최진양 Talking doll, circuit module of talking doll and voice service system based on the same

Also Published As

Publication number Publication date
KR20200119821A (en) 2020-10-20

Similar Documents

Publication Publication Date Title
Bevilacqua et al. Wireless sensor interface and gesture-follower for music pedagogy
Pieraccini The voice in the machine: building computers that understand speech
CN100352622C (en) Robot device, information processing method, and program
KR100906136B1 (en) Information processing robot
WO2002045916A1 (en) Robot device, method for controlling motion of robot device, and system for controlling motion of robot device
JP2017201342A (en) Language Learning Robot Software
JP5404781B2 (en) Interactive toys
JP5020593B2 (en) Foreign language learning communication system
JP2011528246A5 (en)
WO2020159073A1 (en) Conversation-based foreign language learning method using reciprocal speech transmission through speech recognition function and tts function of terminal
WO2019132092A1 (en) Plush doll robot with voice recognition function
JP2001242780A (en) Information communication robot device, information communication method, and information communication robot system
WO2015037871A1 (en) System, server and terminal for providing voice playback service using text recognition
US20230230493A1 (en) Information Processing Method, Information Processing System, and Recording Medium
US20210319715A1 (en) Information processing apparatus, information processing method, and program
KR20010007842A (en) The system and method of a dialogue form voice and multi-sense recognition for a toy
Li et al. Designing a realistic peer-like embodied conversational agent for supporting children\textquotesingle s storytelling
Angulo et al. Aibo jukeBox–A robot dance interactive experience
WO2015076483A1 (en) Control system for toys through scenario command
WO2020111835A1 (en) User device and education server included in conversation-based education system
KR20020068835A (en) System and method for learnning foreign language using network
US20040072498A1 (en) System and method for controlling toy using web
KR20200085433A (en) Voice synthesis system with detachable speaker and method using the same
KR100591465B1 (en) Network based robot system playing multimedia content having motion information selected by the optical identification device
KR20200064021A (en) conversation education system including user device and education server

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18897167

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18897167

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 21/01/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18897167

Country of ref document: EP

Kind code of ref document: A1