EP2321817A1 - Système d apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l éducation des enfants - Google Patents

Système d apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l éducation des enfants

Info

Publication number
EP2321817A1
EP2321817A1 EP09770405A EP09770405A EP2321817A1 EP 2321817 A1 EP2321817 A1 EP 2321817A1 EP 09770405 A EP09770405 A EP 09770405A EP 09770405 A EP09770405 A EP 09770405A EP 2321817 A1 EP2321817 A1 EP 2321817A1
Authority
EP
European Patent Office
Prior art keywords
child
robot
learning
cpu
letter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP09770405A
Other languages
German (de)
English (en)
Other versions
EP2321817A4 (fr
Inventor
Kyung-Chul Shin
Seong-Ju Park
Kyoung-Seon Lee
Eun-Ja Hyun
So-Yune Kim
Sie-Kyung Jang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yujin Robot Co Ltd
Original Assignee
Yujin Robot Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yujin Robot Co Ltd filed Critical Yujin Robot Co Ltd
Publication of EP2321817A1 publication Critical patent/EP2321817A1/fr
Publication of EP2321817A4 publication Critical patent/EP2321817A4/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H3/00Dolls
    • A63H3/28Arrangements of sound-producing means in dolls; Means in dolls for producing sounds
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B7/00Electrically-operated teaching apparatus or devices working with questions and answers
    • G09B7/02Electrically-operated teaching apparatus or devices working with questions and answers of the type wherein the student is expected to construct an answer to the question which is presented or wherein the machine gives an answer to the question presented by a student
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H2200/00Computerized interactive toys, e.g. dolls

Definitions

  • the present invention relates to an interactive learning system using a robot and a method of operating the same. More particularly, the present invention relates to an interactive learning system using a robot and a method of operating the same, which develops suitable content for an interaction function of the robot and provides an interactive learning situation between a child and the robot while the child learns using the robot, instead of allowing the robot to present Internet content per se or to perform a unilateral action. Since learning development during babyhood is achieved through direct experience and interaction, the present invention provides an enhanced learning system adapted to babyhood based on interaction between a robot and a child.
  • the media involve all types and channels used for delivery in a process of delivering information. Further, it is known that the media used in a learning situation represents channels for educational interaction between a teacher and a child. Particularly, when the media are used in delivering content for a certain instructional purpose, they are called instructional media.
  • Korean Patent Application No. 10-2006-0118105 (entitled “robot system for learning-aids and teacher-assistants” and hereinafter referred to as a "prior invention”) discloses a robot capable of improving the aforementioned educational effect.
  • the above robot system provide self-control and a creative learning environment while freely moving in a classroom and having a simple conversation with students, it also behaves like a person assisting a teacher, thereby effectively performing education based on information & communication technology (ICT) introduced to improve educational efficiency for the students.
  • ICT information & communication technology
  • the robot system of the prior invention unilaterally provides learning content through a robot and a projection television, and a student concentrates on and learns the provided content, so that educational effect thereof can be enhanced when the student over a certain age is a recipient of the education.
  • a child under a certain age is a recipient of this system, the system is unsuitable for the child in view of concentration duration, habit, and the like.
  • one important factor for a child in learning with this system is that the child must voluntarily participate in learning.
  • one-way education which provides only the educational content as in the robot system of the prior invention, it is difficult to satisfy the factor in learning.
  • the present invention is conceived to solve the problems of the related art.
  • An aspect of the invention is to provide an interactive learning system and a method of operating the same, which enable interaction between a child and a robot to be achieved while teaching the child using the robot to induce the child to actively participate in learning and extend learning duration by causing the child to recognize the robot as a playmate.
  • Another aspect of the invention is to provide an interactive learning system and a method of operating the same, which enables interaction between a child and a robot to be achieved through letter board recognition and total physical response teaching while teaching the child using the robot to induce the child to actively participate in learning and extend learning duration by inducing the child to raise a letter board on which a letter is written when the letter is output as a sound or image of the robot or teaching materials.
  • the invention provides an interactive learning system using a robot which outputs a sound and an image to teach a child and performs travel and motion by driving a motor.
  • the interactive learning system includes: an input unit that inputs a signal using a sensor for sensing a figure and voice of a child and a teaching material selected to teach a child and that inputs content to teach a child or inputs a signal when directly touched by a child during learning; a central processing unit (CPU) that determines an action corresponding to a signal input through the input unit and controls elements to progress a process of teaching a child; a data table that provides control data called by the CPU when a signal input with respect to a child is generated in the CPU; and a drive unit that receives a control signal and drives an audio unit, a video unit, and the motor of the robot to perform activities of the robot required for learning when the CPU transmits the control signal based on the control data provided from the data table.
  • CPU central processing unit
  • the data table includes a learning data table that stores data of a process for progressing the learning using the teaching material and stores standardized data of activities of the robot while performing the process; an action pattern data table that previously stores an action pattern of a child, which can be generated during learning, and stores the action pattern of the child generated in real time and observed by the input unit; a corresponding pattern data table that previously stores data about activities of the robot to be performed corresponding to the action pattern of a child and stores a situation when the activities of the robot are performed by a corresponding approximate value corresponding to the action pattern of the child generated in real time; and a driving data table that previously stores basic data used by the CPU to drive the robot to do the activities.
  • the CPU may further include a storage medium to store a total learning process performed by the robot and a new data input generated from an outside.
  • the invention provides a method of operating an interactive learning system using a robot.
  • the method includes: if an input unit senses appearance of a child and inputs an appearance signal to a central processing unit (CPU), inducing the child to voluntarily greet the robot to perform a greeting process in response to activities of the robot according to an instruction of the CPU and presenting a teaching material to the child through activities of the robot according to an instruction of the CPU to induce the child to select the teaching material; sensing the teaching material selected by the child through the input unit, calling, by the CPU, learning data from a learning data table based on the sensed data and allowing a reading process to be performed according to a voice of the robot to progress the learning based on the teaching material selected by the child; while or after reading the teaching material, performing extended activities other than reading based on the learning data of the learning data table through activities of the robot according to an instruction of the CPU; if the learning process is completed, performing a finishing activity to notify completion of the learning through activities of the robot according to an instruction of the
  • the CPU may call the action pattern data of the child and the corresponding pattern data from the action pattern data table and allow the robot to respond to the expected action of the child through activities of the robot according to the instruction of the CPU, and if an unexpected action of the child is sensed by the input unit, the CPU may determine an approximate pattern value previously set and input to the action pattern data and the corresponding pattern data and allow activities of the robot based on the approximate pattern value to be performed.
  • the reading process may include a basic reading process where teaching material content is read and a careful reading process where at least one of re-reading, reading together, section reading, enunciated reading, and repeating is performed while performing the basic reading process.
  • the performing extended activities includes at least one of a process of reading the teaching material content by changing a letter into an illustration, a process of changing the read letter into an illustration, a process of outputting the read letter as an image in a stroke order, and a process of expressing encouragement if the child says his or her impressions after reading the teaching material.
  • the invention provides an interactive learning system based on a total physical response (TPR) learning model using a robot that outputs a sound and an image to teach a child and performs travel and motion by driving a motor.
  • the system includes: a letter board on which a certain letter is previously written to be provided by the robot to a child in the form of a sound, an image or combination of the sound and the image; and a robot that recognizes the letter written on the letter board and induces a child to raise the letter board, so that at least one of the sound, the image, travel and motion can be provided to the child according to results of reading the letter from the letter board to respond to the results.
  • TPR total physical response
  • the robot includes: an input unit that includes a letter scanner to read the letter on the letter board when a child raises the letter board, inputs a signal using a sensor for sensing a figure and voice of a child and a teaching material selected to teach a child, and inputs content to teach a child or inputs a signal when directly touched by a child during learning; a central processing unit (CPU) that determines an action corresponding to a signal input through the input unit and controls elements to progress a process of teaching a child; a data table that provides control data called by the CPU when a signal input with respect to a child is generated in the CPU; and a drive unit that receives a control signal and drives an audio unit, a video unit, and the motor of the robot to perform activities of the robot required for learning when the CPU transmits the control signal based on the control data provided from the data table.
  • an input unit that includes a letter scanner to read the letter on the letter board when a child raises the letter board, inputs a signal using a sensor for
  • the data table may include a letter scan data table that stores input data about a shape of the letter to recognize the letter written on the letter board; a learning data table that stores data of a process for progressing the learning with the teaching material and stores standardized data of activities of the robot while performing the process; an action pattern data table that previously stores an action pattern of a child, which can be generated during learning, and stores the action pattern of the child generated in real time and observed by the input unit; a corresponding pattern data table that previously stores data about activities of the robot to be performed corresponding to the action pattern of a child and stores a situation when the activities of the robot are performed by a corresponding approximate value corresponding to the action pattern of the child generated in real time; and a driving data table that previously stores basic data for allowing the CPU to drive the robot to perform the activities.
  • the invention provides a method of operating an interactive learning system using a robot that outputs a sound and an image and performs travel and motion by driving a motor.
  • the method includes: performing a learning process using a teaching material for allowing a child to learn reading and vocabulary through the robot; while or after reading the teaching material, inducing a child to raise a letter board, on which a certain letter is written, through a voice, an image or combination of the voice through a voice according to an instruction of a central processing unit (CPU) based on learning data of a learning data table; and if the child raises the letter board, allowing a letter scanner to read the letter on the letter board and transmit a reading signal to the CPU, and calling data from the letter scan data table and expressing determined results about the letter selected by the child through activities of the robot according to an instruction of the CPU.
  • CPU central processing unit
  • the performing the learning may include: if an input unit senses appearance of a child and inputs an appearance signal to a central processing unit (CPU), inducing the child to voluntarily greet the robot to perform a greeting process in response to activities of the robot according to an instruction of the CPU and presenting a teaching material to the child through activities of the robot according to an instruction of the CPU to induce the child to select the teaching material; sensing the teaching material selected by the child through the input unit, calling, by the CPU, learning data from a learning data table based on the sensed data and allowing a reading process to be performed according to a voice of the robot to progress the learning based on the teaching material selected by the child; while or after reading the teaching material, performing extended activities other than reading based on the learning data of the learning data table through activities of the robot according to an instruction of the CPU; if the learning process is completed, performing a finishing activity to notify completion of the learning through activities of the robot according to an instruction of the CPU; and, when switching between the foregoing operations, performing a switching operation to attract
  • the CPU may call action pattern data of the child and corresponding pattern data from an action pattern data table and allow the robot to respond to the expected action of the child through activities of the robot according to an instruction of the CPU, and if an unexpected action of the child is sensed by the input unit, the CPU may determine an approximate pattern value previously set and input to the action pattern and the corresponding pattern data and allow activities of the robot based on the approximate pattern value to be performed.
  • the performing the extended activities may include performing total physical response (TPR) learning by making the child act corresponding to the letter after the child reads the letter raised by the child or by inducing the child to follow action of the robot after the robot acts corresponding to the letter.
  • TPR total physical response
  • the learning system employs an instructional design using an intelligent robot when educating a child such that the child recognizes the robot as a playmate, thereby increasing attention span during learning through natural participation and providing more effective learning.
  • the learning system allows a child to learn a language through physical activity based on playing-mode learning with a letter board, thereby ensuring more effective learning through a total physical response (TPR) method.
  • TPR total physical response
  • Fig. 1 is a block diagram of an interactive learning system according to one exemplary embodiment of the present invention.
  • Fig. 2 is a block diagram of data tables in the interactive learning system according to one exemplary embodiment of the present invention.
  • Fig. 3 illustrates a robot to which the interactive learning system of Fig. 1 is applied.
  • Fig. 4 is a table showing operating status of the robot of Fig. 3.
  • Fig. 5 shows a flow of operating the interactive learning system according to one exemplary embodiment of the present invention
  • Fig. 6 shows a child selecting teaching material through a robot during operation of the interactive learning system according to one exemplary embodiment of the present invention.
  • Fig. 7 is a block diagram of an interactive learning system according to another exemplary embodiment of the present invention.
  • Fig. 8 is a block diagram of data tables in the interactive learning system according to another exemplary embodiment of the present invention.
  • Fig. 9 is a flowchart of letter board learning operation in the interactive learning system according to another exemplary embodiment of the present invention.
  • Fig. 10 shows the letter board learning operation in the interactive learning system according to another exemplary embodiment of the present invention.
  • Fig. 11 shows results from the learning operation of Fig. 10.
  • Fig. 1 is a block diagram of an interactive learning system according to one exemplary embodiment of the present invention
  • Fig. 2 is a block diagram of data tables in the interactive learning system according to one exemplary embodiment of the present invention.
  • an interactive learning system includes an input unit 10, a central processing unit (CPU) 20, a data table 30, and a drive unit 40.
  • This system is embodied by an internal circuit of a robot.
  • the robot includes a speaker 1 for outputting a sound; a monitor 2 and a light emitting diode (LED) module 3 for outputting an image; and a travel unit 4 and a motion unit 5 for performing travel and motion by driving a motor.
  • the travel unit 4 includes a motor, a wheel, a power transmission unit, etc.
  • the motion unit 5 is configured to express behaviors of an arm, a leg, winking, etc., using a motor (or a power means such as a cylinder or the like).
  • the robot using the motion unit 5 includes an emotion expression module, an image composite module, and a driving module.
  • a robot has a face, a head, arms, and a wheel, which can be driven in various ways as shown in Fig. 4 to achieve basic actions for driving the travel unit 4 and the motion unit 5 described below.
  • the head, the arm, the wheel or the like may be driven by the power means, such as the motor, and change in the face may be expressed by a liquid crystal display (LCD) or a light emitting diode (LED) module provided in the form of eyes on the face.
  • LCD liquid crystal display
  • LED light emitting diode
  • the input unit 10 includes a sensor 11, a microphone 12, a camera 15, and a touch screen 16.
  • the input unit 10 employs the camera 15 to sense a figure of a child who uses the robot for learning and a teaching material selected for teaching the child; the microphone 12 to sense a voice of the child; and the sensor 11 to sense child'a entrance, exit, movement or the like.
  • a child can input a signal through an interface unit 22.
  • the respective elements of the input unit 10 are generally installed throughout the robot from head to body, but for convenience of maintenance, all elements may be installed in the body.
  • the sensor 11 includes an image sensor for sensing a figure of a child, teaching materials, etc., and a stereo microphone for sensing a position of a sound source based on a child's voice.
  • an amplifier 13 and a converter 14, connected to the sensor 11 and the microphone 12, are configured to amplify an analog signal and convert it into a digital signal.
  • the touch screen 16 allows a child's guardian or educator to input educational content for a child or allows a child to directly touch and input a signal while progressing the learning.
  • the CPU 20 may include a microcomputer, a microprocessor, or the like to determine whether action corresponds to a signal input through the input unit 10, and to progress a process for teaching a child.
  • the CPU 20 accesses a personal computer (PC), the Internet or a network via a communication line when receiving learning data or the like, and is connected with a universal serial bus (USB) port (mostly, installed in the body of the robot) or the like when updating the data.
  • PC personal computer
  • USB universal serial bus
  • the CPU 20 includes a storage medium 21 such as a memory, a hard disk and a compact disc (CD) (involving a CD player) to store an overall learning process performed by the robot and new data input from outside.
  • a storage medium 21 such as a memory, a hard disk and a compact disc (CD) (involving a CD player) to store an overall learning process performed by the robot and new data input from outside.
  • CD compact disc
  • the interface unit 22 transmits the input data of the input unit 10 in a form that can be easily processed by the CPU 20.
  • the data table 30 stores various control data called by the CPU 20 when a signal input with respect to a child is generated in the CPU 20.
  • the data table 30 includes a learning data table 31, an action pattern data table 32, a corresponding pattern data table 33, and a driving data table 34.
  • the learning data table 31 stores data of a process for progressing learning with the teaching material and stores standardized data of activities of the robot while performing the process.
  • the action pattern data table 32 previously stores an action pattern of a child which can be generated during learning and stores the action pattern of the child generated in real time and observed by an associated element of the input unit 10.
  • the corresponding pattern data table 33 previously stores data about activities of the robot to be performed corresponding to the action pattern of a child and stores a situation when the activities of the robot is performed by a corresponding approximate value corresponding to the action pattern of the child generated in real time.
  • the driving data table 34 previously stores basic data for allowing the CPU 20 to drive the activities of the robot.
  • the data table 34 serves along with the CPU 20 as a controller or an actual operation processor to actually treat an educational interaction.
  • the data table may be a readable and writable memory, hard disk or the like.
  • the drive unit 40 receives the control signal and drives an audio unit, a video unit and the motor of the robot based on the control signal, thereby allowing the robot to output a sound and an image or to perform travel and motion (hereinafter, referred to as "activities of the robot").
  • the sound and the image output by the robot may be achieved by the speaker 1, the monitor 2 and the LED module 3, and the travel and the motion of the robot may be achieved by the travel unit 4 and the motion unit 5.
  • the drive unit 40 includes an audio driver 41, a video driver 42, and a motor driver 43.
  • the respective elements of the robot for operating the interactive learning system have functions as shown in Table 1.
  • Fig. 5 shows a flow of operating the interactive learning system according to one exemplary embodiment of the present invention.
  • the learning system senses appearance of a child through the input unit 10 and inputs an appearance signal to the CPU 20 to induce the child to voluntarily greet the robot through activities of the robot according to an instruction of the CPU 20, thereby performing a greeting process at operation S1.
  • the robot presents a teaching material to the child through activities of the robot according to an instruction of the CPU 20 and induces the child to select the teaching material at operation S2.
  • the teaching material generally includes a book such as a picture book or the like, and classification data relating to a cover, pages and the like of the book are previously stored in the learning data table 31 so as to be identified by the sensor 11 or the camera 15 of the input unit 10.
  • the input unit 10 senses the selected teaching material through the sensor 11, the camera 15, etc., and the CPU 20 calls learning data from the learning data table 31 based on the sensed data and allows a reading process to be processed according to a voice of the robot to progress the learning based on the teaching material selected by the child at operation S3-S4.
  • the reading process During the reading process, the child reads the book or the teaching material while watching teaching material content output through the monitor 2.
  • the reading process is divided into a basic reading process at operation S3 and a careful reading process where at least one of re-reading, reading together, section reading, enunciated reading, and repeating is performed while performing the basic reading process at operation S4.
  • the basic reading is a process for reading the teaching material by a simple pattern; the reading together is performed as the robot outputs a sound such as "shall we read together?" or the like through the speaker 1; the section reading allows a letter, which is difficult to pronounce, or helpful vocabulary, to be displayed and read letter by letter; and the enunciated reading is a process of reading a certain word while changing a tone or a color output as an image.
  • the careful reading process may be achieved by performing the one or more processes simultaneously or independently.
  • the robot While or after reading the teaching material, the robot performs extended activities other than reading based on the learning data of the learning data table 31 through activities of the robot according to an instruction of the CPU 20 at operation S5.
  • Such extended activities may include a process of reading the teaching material content by changing a letter into an illustration, a process of changing the read letter into an illustration, a process of outputting the read letter as an image in a stroke order, a process of expressing encouragement if a child says his or her impressions after reading the teaching material, etc.
  • the processes of the extended activities may be achieved by performing the one or more processes simultaneously or independently.
  • the robot performs a finishing activity to notify completion of learning through activities of the robot according to an instruction of the CPU 20 at operation S6.
  • the finishing activity may induce a farewell address as in the foregoing greeting process or induce the child to answer a question about repetition of the learning process.
  • the robot performs a switching operation to attract a child's attention through activities of the robot according to an instruction of the CPU 20 based on the data stored in the data table 30 when switching between the operations at operation S7.
  • the switching operation may be achieved by an activity of outputting music to a child through the speaker 1 or allowing a child to observe an interesting image on the monitor 2. Furthermore, the switching operation may be achieved by inducing the travel unit 4, the motion unit 5, etc. of the robot to perform a certain action.
  • the CPU 20 calls the action pattern data of the child and the corresponding pattern data from the action pattern data table 32 and allows the robot to respond to the expected action of the child through activities of the robot according to an instruction of the CPU 20.
  • the CPU 20 determines an approximate pattern value previously set and input to the action pattern data and the corresponding pattern data and allows the activities of the robot based on the approximate pattern value to be performed.
  • the approximate pattern value is the most approximate to the previously input data in the action pattern data table 32 with regard to the unexpected action of a child, as determined by the CPU 20.
  • the activities of the robot are performed by calling the corresponding pattern data 33 corresponding to the approximate pattern value.
  • the activities of the robot may be achieved by performing at least one of the audio output, the video output, and the travel and motion of the robot simultaneously or independently through the drive unit 40.
  • test For the evaluation, four-year-old children were recruited and divided into an evaluation experimental group (for the learning system according to one embodiment of the invention) and a control group (for a multimedia-type learning system using a computer system). The test was carried out using the same book in a comfortable room for the children.
  • test results of the experimental group and the control group were obtained based on statistics by a well-known official test method.
  • the experimental group and the control group have levels (t) of significance of .008 and .93 in the post-test tale-understanding, respectively, which shows that only the experimental group is significantly increased in the tale-understanding as compared with the pre-test tale-understanding.
  • the experimental group and the control group have levels (t) of significance of .00 and 1.00 in the post-test story composing ability, respectively, which shows that only the experimental group is significantly and largely increased in the story composing ability as compared with the pre-test story composing ability.
  • the experimental group and the control group have levels (t) of significance of .00 and .11 in the post-test word reading ability, respectively, which shows that only the experimental group is significantly and largely increased in the word reading ability as compared with the pre-test word reading ability
  • Fig. 7 is a block diagram of an interactive learning system according to another exemplary embodiment of the invention
  • Fig. 8 is a block diagram of data tables in the interactive learning system according to another exemplary embodiment of the invention.
  • the interactive learning system includes a letter board L and a robot R.
  • the letter board L On the letter board L, a certain letter is written to be provided by the robot to a child in the form of a sound, an image or combination of the sound and the image.
  • the letter board L is made of a light nontoxic material so as to be easily raised by a child and is rounded at corners thereof to prevent a child form getting hurt.
  • a background color and a color of a letter written on the letter board L are set to be photographed or scanned by a letter scanner 17 of the robot R described below.
  • the robot R recognizes the letter written on the letter board L and induces a child to raise the letter board L, so that at least one of the sound, the image, the travel and the motion can be provided to the child according to results of reading the letter from the letter board L, thereby responding to the results.
  • the robot R includes configurations required for a child'a learning, which include an input unit 10, a central processing unit 20, a data table 30 and a drive unit 40.
  • the camera 15 performs not only the foregoing function but also a function of inputting a letter image of the letter board L, so that the letter scanner 17 can read the letter on the letter board L based on a letter image signal.
  • the letter scanner 17 may include a video signal processor or the like to recognize the previously written letter.
  • the data table 30 is configured to store various control data called by the CPU 20 when a signal input with respect to a child is generated in the CPU 20.
  • the data table 30 includes a letter scan data table 35, a learning data table 31, an action pattern data table 32, a corresponding pattern data table 33, and a driving data table 34.
  • the letter scan data table 35 stores input data about the shape of the letter so as to recognize the letter written on the letter board.
  • each element of the robot has the same function as those of Table 1, but there is a difference in that the motion control in Table 1 employs the TPR method, and object recognition is required for the learning using the letter board.
  • Fig. 9 is a flowchart of a letter board learning operation in the interactive learning system according to another exemplary embodiment of the present invention. Also, Tables 6 to 9 are a story board showing a process of performing the letter board learning operation and the TPR method.
  • a first item of the table is a progressing order of the story and a second item shows a scene for progressing the story. Further, a third item is a scenario and a fourth item shows cautions of the scenario with regard to each scene.
  • Table 6 7 Recognized letter board withsentence ⁇ Put botharms up ⁇ (if a child inputs a letter board)[TTS] hmm, it's a (directive)/I will dothis![Face] happy[Head] nodding[Arms] both arms are waved up and down[TTS] Roby ⁇ / Roby ⁇ (move by instruction)(ex., if a card "put your arms up” isinput[TTS] put both arms/ up[Arms] both arms are put up[Face] happy)[TTS] ok/ it's ok/ to do like this. [Face] normal[Head] nodding[TTS] Please select/ plays/ for me/ oneby one/ okay?
  • Table 9 11 A child's face appears(picture in anattendance book isusable) ⁇ child's picture ⁇ [TTS] (name)!/ it is very/ interesting[Face] happy[Arms] swing cross[TTS] Now/ look at/my stomach, and moveaccording to theletter board on thescreen[Face] happy[TTS] Now/ are youready?[Arms] swing alittle forward andbackward 12 ⁇ Give me a handclap ⁇ Should appear everytime ⁇ check[TTS] (name) ⁇ / (name) ⁇ [Face] nodding[Arms] swing cross(the screen shows achild one of theaction lists)(one second after)[TTS] (reading theletter board)[Face] nodding[Arms] cross swingCheck [Screen] a key word appears on thescreen while being pronounced, anda hint image is shown together withreading the other sentence.
  • the letter board learning operation is performed to further improve learning efficiency during the interactive learning operation.
  • Such a letter board learning operation may be performed in the course of the basic reading process, the careful reading process, and the extended reading process.
  • the CPU 20 While or after reading the teaching material, the CPU 20 provides a certain letter through a voice, an image or combination of the voice and the image based on the learning data of the learning data table 31 at operation S10.
  • the letter scanner 17 reads the letter on the letter board L and transmits a reading signal to the CPU 20 at operation S12.
  • the reading of the letter on the letter board L may be achieved by processing a video signal input by the camera 15 in the letter scanner 17 and inputting it to the CPU 20.
  • the CPU 20 After receiving the signal, the CPU 20 calls data from the letter scan data table 35 and expresses determined results about the letter selected or written by the child through activities of the robot at operation S13.
  • the result expression may be achieved by driving the motion unit 5 of the robot to change the expression or the like of the robot along with the expression such as "Great” or "Well, that is not the letter. Let's try again” through a sound or an image from the robot R.
  • the driving of the motion unit 5 of the robot may be applied to the total physical response (TPR) learning model according to recognition of the letter board. That is, the TPR learning model causes the robot induce a child to act as shown in Tables 6 to 9 in response to sentences displayed on the letter board, thereby improving language learning efficiency.
  • TPR total physical response
  • the robot R induces the child to read the letter through the foregoing processes and act corresponding to the read letter.
  • the robot R performs the action corresponding to the letter through the motion unit 5, thereby making a child follow it.
  • the robot R reads the letter through the foregoing processes and performs an action of shaking its body through the motion unit 5, in order to induce a child to follow the action, thereby enabling the TPR learning.
  • Example 1 where development in tale understanding ability of a child was evaluated by operating the learning system using the robot
  • Example 2 where development in story composing ability of a child was evaluated by the same process as Example 1
  • Example 3 where development in word reading ability of a child was evaluated, show the same results as Examples 1 to 3 according to the foregoing embodiment shown in Tables 3 to 5, and thus repetitious descriptions thereof will be omitted herein.
  • the interactive learning system of this embodiment is also operated using the robot, so that a child group having experienced the interactive learning is superior to a child group having experienced the unilateral learning based on the multimedia such as a computer system in tale understanding ability, story composing ability, and word reading ability.
  • the multimedia such as a computer system in tale understanding ability, story composing ability, and word reading ability.
  • the interactive learning by the system according to this embodiment can be used with very high efficiency in many educational institutions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Physics & Mathematics (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Toys (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

L’invention concerne un système d’apprentissage interactif utilisant un robot, comprenant une unité d’entrée (10) permettant l’entrée d’un signal au moyen d’un détecteur conçu pour détecter la silhouette et la voix d’un enfant et du matériel pédagogique et permettant l’entrée d’un contenu utilisé pour l’apprentissage de l’enfant ou permettant l’entrée d’un signal par contact direct de l’enfant; une CPU (20) qui détermine une action correspondant à un signal entré par l’unité d’entrée et qui commande des éléments en vue de faire avancer un processus d’apprentissage; une table de données (30) qui renvoie des données de commande sollicitées par la CPU (20) lorsque celle-ci génère un signal correspondant à une entrée par l’enfant; et une unité pilote (40) qui reçoit un signal de commande et pilote une unité audio, une unité vidéo et un moteur du robot permettant au robot de réaliser des activités relatives à l’apprentissage lorsque la CPU (20) transmet un signal de commande basé sur les données de commande renvoyées par la table de données (30). L’invention concerne également un procédé de fonctionnement du système d’apprentissage interactif.
EP09770405.0A 2008-06-27 2009-06-26 Système d apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l éducation des enfants Withdrawn EP2321817A4 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR20080062034 2008-06-27
KR20080062033 2008-06-27
PCT/KR2009/003464 WO2009157733A1 (fr) 2008-06-27 2009-06-26 Système d’apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l’éducation des enfants
KR20090057392A KR101088406B1 (ko) 2008-06-27 2009-06-26 유아교육시 로봇을 이용한 양방향 학습 시스템 및 그 운영방법

Publications (2)

Publication Number Publication Date
EP2321817A1 true EP2321817A1 (fr) 2011-05-18
EP2321817A4 EP2321817A4 (fr) 2013-04-17

Family

ID=41812317

Family Applications (1)

Application Number Title Priority Date Filing Date
EP09770405.0A Withdrawn EP2321817A4 (fr) 2008-06-27 2009-06-26 Système d apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l éducation des enfants

Country Status (4)

Country Link
EP (1) EP2321817A4 (fr)
KR (1) KR101088406B1 (fr)
CN (1) CN102077260B (fr)
WO (1) WO2009157733A1 (fr)

Families Citing this family (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102236981A (zh) * 2010-04-27 2011-11-09 上海以太软件有限公司 具有大尺寸液晶屏和触摸屏的数码智力开发机
KR101119030B1 (ko) * 2010-05-12 2012-03-13 (주) 퓨처로봇 지능형 로봇 장치의 서비스 시나리오 편집 방법, 그 방법을 실행하기 위한 프로그램을 기록한 컴퓨터 판독가능한 기록매체, 지능형 로봇 장치 및 지능형 로봇의 서비스 방법
CN101833884B (zh) * 2010-05-17 2011-12-07 博雅创世(北京)智能科技有限公司 一种机器人教育平台
CN201940040U (zh) * 2010-09-27 2011-08-24 深圳市杰思谷科技有限公司 家用机器人
WO2012056459A1 (fr) * 2010-10-28 2012-05-03 Visionstory Ltd Appareil pour l'éducation et le divertissement
US9126122B2 (en) * 2011-05-17 2015-09-08 Zugworks, Inc Doll companion integrating child self-directed execution of applications with cell phone communication, education, entertainment, alert and monitoring systems
CN102522008A (zh) * 2011-11-23 2012-06-27 康佳集团股份有限公司 一种多媒体互动教学方法及其系统、一种电视机
KR101333532B1 (ko) * 2011-12-05 2013-11-28 (주)세스넷 학습 장치 및 학습 방법
KR101209012B1 (ko) * 2012-01-31 2012-12-24 한성대학교 산학협력단 캐릭터 로봇을 이용한 교육용 연극 인터페이스 장치
KR101344727B1 (ko) * 2012-03-02 2014-01-16 주식회사 유진로봇 지능형 로봇 제어 장치 및 방법
CN102663904A (zh) * 2012-04-20 2012-09-12 江苏奇异点网络有限公司 儿童娱乐系统
CN102819969B (zh) * 2012-08-15 2014-11-26 魔方天空科技(北京)有限公司 多媒体教育平台的实现方法及多媒体教育平台系统
KR101515178B1 (ko) * 2013-01-14 2015-04-24 주식회사 케이티 얼굴에 기반하는 사용자 인터페이스를 제공하는 로봇 및 로봇의 제어 방법
KR101544044B1 (ko) 2013-09-16 2015-08-13 이호현 교육특화디바이스용 확장형 교구 로봇
CN103777595A (zh) * 2013-12-30 2014-05-07 深圳市德宝威科技有限公司 机器人系统及机器人办公、教学、设计、工程、家庭系统
CN104252287A (zh) * 2014-09-04 2014-12-31 广东小天才科技有限公司 一种交互装置和基于交互装置的提高表达能力的方法
CN104575141A (zh) * 2015-01-20 2015-04-29 三峡大学 人机互动课堂辅助教学仪
WO2016206643A1 (fr) * 2015-06-26 2016-12-29 北京贝虎机器人技术有限公司 Procédé et dispositif de commande de comportement interactif de robot et robot associé
CN105872828A (zh) * 2016-03-30 2016-08-17 乐视控股(北京)有限公司 一种电视交互学习的方法及装置
CN105719519A (zh) * 2016-04-27 2016-06-29 深圳前海勇艺达机器人有限公司 一种分等级教学功能的机器人
KR101904453B1 (ko) * 2016-05-25 2018-10-04 김선필 인공 지능 투명 디스플레이의 동작 방법 및 인공 지능 투명 디스플레이
CN105894873A (zh) * 2016-06-01 2016-08-24 北京光年无限科技有限公司 一种面向智能机器人的儿童教学方法和装置
CN106057023A (zh) * 2016-06-03 2016-10-26 北京光年无限科技有限公司 一种面向智能机器人的儿童教学方法和装置
KR101983728B1 (ko) * 2016-07-15 2019-06-04 주식회사 토이트론 카드를 인식하여 명령을 수행하는 스마트 토이를 운용하는 장치 및 그 방법
CN106097793B (zh) * 2016-07-21 2021-08-20 北京光年无限科技有限公司 一种面向智能机器人的儿童教学方法和装置
CN107784354B (zh) 2016-08-17 2022-02-25 华为技术有限公司 机器人的控制方法及陪伴机器人
CN106295217A (zh) * 2016-08-19 2017-01-04 吕佳宁 一种孕育机器人
CN106205237A (zh) * 2016-08-31 2016-12-07 律世刚 基于动作反应和绘画反应的第二母语的训练方法及装置
WO2018044230A1 (fr) * 2016-09-02 2018-03-08 Tan Meng Wee Appareil et système robotisés de formation
CN106297436A (zh) * 2016-09-12 2017-01-04 上海夫子云教育投资股份有限公司 一种智能机器人视频课程推送系统
CN106251717A (zh) * 2016-09-21 2016-12-21 北京光年无限科技有限公司 智能机器人语言跟读学习方法和装置
CN106393113A (zh) * 2016-11-16 2017-02-15 上海木爷机器人技术有限公司 机器人和机器人的交互控制方法
CN107067835A (zh) * 2016-11-23 2017-08-18 河池学院 一种儿童语音教育机器人
KR20180089667A (ko) 2017-02-01 2018-08-09 주식회사 시공미디어 코딩 교육 제공을 위한 로봇
CN107030691B (zh) 2017-03-24 2020-04-14 华为技术有限公司 一种看护机器人的数据处理方法及装置
CN107369341A (zh) * 2017-06-08 2017-11-21 深圳市科迈爱康科技有限公司 教育机器人
CN107547925A (zh) * 2017-09-27 2018-01-05 刘伟平 一种视频学习监督系统
KR102191488B1 (ko) * 2017-10-27 2020-12-15 서울대학교산학협력단 전력 및 모션 감응형 교육용 로봇
CN109300341A (zh) * 2018-08-30 2019-02-01 合肥虹慧达科技有限公司 交互式早教机器人及其交互方法
CN109147433A (zh) * 2018-10-25 2019-01-04 重庆鲁班机器人技术研究院有限公司 儿童语言辅助教学方法、装置及机器人
CN109366502B (zh) * 2018-12-17 2022-04-08 广东誉丰教育科技有限公司 一种基于人工智能的网络交互式教育方法及机器人
KR20200076169A (ko) * 2018-12-19 2020-06-29 삼성전자주식회사 놀이 컨텐츠를 추천하는 전자 장치 및 그의 동작 방법
KR102708292B1 (ko) * 2018-12-24 2024-09-23 엘지전자 주식회사 로봇 및 그의 제어 방법
KR102134189B1 (ko) * 2019-07-11 2020-07-15 주식회사 아들과딸 인공지능 로봇을 활용한 도서 콘텐츠 제공 방법 및 장치
CN110619767A (zh) * 2019-09-05 2019-12-27 顾柳泉 智能教育机器人及计算机可读介质
DE112019007598B4 (de) * 2019-09-05 2024-05-08 Mitsubishi Electric Corporation Inferenzeinrichtung, vorrichtung-steuerungssystem und lerneinrichtung
CN110751050A (zh) * 2019-09-20 2020-02-04 郑鸿 一种基于ai视觉感知技术的运动教学系统
TWI833681B (zh) * 2023-10-13 2024-02-21 國立勤益科技大學 主動式單字學習系統

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006062274A1 (fr) * 2004-12-07 2006-06-15 Rivalkorea Co., Ltd. Robot intelligent et systeme de jeu mobile dans lequel est utilise ce robot
US20060257830A1 (en) * 2005-05-13 2006-11-16 Chyi-Yeu Lin Spelling robot
US20090104841A1 (en) * 2007-10-19 2009-04-23 Hon Hai Precision Industry Co., Ltd. Toy robot
US20090157223A1 (en) * 2007-12-17 2009-06-18 Electronics And Telecommunications Research Institute Robot chatting system and method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002351305A (ja) * 2001-05-23 2002-12-06 Apollo Seiko Ltd 語学研修用ロボット
CN2702391Y (zh) * 2004-02-04 2005-05-25 上海科技馆 人与机器人做算术游戏的装置
JP3923053B2 (ja) * 2004-03-31 2007-05-30 ファナック株式会社 ロボット教示装置
CN100559422C (zh) * 2006-12-15 2009-11-11 华南理工大学 具备认字、写字功能的教育机器人的文字识别方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006062274A1 (fr) * 2004-12-07 2006-06-15 Rivalkorea Co., Ltd. Robot intelligent et systeme de jeu mobile dans lequel est utilise ce robot
US20060257830A1 (en) * 2005-05-13 2006-11-16 Chyi-Yeu Lin Spelling robot
US20090104841A1 (en) * 2007-10-19 2009-04-23 Hon Hai Precision Industry Co., Ltd. Toy robot
US20090157223A1 (en) * 2007-12-17 2009-06-18 Electronics And Telecommunications Research Institute Robot chatting system and method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of WO2009157733A1 *

Also Published As

Publication number Publication date
CN102077260A (zh) 2011-05-25
CN102077260B (zh) 2014-04-09
KR20100002210A (ko) 2010-01-06
EP2321817A4 (fr) 2013-04-17
WO2009157733A1 (fr) 2009-12-30
KR101088406B1 (ko) 2011-12-01

Similar Documents

Publication Publication Date Title
WO2009157733A1 (fr) Système d’apprentissage interactif utilisant un robot et son procédé de fonctionnement pour l’éducation des enfants
Hyun et al. Comparative study of effects of language instruction program using intelligence robot and multimedia on linguistic ability of young children
US6517351B2 (en) Virtual learning environment for children
Bray et al. Technology and the diverse learner: A guide to classroom practice
JP2020016880A (ja) ダイナミックストーリー指向のデジタル言語教育方法及びシステム
Freed " This is the fluffy robot that only speaks french": language use between preschoolers, their families, and a social robot while sharing virtual toys
de Souza Jeronimo et al. Comparing social robot embodiment for child musical education
Maxwell Beginning reading and deaf children
CN1279502C (zh) 教育设备
Smith The Promise and Threat of Microcomputers for
McKeown Unlocking Potential: How ICT can support children with special needs
Hyun et al. Young children's perception of IrobiQ, the teacher assistive robot, with reference to speech register
WO2013089356A1 (fr) Matériel pédagogique pour apprentissage de l'anglais et support d'enregistrement ayant un matériel pédagogique pour apprentissage de l'anglais enregistré sur celui-ci
Hayati Take 2, Act 1: Feeding two birds with one scone! The role of role-playing in teaching English
Shang A comparative analysis on alternative approaches to literacy instruction
CN112863267B (zh) 一种英语人机对话系统及学习方法
De Wit et al. Designing and evaluating iconic gestures for child-robot second language learning
Kikuchi et al. Developing multimedia training materials for use with small robot controls at Chubu Polytechnic Center in Japan
JP2003208084A (ja) 外国語の学習装置及び学習方法
Tobin et al. How non-visual modalities can help the young visually impaired child to succeed in visual and other tasks
Koceska et al. Design and Development of Educational Game Using ARCS Model
Sanchez et al. Social Robots in Education to Enhance Social Communications and Interaction Skills of Children with Autism-A Review
Qiu et al. Integrating computer-based multimedia instructional design into teaching international English phonetic symbols
CN1912951A (zh) 教育设备
Reagon Using script-fading procedures to teach children with autism to initiate during free play

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20110126

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL BA RS

DAX Request for extension of the european patent (deleted)
A4 Supplementary search report drawn up and despatched

Effective date: 20130314

RIC1 Information provided on ipc code assigned before grant

Ipc: A63H 17/00 20060101ALI20130314BHEP

Ipc: G09B 5/06 20060101AFI20130314BHEP

Ipc: A63H 3/28 20060101ALI20130314BHEP

Ipc: G09B 7/00 20060101ALI20130314BHEP

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20131015