KR20140084431A - Server and method for providing vocabulary icon for foreign language study service and device for representing video - Google Patents

Server and method for providing vocabulary icon for foreign language study service and device for representing video Download PDF

Info

Publication number
KR20140084431A
KR20140084431A KR1020120153246A KR20120153246A KR20140084431A KR 20140084431 A KR20140084431 A KR 20140084431A KR 1020120153246 A KR1020120153246 A KR 1020120153246A KR 20120153246 A KR20120153246 A KR 20120153246A KR 20140084431 A KR20140084431 A KR 20140084431A
Authority
KR
South Korea
Prior art keywords
word
output
moving picture
icon
words
Prior art date
Application number
KR1020120153246A
Other languages
Korean (ko)
Inventor
한영호
김희경
류창선
박성원
박종세
최대우
Original Assignee
주식회사 케이티
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 케이티 filed Critical 주식회사 케이티
Priority to KR1020120153246A priority Critical patent/KR20140084431A/en
Publication of KR20140084431A publication Critical patent/KR20140084431A/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/20Education
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/06Foreign languages
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/06Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
    • G09B5/065Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit

Abstract

A server for providing a word icon comprises a word collection unit to collect words of a predetermined language stored in a mobile device from the mobile device matched with a video playback device; an output word selection unit to select a first output word related to the collected words among one or more output words output during playback of a video; a word icon generation unit to generate a word icon of the selected first word; and a communications unit to provide the video playback device with information on the time when the first output word is output and the generated word icon.

Description

TECHNICAL FIELD [0001] The present invention relates to a server and a method for providing a word icon for foreign language learning,

The present invention relates to a server and a method for providing a word icon, and a moving picture reproducing apparatus for receiving a word icon.

Internet Protocol Television (IPTV) is a two-way broadcasting and communication service that provides a television receiver with various multimedia services such as video contents such as movies and broadcast programs and Internet search using a high-speed Internet network. It differs from over-the-air broadcasting and cable broadcasting in that viewers can select only the programs they want to view at the time they want. In order to use such IPTV technology, only a television, a set-top box, a dedicated modem, and an Internet line need to be connected.

In recent years, technologies that combine IPTV and foreign language learning have been proposed. In connection with a method for providing a foreign language learning method using IPTV, Korean Patent Laid-Open Publication No. 2012-0031671 discloses a configuration for providing a foreign language dictionary function to a video display device.

However, in providing the foreign language learning method using IPTV, the dictionary data for the foreign language subtitles could not be set to words and idioms desired by the user to learn, and repeated learning could not be performed. Also, words and idioms could not be displayed on the display without scripts for foreign language subtitles.

The present invention provides a server and a method for providing a foreign language learning service capable of effectively increasing a memorization power through repeated learning of words and idioms. And to provide a server and a method for providing a learning service that can be memorized for a long time because it learns by utilizing various sensory organs. And to provide a server and a method for providing a foreign language learning service that helps learning effectively for those who have difficulty in taking time by taking a separate time in learning a foreign language. And to provide a server and a method for providing a foreign language learning service for expanding the VOD market to the education market to generate various kinds of revenue. It is to be understood, however, that the technical scope of the present invention is not limited to the above-described technical problems, and other technical problems may exist.

As a technical means to achieve the above technical object, an embodiment of the present invention includes a word collection unit for collecting words of a predetermined language stored in a mobile device from a mobile device matched with a moving picture reproduction device, A word icon generator for generating a word icon of the selected first output word, and information on the time at which the first output word is output, And providing a word icon to the moving picture reproducing device.

In another embodiment of the present invention, there is provided a method for reproducing a moving picture, comprising: collecting words of a predetermined language stored in a mobile device from a mobile device matched with the moving picture reproducing device; Selecting one output word, generating a word icon of the selected first output word, and providing the generated word icon to the moving picture reproducing device with information of the time at which the first output word is output Method can be provided.

According to another embodiment of the present invention, there is provided a communication system including a communication unit for receiving moving picture data of a moving picture from a word icon providing server, a reproducing unit for reproducing a moving picture based on the received moving picture data, An output word recognition unit for recognizing a first output word output through the speaker and a display unit for displaying a word icon corresponding to the first output word recognized based on the received moving picture data on a reproduction screen of the moving picture and displaying the same It is possible to provide a reproducing device.

According to any of the above-mentioned tasks of the present invention, it is possible to effectively increase the memorization power through repetitive learning of words and idioms. It can be remembered for a long time because it learns by utilizing various sensory organs. It is possible to help students to learn effectively when they are having difficulty in studying in a foreign language. By expanding the VOD market into the education market, it is possible to generate various profits.

1 is a configuration diagram of a foreign language learning service providing system according to an embodiment of the present invention.
2 is a configuration diagram of the word icon providing server 10 shown in FIG.
FIG. 3 is a diagram illustrating a process of generating and displaying a word icon from the word icon providing server 10 of FIG.
4 is a flowchart illustrating a process of creating a segment script in the word icon providing server 50 of FIG.
FIG. 5 is a diagram showing an embodiment of a word icon script generated from the word icon providing server 10 of FIG.
6 is a diagram showing an embodiment of word selection information set in the word application of the mobile device 20 of FIG.
7 is a configuration diagram of the moving picture reproducing device 30 shown in FIG.
8A and 8B are views showing an example in which an icon is displayed on the word display of the moving picture reproducing device 30 according to the embodiment of the present invention.
9 is a flowchart illustrating a process of providing a word icon to a moving image playback device according to an embodiment of the present invention.
10 is a flowchart illustrating a process of reproducing a moving picture based on moving picture data received from a word icon providing server according to an embodiment of the present invention.

Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings, which will be readily apparent to those skilled in the art. The present invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. In order to clearly illustrate the present invention, parts not related to the description are omitted, and similar parts are denoted by like reference characters throughout the specification.

Throughout the specification, when a part is referred to as being "connected" to another part, it includes not only "directly connected" but also "electrically connected" with another part in between . Also, when an element is referred to as "comprising ", it means that it can include other elements as well, without departing from the other elements unless specifically stated otherwise.

1 is a configuration diagram of a foreign language learning service providing system according to an embodiment of the present invention. 1, a foreign language learning service providing system 1 includes a word icon providing server 10, a mobile device 20, a moving picture reproducing device 30, a cloud server 40 and a script server 50 .

1 constituting the foreign language learning service providing system 1 can be connected to each other through the network 15. 1, the word icon providing server 10, the cloud server 40 and the script server 50 are connected to the mobile device 20 and the moving picture reproducing device 30 via the network 15. [ Lt; / RTI > Such a network means a connection structure in which information can be exchanged between each node such as terminals and servers. As an example of a network, there are an Internet, a LAN (Local Area Network), a Wireless LAN Local Area Network (WAN), Wide Area Network (WAN), Personal Area Network (PAN), and the like. It should be noted that the word icon providing server 10, the cloud server 40, the script server 50, the mobile device 20 and the moving picture reproducing device 30 disclosed in Fig. 1 are limited to those shown in Fig. 1 no.

The word icon providing server 10 relays data among other components in the foreign language learning service providing system 1. [ For example, the word icon providing server 10 can receive the foreign language words and idioms stored in the word application of the mobile device 20 from the cloud server 40. For example, the word icon providing server 10 may receive the text of the output word while the moving image is being reproduced from the script server 50. In this case, the foreign language means languages other than the native language such as English, French, Spanish, and Italian.

The mobile device 20 transmits the words of the predetermined language to the cloud server 40. [ Words in a given language may be stored through the word application of the mobile device 20, and an example of a word in a given language may be English. At this time, the word application may include at least one of a word list, an idiom list, an input time, a source, a repetition count, an exposed presence user setting, and a recently viewed VOD information. For example, words in a predetermined language are sequentially displayed in a word list, and the input time indicates a time when the word is input. When the source is directly input by the user, it is displayed as a user, and when a word of a predetermined language is extracted from the word list, the name of the word list is displayed. The number of repetition times indicates the number of times a word of a predetermined language is exposed on the display of the moving picture reproducing device 30. If the number of repetitions exceeds the number of repetition times, the word is no longer displayed on the moving picture reproducing device 30. The user setting of presence / absence of exposure is to set whether or not the user wants to expose words and idioms to the video playback apparatus. If the recently viewed VOD information includes the word and the idiom in the recently viewed VOD, the related VOD name and the VOD turn can be displayed.

The moving picture reproducing device 30 receives the moving picture data from the word icon providing server 10. At this time, the moving picture data may include the original moving picture data of the moving picture, the information at the time of outputting the first output word, and the word icon of the first output word. One example of the moving picture reproducing device 30 may be any one of IPTV (Internet Protocol Television), Smart TV (Smart TV), and Connected TV.

The cloud server 40 registers words of a predetermined language stored in the word application of the mobile device 20 from the mobile device 20. [ In addition, the cloud server 40 transmits the registered word of the predetermined language to the word icon providing server 10. At this time, the cloud server 40 is connected to the mobile device 20 and the word icon providing server 10 via the network 15. An example of the network 15 is the Internet, a LAN (Local Area Network) ), Wireless LAN (Local Area Network), WAN (Wide Area Network), PAN (Personal Area Network), 3G, 4G, LTE, Wi-Fi and the like.

The script server 50 includes text of one or more output words that are output while a moving image is played back. At this time, the text of the output word may be a subtitle script. In addition, the script server 50 may send one or more output word texts to the word icon providing server 10.

The word icon providing server 10 generates a segment script based on the script or sound received from the script server 50. [

The word icon providing server 10 generates a word icon by matching a word of a predetermined language received from the cloud server 40 with a segment script. At this time, the word icon providing server 10 transmits the moving picture data including the generated word icon and the point of view information to the moving picture reproducing device 30.

When the moving picture is reproduced, the moving picture reproducing device 30 can display the word icon in accordance with the time information included in the segment script. At this time, the word icon can be displayed by either a dynamic point or a fixed point by the user. When a moving picture is being played back on the moving picture reproducing device 30, if the word icon is set as a dynamic point, the word can be exposed around the mouth of the speaking person who uttered the word. In addition, if the word icon is set as a fixed point through the moving picture reproducing device 30, the word icon can be fixed to a specific position on the display of the moving picture reproducing device 30 and exposed.

In addition, when a plurality of word icons are exposed on the display of the moving picture reproducing device 30, any word icon can be selected through the user interface. The display of the video playback device 30 may display a plurality of word icons and definitions of word icons selected from the user interface.

2 is a configuration diagram of the word icon providing server 10 shown in FIG. 2, the word icon providing server 10 may include a word collecting unit 110, an output word selecting unit 120, a word icon generating unit 130, a communication unit 140, and a database 150 have.

The word collection unit 110 collects words of a predetermined language stored in the mobile device 20 from the mobile device 20 matched with the moving picture reproduction device 30. [ At this time, the word collection unit 110 collects words of a predetermined language registered in the cloud server 40 through the word application of the mobile device 20. [ For example, the word collection unit 110 may collect words of a predetermined language stored in the word application of the mobile device 20, and the word collection unit 110 may collect the word of the mobile device 20 through the word application of the mobile device 20, The words of the predetermined language registered in the server 40 can be collected from the cloud server 40. [

The output word collection unit 120 selects a first output word associated with the collected word among one or more output words output during the reproduction of the moving image. The output word collection unit 120 can receive script data based on moving picture data from the script server 50. [ The output word collection unit 120 can determine whether or not script data of the moving picture data received from the script server 50 exists. The output word collection unit 120 may generate the segment script through word alignment when the script of the moving image data exists, and when the script data of the moving image data does not exist, Based on the acoustic data of the speech recognition system.

At this time, the speech recognition system may include a HMM algorithm (Hidden Markov Model), a Viterbi Search and State Alignment, and a Noise Cancelation End Point Detect. The HMM algorithm (Hidden Markov Model) is the most successful model in the field of speech recognition and is mainly used for pattern recognition (pattern recognition) or character recognition. The HMM algorithm (Hidden Markov Model) assumes that the speech signal is generated by the Markov model, and estimates the parameter of the model in the learning stage. The recognizer uses the estimated parameters to find the most suitable model for the unknown input speech. An example of the model used for recognition may be a phoneme or a word itself. The Viterbi Search and State Alignment can be used to find the sequence of words as an error correction algorithm. Finally, the Noise Cancellation End Point Detection method is a method of detecting the presence of voice in a low SNR environment and detecting an accurate end point.

The output word selection unit 120 selects a first output word associated with a word collected in the word collection unit 110 among one or more output words output during the reproduction of the moving image. For example, the output word selection unit 120 can select the first output word based on the word selection information input from the mobile device 20 or the moving picture reproduction device 30. [ The word selection information includes at least one of a word list, an input time, a source, a repetition count, an exposure setting user setting, and a recent viewing VOD information.

When the first output word is selected by the output word selection unit 120, one or more output words may be output through the speaker of the moving picture reproduction device 30 when the speaker included in the moving image utters the output word.

The word icon generation unit 130 generates a word icon of the first output word selected by the output word selection unit 120. At this time, the word of the predetermined language may be an English word.

The communication unit 140 may transmit a predetermined word registered in the cloud server 40 to the word collection unit 110. [ In this case, the word of the predetermined language registered in the cloud server 40 may be a word stored through the word application of the mobile device 20. [

The communication unit 140 may transmit the moving picture data including the information at the time when the first output word is output from the output word selection unit 120 and the word icon generated by the word icon generation unit 130, To the moving picture reproducing device (30). For example, the communication unit 140 may transmit information of a time when a foreign language word is output and an icon of a foreign language word to the moving picture reproducing device 30, and the moving picture reproducing device 30 displays an icon of the transmitted foreign word Can be displayed. The database 150 may include moving picture data to be played back by the moving picture reproducing device 30. [ As an example, the database 150 may store VOD movie data.

FIG. 3 is a diagram illustrating a process of generating and displaying a word icon from the word icon providing server 10 of FIG.

The word collection unit 110 collects words of a predetermined language registered in the cloud server 40 from the mobile device 20. At this time, the word of the predetermined language may be a word stored through the word application of the mobile device 20. The word in a certain language may be an English word poor.

The output word selection unit 120 may select a first output word associated with a word collected from the word collection unit 110 among at least one output word output during reproduction of the moving image. The output word selection unit 120 may further include an output word identification unit 121 and a selection unit 122. The output word identification unit 121 can identify the text of one or more output words output during the reproduction of the moving image based on the moving image data included in the database 150. [ For example, the output word identification unit 121 can identify the text of the English word " poor "

The output word identification unit 121 may generate a segment script including the point-in-time information at which one or more output words are output, and the output word identification unit 121 may identify one or more output words based on the generated segment script can do.

At this time, the viewpoint information includes a text of the first output word, a time point at which the first output word starts, a point at which the first output word ends, a method in which the word icon of the first output word is displayed, May be included in at least one of the displayed positions.

The output word identification unit 121 can generate the segment script based on the script data or the sound data included in the moving picture data of the database 150. [ The output word identification unit 121 may determine whether or not script data of moving picture data exists. At this time, if the script data exists in the moving picture data, the output word identification unit 121 can generate a segment script through word alignment. The output word identification unit 121 may generate a segment script through voice recognition when script data does not exist in the moving picture data.

In this case, the speech recognition may include an HMM algorithm (Hidden Markov Model), a Viterbi Search and State Alignment, and a Noise Cancelation End Point Detect. The HMM algorithm (Hidden Markov Model) is a model of the speech recognition field and is mainly used in pattern recognition or character recognition. The HMM algorithm (Hidden Markov Model) assumes that the speech signal is generated by the Markov model, and estimates the parameter of the model in the learning stage. The recognizer uses the estimated parameters to find the best model for the unknown input speech. An example of the model used for recognition may be the phoneme or the word itself. The Viterbi Search and State Alignment is a search for a sequence of words as an error correction algorithm. Finally, the Noise Cancellation End Point Detection method is a method of detecting the presence of voice in a low SNR environment and detecting an accurate end point.

For example, the output word identification unit 121 can extract a text of a plurality of output words "He is not only poor, but also lazy" through the script data included in the moving picture data. At this time, the output word identification unit 121 can generate a segment script including the time information of output words of He, is, not, only, poor, but, also, and lazy. The output word identification unit 121 can identify one or more output words based on the generated segment script and transmit the identified output word to the selection unit 122. [

The selecting unit 122 may compare each of the one or more output words transmitted from the output word identifying unit 121 with the words collected from the word collecting unit 110. At this time, the selecting unit 122 can select a first output word associated with the word collected from the word collecting unit 110 among the one or more output words transmitted from the output word identifying unit 121. [ At this time, the associated first output word may be the same or derived word from the collected word. For example, when the words He, is, not, only, poor, but, also, and lazy are transmitted from the output word identification unit 121 and words poor, smart, and recognition are collected from the word collection unit 110 The selecting unit 122 may select poor, which is the same word as the collected word, among the one or more output words transmitted. In another example, when the word It, is, recognized is transmitted from the output word identification unit 121 and the word recognize is collected from the word collection unit 110, the selection unit 122 selects one or more You can choose recognized word, which is a word derived from collected words. In this case, examples of words derived from recognize are various, such as recognized, recognizing, and recognition.

The word icon generation unit 130 may generate a word icon of the first output word selected from the selection unit 122. [ For example, the word icon generation unit 130 may generate an icon of the word " poor ", which is the first output word. The word icon generation unit 130 may transmit the generated word icon to the communication unit 140 by including the generated word icon in the moving image data.

The communication unit 140 may transmit the moving picture data including the word icon to the moving picture reproducing device 30. [ In this case, the moving picture data may include information on a time point at which the first output word is output, a generated word icon, and learning data associated with the first output word. For example, the communication unit 140 may transmit the moving picture data including the word icon of the word poor and the information of the point of time when the poor is output to the moving picture reproducing device 30. The moving picture reproducing device 30 displays (31) a word icon on the display based on the information at the time when the first output word transmitted from the communication unit 140 is outputted.

In another embodiment of the present invention, the output word identifier 121 may exist in a separate segment server. The words collected through the word icon providing server 10 are transmitted to the segment server, and the segment server can identify one or more output words output during reproduction of the moving image and generate the segment script.

4 is a flowchart illustrating a process of creating a segment script in the word icon providing server 50 of FIG. Referring to FIG. 4, the shortcut icon transmission server 10 extracts script data or sound data from the script server 50 (S410) and determines whether there is script data (S420). If the script data exists, the segment subtitle script is generated (S430) through word alignment (S421). If the script data does not exist (S420), speech recognition is used (S422) to generate a segment script (S430).

FIG. 5 is a diagram showing an embodiment of a word icon script generated from the word icon providing server 10 of FIG. 5, the word icon script may include a plurality of word lists 510, a start point 520, an end point 530, an exposure scheme 540, and an exposure position 550.

The word list 510 indicates a word of a predetermined language registered in the cloud server 40. At this time, the words of the predetermined language can be stored in the word application of the mobile device 20. [ The word list 510 may indicate the order in which they are exposed to moving picture data, and may be arranged in the vertical direction.

The start point 520 and the end point 530 represent information at the time when the first output word identified based on the moving image data is outputted. For example, the word lazy starts at 10:21:34 and ends at 10:22:34.

The exposure scheme 540 indicates how the word icon of the first output word is displayed. The word icon may be displayed by a user as either a dynamic point or a fixed point. When a moving picture is being played back on the moving picture reproducing device 30, if the word icon is set as a dynamic point, the word can be exposed around the mouth of the speaking person who uttered the word. In addition, if the word icon is set as a fixed point, the word icon can be fixed to a specific position on the display of the moving picture reproducing device 30 and exposed.

The exposure position 550 indicates a position at which the word icon of the first output word is displayed, and may be displayed in two-dimensional coordinates expressed by XY.

6 is a diagram showing an embodiment of word selection information set in the word application of the mobile device 20 of FIG. Referring to FIG. 6, the word selection information may include a word list 610, an input time 620, a source 630, a repetition number 640, an exposure availability user setting 650, and recent viewing VOD information.

The word list 610 represents the order stored in the word application of the mobile device 20 and may be arranged in the longitudinal direction.

The input time 620 represents the time in which a word of a certain language is stored in the word application of the mobile device 20. For example, if the word lazy is stored on November 10, 2012, it might look like 2012.11.10.

The source 630 represents a subject that stores a word of a predetermined language stored in the word application of the mobile device 20, and may include a user or a word list collection. For example, when the user directly stores the word in the word application of the mobile device 20, the word is displayed as a user, and if the word is stored in the word list, the word list is displayed.

The number of repetition times 640 displays the number of exposures of the word icon corresponding to a word of a predetermined language on the display of the moving picture reproducing apparatus 30. [ At this time, the number of repetition times 640 is determined by the user, and when the number of repetition times 640 is exceeded, the moving picture reproducing apparatus 30 does not display the word icon on the display.

The exposure availability user setting 650 is to set whether or not the word icon corresponding to a word of a predetermined language is exposed on the display of the moving picture reproducing device 30.

The latest viewing VOD information 660 displays viewing information of a moving picture in which a word icon corresponding to a word of a predetermined language is exposed. At this time, the latest viewing VOD information 660 may include the title and the rotation of the moving image.

7 is a configuration diagram of the moving picture reproducing device 30 shown in FIG. 7, the moving picture reproducing device 30 may include a communication unit 310, a reproducing unit 320, an output word recognizing unit 330, a display unit 340, and a database 350.

The communication unit 310 receives moving image moving image data from the word icon providing server 10. In this case, the moving picture data may include original moving picture data of the moving picture, jumbo at the time of outputting the first output word, and a word icon of the first output word. The reproduction unit 320 reproduces the moving picture based on the moving picture data received from the word icon providing server 10. [ For example, the playback unit may play back the VOD movie received from the word icon providing server 10. [

The output word recognizing unit 330 can recognize the first output word output through the speaker during the reproduction of the moving image based on the moving image data received from the word icon providing server 10. [ For example, when one speaker in the moving image utters the word poor and the word poor is output through the speaker, the output word recognizing unit 330 can recognize the word poor, which is uttered by the speaker.

The display unit 340 can overlay the word icon corresponding to the first output word recognized by the output word recognizing unit 330 on the moving picture playback screen based on the moving picture data received from the word icon providing server 10 have. At this time, the position of the word icon is determined by the user and can be displayed as either a dynamic point or a fixed point. When a moving picture is being played back on the moving picture reproducing device 30, if the word icon is set as a dynamic point, the word can be exposed around the mouth of the speaking person who uttered the word. In addition, if the word icon is set as a fixed point, the word icon can be fixed to a specific position on the display of the moving picture reproducing device 30 and exposed.

8A and 8B are views showing an example in which an icon is displayed on the word display of the moving picture reproducing device 30 according to the embodiment of the present invention. Referring to FIG. 8A, a word icon of " poor " is displayed in the form of a dynamic point through time information included in a segment script. In addition, when a word icon included in the moving picture data is exposed to the moving picture reproducing device 30, the exposed word icon may be displayed on the display of the moving picture reproducing device 30 and displayed as a plurality of word icons. 8B, when a plurality of word icons including numbers are displayed on the display of the moving picture reproducing device 30 and the moving picture reproducing device 30 receives any one of the numbers from the user interface, the moving picture reproducing device 30 May display a definition of a word icon corresponding to the number entered in the display based on the entered number.

9 is a flowchart illustrating a process of providing a word icon to the moving picture reproducing device 30 according to an embodiment of the present invention.

Referring to FIG. 9, the word icon providing server 10 collects words of a predetermined language stored in the mobile device 20 from the mobile device 20 matched with the moving picture reproducing device 30 (S910). Then, the word icon providing server 10 selects a first output word that is the same as the collected word among the one or more output words output during playback of the moving image (S920), and generates a word icon of the selected first output word )do. Thereafter, the word icon providing server 10 transmits the information of the time at which the first output word is output and the generated word icon to the moving picture reproducing device 30 (S940). 10 is a flowchart illustrating a process of reproducing a moving image based on moving image data received from a word icon providing server 10 according to an embodiment of the present invention.

10, the moving picture reproducing device 30 receives the moving picture data of the moving picture from the word icon providing server 10 (S1010), and reproduces the moving picture based on the received moving picture data (S1020). Then, the moving picture reproducing device 30 recognizes a first output word outputted through the speaker during the reproduction of the moving picture based on the received moving picture data (S1030), and recognizes the first output word recognized based on the received moving picture data (S1040). ≪ / RTI >

 The word icon providing method and the moving picture reproducing device reproducing method according to the embodiments described with reference to FIGS. 9 to 10 may also be implemented in the form of a recording medium including a command executable by a computer such as a program module executed by a computer . Computer readable media can be any available media that can be accessed by a computer and includes both volatile and nonvolatile media, removable and non-removable media. In addition, the computer-readable medium may include both computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Communication media typically includes any information delivery media, including computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave, or other transport mechanism.

It will be understood by those skilled in the art that the foregoing description of the present invention is for illustrative purposes only and that those of ordinary skill in the art can readily understand that various changes and modifications may be made without departing from the spirit or essential characteristics of the present invention. will be. It is therefore to be understood that the above-described embodiments are illustrative in all aspects and not restrictive. For example, each component described as a single entity may be distributed and implemented, and components described as being distributed may also be implemented in a combined form.

The scope of the present invention is defined by the appended claims rather than the detailed description and all changes or modifications derived from the meaning and scope of the claims and their equivalents are to be construed as being included within the scope of the present invention do.

10: Word icon providing server
20: Mobile device
30: Video playback device
40: Cloud server
50: Script Server
110: word collection unit
120: output word selection unit
121: output word identification unit
122:
130: word icon generation unit
140:

Claims (16)

A server for providing a word icon of a word outputted at the time of video playback to a video playback device,
A word collection unit for collecting words of a predetermined language stored in the mobile device from a mobile device matched with the moving picture reproduction device;
An output word selector for selecting a first output word associated with the collected word from among one or more output words output during playback of the moving image;
A word icon generator for generating a word icon of the selected first output word; And
And a communication unit for providing information on a time point at which the first output word is output and the generated word icon to the moving picture reproducing device.
The method according to claim 1,
Wherein the word collection unit collects words of the predetermined language through a word application installed in the mobile device.
The method according to claim 1,
The word collecting unit collects words of the predetermined language through a cloud server,
Wherein the cloud server registers words of a predetermined language stored in the mobile device from the mobile device.
The method according to claim 1,
Wherein the output word selector comprises:
An output word identification unit for identifying one or more output words to be output while the moving image is reproduced based on moving image moving image data; And
And a selector for selecting a first output word associated with the collected one of the identified one or more output words by comparing the collected words with each of the identified one or more output words.
5. The method of claim 4,
Wherein the output word identification unit generates a segment script including a text of one or more output words output while the moving image is reproduced and a time point at which each of the one or more output words is output based on the moving image data, And identifying one or more output words based on the script.
6. The method of claim 5,
Wherein the output word identification unit generates the segment script through word alignment of script data included in the moving picture data.
6. The method of claim 5,
Wherein the output word identification unit generates the segment script through speech recognition of the sound data included in the moving picture data.
5. The method of claim 4,
Wherein the output word identification unit receives from the segment server a segment script including a text of one or more output words output during playback of the moving image from the script server and a time point at which each of the one or more output words is output, And identifying one or more output words based on the segment script.
The method according to claim 1,
Wherein the word of the predetermined language is an English word.
The method according to claim 1,
Wherein the communication unit includes information on a time point at which the first output word is outputted to the moving picture data of the moving picture and the generated word icon, and provides the word icon to the moving picture reproduction device.
11. The method of claim 10,
The information of the time point at which the first output word is output may include at least one of a text of the first output word, a time at which the first output word starts, a time at which the first output word ends, And a position at which a word icon of the first output word is displayed,
Wherein the moving picture data includes information on a time at which the first output word is output, the generated word icon, and learning data associated with the first output word.
The method according to claim 1,
Wherein the output word selection unit selects the first output word based on the word selection information input from the mobile device or the moving picture reproduction device.
13. The method of claim 12,
Wherein the one or more output words are output through a speaker of the moving picture reproducing device when a speaker included in the moving picture is fired.
A method of providing a word icon of a word to be output at the time of moving picture reproduction to a moving picture reproducing device,
Collecting words of a predetermined language stored in the mobile device from a mobile device matched with the moving picture reproducing device;
Selecting a first output word associated with the collected word from one or more output words output during playback of the moving image;
Generating a word icon of the selected first output word; And
Providing information on a time point at which the first output word is output and the generated word icon to the moving picture reproducing device.
1. A moving picture reproducing device for reproducing a moving picture based on moving picture data received from a word icon providing server,
A communication unit for receiving the moving picture data of the moving picture from the word icon providing server;
A reproducing unit for reproducing the moving image based on the received moving image data;
An output word recognizer that recognizes a first output word output through the speaker during the reproduction of the moving image based on the received moving image data; And
And a display unit for overlaying a word icon corresponding to the recognized first output word on the reproduction screen of the moving picture based on the received moving picture data.
16. The method of claim 15,
Wherein the moving picture data includes original moving picture data of the moving picture, information at a time when the first output word is outputted, and a word icon of the first output word.
KR1020120153246A 2012-12-26 2012-12-26 Server and method for providing vocabulary icon for foreign language study service and device for representing video KR20140084431A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020120153246A KR20140084431A (en) 2012-12-26 2012-12-26 Server and method for providing vocabulary icon for foreign language study service and device for representing video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
KR1020120153246A KR20140084431A (en) 2012-12-26 2012-12-26 Server and method for providing vocabulary icon for foreign language study service and device for representing video

Publications (1)

Publication Number Publication Date
KR20140084431A true KR20140084431A (en) 2014-07-07

Family

ID=51734303

Family Applications (1)

Application Number Title Priority Date Filing Date
KR1020120153246A KR20140084431A (en) 2012-12-26 2012-12-26 Server and method for providing vocabulary icon for foreign language study service and device for representing video

Country Status (1)

Country Link
KR (1) KR20140084431A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102619562B1 (en) * 2023-03-27 2023-12-29 주식회사 트리거스 System for providing foreign language learning reward service using lock screen

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102619562B1 (en) * 2023-03-27 2023-12-29 주식회사 트리거스 System for providing foreign language learning reward service using lock screen

Similar Documents

Publication Publication Date Title
KR102428911B1 (en) Method and system for correcting input generated using automatic speech recognition based on speech
DK3175442T3 (en) SYSTEMS AND METHODS FOR PERFORMING ASR IN THE PRESENCE OF HETEROGRAPHS
US11227620B2 (en) Information processing apparatus and information processing method
CN108292314B (en) Information processing apparatus, information processing method, and program
CN108012173A (en) A kind of content identification method, device, equipment and computer-readable storage medium
CN101465068A (en) Method for the determination of supplementary content in an electronic device
US11800202B2 (en) Systems and methods for generating supplemental content for a program content stream
JP6202815B2 (en) Character recognition device, character recognition method, and character recognition program
JP6654718B2 (en) Speech recognition device, speech recognition method, and speech recognition program
KR102145370B1 (en) Media play device and method for controlling screen and server for analyzing screen
KR101962126B1 (en) Multimedia device for accessing database according to result of voice recognition and method for controlling the same
JP2006337490A (en) Content distribution system
KR102396263B1 (en) A System for Smart Language Learning Services using Scripts
KR20140084431A (en) Server and method for providing vocabulary icon for foreign language study service and device for representing video
JP4654438B2 (en) Educational content generation device
KR20140084463A (en) Apparatus and method for displaying image of narrator information and, server for editing video data
US20230196934A1 (en) Apparatus and method for supporting language learning using video
CN109977239B (en) Information processing method and electronic equipment
KR20110007384A (en) System for studying a foreign language to extract video data and method thereof
JP2013088738A (en) Device, method and program for display control of sign-language interpreter video
JP2023049066A (en) Language education animation system
JP2017060059A (en) Control program, storage medium, portable communication equipment, program-related information provision device, and program-related information display method
KR20140099630A (en) Method and Apparatus for speaking learning

Legal Events

Date Code Title Description
WITN Withdrawal due to no request for examination