US20160180741A1 - Pronunciation learning device, pronunciation learning method and recording medium storing control program for pronunciation learning - Google Patents
Pronunciation learning device, pronunciation learning method and recording medium storing control program for pronunciation learning Download PDFInfo
- Publication number
- US20160180741A1 US20160180741A1 US14/841,565 US201514841565A US2016180741A1 US 20160180741 A1 US20160180741 A1 US 20160180741A1 US 201514841565 A US201514841565 A US 201514841565A US 2016180741 A1 US2016180741 A1 US 2016180741A1
- Authority
- US
- United States
- Prior art keywords
- pronunciation
- example sentence
- word
- data
- output
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/06—Foreign languages
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
Definitions
- the present invention relates to a pronunciation learning device and a control program thereof. More particularly, the present invention relates to a pronunciation learning device which has a function which allows a user to efficiently learn pronunciations, and a control program thereof.
- Recent electronic dictionaries have, for example, a function of outputting a pronunciation of a specific example sentence as disclosed in JP 2013-37251 A or a function of searching for a word in an example sentence, vocally outputting the example sentence and displaying a translation for learning a pronunciation of words as disclosed in JP 2006-268501 A. Therefore, some electronic dictionaries are used not only for dictionaries but also for pronunciation learning devices.
- the conventional pronunciation learning device can not only output pronunciations of words but also output pronunciations of example sentences including the words.
- such a conventional pronunciation learning device cannot associate and provide pronunciations of words and a pronunciation of an example sentence including these words.
- a conventional pronunciation learning device does not have a function of associating pronunciations of words and a pronunciation of an example sentence including these words. Therefore, since there is a problem that it is bothersome for a user to operate the pronunciation learning device to output a pronunciation per individual word included in the example sentence, the user cannot efficiently learn the pronunciations.
- the present invention has been made in light of such a situation. It is therefore an object of the present invention to provide a pronunciation learning device which can associate and provide pronunciations of words and a pronunciation of an example sentence including these words and, consequently, provide a function of enabling a user to efficiently learn pronunciations.
- a pronunciation learning device includes: an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words; an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentence text storage unit with pronunciation data as a pronunciation-associated example sentence; a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation; a pronunciation-associated example sentence registering unit configured to extract from the example sentence pronunciation storage unit a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence; and an example sentence pronunciation output unit configured to read from the example sentence pronunciation storage unit pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering unit, and to vocally output the read pronunciation data.
- the present invention can realize a pronunciation learning device which can associate and provide pronunciations of words and a pronunciation of an example sentence including these words, and which allows a user to efficiently learn pronunciations.
- FIG. 1 is a block diagram showing a configuration of an electronic circuit of a pronunciation learning device according to a first embodiment of the present invention
- FIG. 2 is a front view showing an external appearance configuration in case where the pronunciation learning device according to the first embodiment of the present invention is realized as an electronic dictionary device;
- FIG. 3 is a front view showing an external appearance configuration in case where the pronunciation learning device according to the first embodiment of the present invention is realized as a tablet terminal;
- FIG. 4 is a flowchart showing processing (part 1 ) of the pronunciation learning device according to the first embodiment of the present invention
- FIG. 5 is a flowchart showing processing (part 2 ) of the pronunciation learning device according to the first embodiment of the present invention
- FIG. 6 is a flowchart showing processing (part 3 ) of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 7 is a flowchart showing processing (part 4 ) of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 8 is a flowchart showing processing (part 5 ) of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 9 is a view showing an example (part 1 ) of a display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 10 is a view showing an example (part 2 ) of the display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 11 is a view showing an example (part 3 ) of the display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 12 is a view showing an example (part 4 ) of the display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 13 is a view showing an example (part 5 ) of the display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 14 is a view showing an example (part 6 ) of the display screen of the pronunciation learning device according to the first embodiment of the present invention.
- FIG. 15 is a conceptual diagram showing a configuration example of a pronunciation learning device according to a second embodiment of the present invention.
- a pronunciation learning device according to a first embodiment of the present invention will be described.
- FIG. 1 is a block diagram showing an electronic circuit of a pronunciation learning device 10 according to the first embodiment of the present invention.
- This pronunciation learning device 10 is a device which is preferable to learn pronunciations of foreign languages in particular, and includes a CPU 11 , a memory 12 , a recording medium reading unit 14 , a key input unit 15 , a main screen 16 , a sub screen 17 , a speaker 18 a and a microphone 18 b which are connected with each other through a communication bus 19 .
- the CPU 11 controls an operation of the pronunciation learning device 10 according to a pronunciation learning processing control program 12 a stored in advance in the memory 12 , the pronunciation learning processing control program 12 a read to the memory 12 from an external recording medium 13 such as a ROM card through the recording medium reading unit 14 , or the pronunciation learning processing control program 12 a downloaded from a web server (a program server in this case) on the Internet and read to the memory 12 .
- a pronunciation learning processing control program 12 a stored in advance in the memory 12
- the pronunciation learning processing control program 12 a read to the memory 12 from an external recording medium 13 such as a ROM card through the recording medium reading unit 14
- the pronunciation learning processing control program 12 a downloaded from a web server (a program server in this case) on the Internet and read to the memory 12 .
- the pronunciation learning processing control program 12 a also includes a communication program for performing data communication with each web server on the Internet or a user PC (Personal Computer) externally connected to the pronunciation learning device 10 . Further, the pronunciation learning processing control program 12 a is activated according to an input signal corresponding to a user's operation from the key input unit 15 , an input signal corresponding to a user's operation from the main screen 16 or the sub screen 17 having a touch panel color display function, a communication signal for the web server on the externally connected Internet or a connection communication signal for the recording medium 13 such as an EEPROM, a RAM or a ROM externally connected through the recording medium reading unit 14 .
- the memory 12 includes a dictionary database 12 b , an example sentence text storage area 12 c , an example sentence pronunciation storage area 12 d , a word registration area 12 e , a pronunciation-associated example sentence registration area 12 f , and a user pronunciation registration area 12 g.
- dictionary database 12 b dictionaries (an English-Japanese dictionary, a Japanese-English dictionary, an English-English dictionary, a Chinese-Japanese dictionary, a Japanese-Chinese dictionary, Chinese phrases and a Chinese-Chinese dictionary) and phrases of learning target foreign languages such as English or Chinese are stored.
- a dictionary stores pieces of general dictionary information such as parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, word pronunciation data, meanings, a usage, example sentences, and an example sentence pronunciation data per word.
- phrases information such as example sentences, meanings and pronunciations are stored per situation such as a travel, business, life and cooking.
- the numbers of dictionaries and phrases are not limited to singular forms.
- the dictionary database 12 b may store a plurality of dictionaries of similar types like a plurality of English-Japanese dictionaries.
- the example sentence text storage area 12 c is a storage area in which example sentence texts are extracted from the dictionaries and the phrases stored in the dictionary database 12 b under control of the pronunciation learning processing control program 12 a , and are stored together with names of sources (e.g. English-Japanese Dictionary A).
- the example sentence text storage area 12 c configures an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words.
- the example sentence pronunciation storage area 12 d is a storage area in which each example sentence text stored in the example sentence text storage area 12 c is associated with the pronunciation data and stored as a pronunciation-associated example sentence under control of the pronunciation learning processing control program 12 a .
- the example sentence pronunciation storage area 12 d configures an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentence text storage area 12 c with pronunciation data as a pronunciation-associated example sentence.
- the word registration area 12 e is a registration area in which a word of pronunciation data vocally output from the speaker 18 a is registered under control of the pronunciation learning processing control program 12 a .
- the words registered in advance are basic words for which the user does not need to practice, and correspond to, for example, “I”, “to” and “the” in case of English.
- the word registration area 12 e configures a word registering unit configured to register the word corresponding to the pronunciation data vocally output by the speaker 18 a .
- the speaker 18 a configures a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation.
- the pronunciation-associated example sentence registration area 12 f is a registration area in which a pronunciation-associated example sentence including the pronunciation data of the word registered in the word registration area 12 e is extracted from the example sentence pronunciation storage area 12 d and the extracted pronunciation-associated example sentence is registered under control of the pronunciation learning processing control program 12 a .
- One of items of pronunciation data of pronunciation-associated example sentences registered in the pronunciation-associated example sentence registration area 12 f is read from the example sentence pronunciation storage area 12 d and is vocally output from the speaker 18 a under control of the pronunciation learning processing control program 12 a performed by a user's operation.
- the pronunciation-associated example sentence registration area 12 f configures a pronunciation-associated example sentence registering unit configured to extract from the example sentence pronunciation storage area 12 d a pronunciation-associated example sentence including the pronunciation data of the word registered in the word registering area 12 e , and to register the extracted pronunciation-associated example sentence.
- the speaker 18 a also configures an example sentence pronunciation output unit configured to read from the example sentence pronunciation storage area 12 d pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering area 12 f , and to vocally output the read pronunciation data.
- the user pronunciation registration area 12 g is a storage area in which pronunciation data obtained by the microphone 18 b and pronounced by the user is stored.
- the example sentence text storage area 12 c , the example sentence pronunciation storage area 12 d , the word registration area 12 e and the pronunciation-associated example sentence registration area 12 f are preferably provided per language.
- the pronunciation learning processing control program 12 a which is applied to the pronunciation learning device 10 according to the first embodiment of the present invention controls an operation performed by a conventional electronic dictionary or pronunciation learning device, and, the following operation by using these dictionary database 12 b , example sentence text storage area 12 c , the example sentence pronunciation storage area 12 d , the word registration area 12 e and the pronunciation-associated example sentence registration area 12 f.
- the pronunciation learning processing control program 12 a causes a new example sentence text to be stored in the example sentence text storage area 12 c.
- the pronunciation learning processing control program 12 a causes each example sentence text stored in the example sentence text storage area 12 c to be associated as a pronunciation-associated example sentence with pronunciation data and stored in the example sentence pronunciation storage area 12 d.
- the pronunciation learning processing control program 12 a causes the speaker 18 a to vocally output pronunciation data of a word specified by a user's operation by using the key input unit 15 .
- the word corresponding to pronunciation data vocally output from the speaker 18 a is registered in the word registration area 12 e.
- the pronunciation learning processing control program 12 a causes a pronunciation-associated example sentence including the pronunciation data of the word registered in the word registration area 12 e to be extracted from the example sentence pronunciation storage area 12 d , and causes the extracted pronunciation-associated example sentence to be registered in the pronunciation-associated example sentence registration area 12 f.
- the pronunciation learning processing control program 12 a causes pronunciation data specified by a user's operation among pronunciation-associated example sentences registered in the pronunciation-associated example sentence registration area 12 f , to be read from the example sentence pronunciation storage area 12 d , and causes the speaker 18 a to vocally output the pronunciation data.
- the word specified by the user's operation using the key input unit 15 and corresponding to the pronunciation data vocally output from the speaker 18 a is registered in the word registration area 12 e.
- the pronunciation learning processing control program 12 a In response to registration of the word in the word registration area 12 e , the pronunciation learning processing control program 12 a causes the main screen 16 or the sub screen 17 to display a list of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registration area 12 f.
- the pronunciation learning processing control program 12 a causes the main screen 16 or the sub screen 17 to display a list of words registered in the word registration area 12 e.
- the pronunciation learning processing control program 12 a causes a pronunciation-associated example sentence including the pronunciation data of the word specified and selected by the user among a list of the words displayed on the main screen 16 or the sub screen 17 , to be extracted from the example sentence pronunciation storage area 12 d , and causes the main screen 16 or the sub screen 17 to display a list of the pronunciation-associated example sentences.
- the pronunciation learning processing control program 12 a causes the main screen 16 or the sub screen 17 to display the pronunciation-associated example sentences including the word whose pronunciation changes in a state where the word is identifiable among other words.
- the pronunciation learning processing control program 12 a also controls an operation performed by the conventional electronic device or pronunciation learning device, in addition to these operations. However, the operation in these conventional techniques will not be described in this description.
- FIG. 2 is a front view showing an external appearance configuration in case where the pronunciation learning device 10 according to the first embodiment of the present invention is realized as an electronic dictionary device 10 D.
- the CPU 11 , the memory 12 and the recording medium reading unit 14 are built in a lower stage side of a device main body which is opened and closed, the key input unit 15 , the sub screen 17 , the speaker 18 a and the microphone 18 b are provided, and the main screen 16 is provided at an upper stage side.
- the key input unit 15 further includes character input keys 15 a , various dictionary specifying keys 15 b , a [Translation/Enter] key 15 c and a [Back/List] key 15 d.
- the main screen 16 of the electronic dictionary device 10 D in FIG. 2 shows an example where the user specifies the English-Japanese dictionary (e.g. English-Japanese Dictionary A) by using the dictionary specifying key 15 b , and, when a search word “apply” is input from the key input unit 15 , explanation information d 1 (e.g. parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, meanings, a usage, and example sentences) of the search word obtained from English-Japanese dictionary (e.g. Genius English-Japanese Dictionary) data stored in the dictionary database 12 b is displayed.
- English-Japanese dictionary e.g. English-Japanese Dictionary A
- various function selection icons I are displayed vertically in a row.
- a “Listen” icon I 1 a “Listen/Compare” icon I 2 , a “Read” icon I 3 and an “Example Sentence” icon I 4 are displayed as examples of these function selection icons I.
- arbitrary icons can also be additionally provided.
- FIG. 3 is a front view showing an external appearance configuration in case where the pronunciation learning device 10 according to the first embodiment of the present invention is realized as a tablet terminal 10 T.
- the CPU 11 , the memory 12 and the recording medium reading unit 14 are built in a terminal main body, and various icons and a software keyboard displayed on the main screen 16 when necessary function as the key input unit 15 . Consequently, it is possible to realize the same function as the electronic dictionary device 10 D shown in FIG. 2 .
- the pronunciation learning device 10 can be realized as not only a mobile device for an electronic dictionary shown in FIG. 2 (electronic dictionary device 10 D) and the tablet terminal 10 T having the dictionary function shown in FIG. 3 , but also so-called electronic devices such as mobile telephones, electronic books and mobile game machines.
- the pronunciation learning device 10 when used to search in a dictionary (S 1 : Yes), the user specifies the English-Japanese dictionary (e.g. English-Japanese Dictionary A) by using the dictionary specifying key 15 b , and inputs a search word “apply” by using the key input unit 15 (S 2 ).
- the English-Japanese dictionary e.g. English-Japanese Dictionary A
- the search word (“apply”) is searched in the specified dictionary (e.g. English-Japanese Dictionary A) stored in the dictionary database 12 b , and the explanation information d 1 of the search word (e.g. parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, meanings, a usage, and example sentences) is displayed on the main screen 16 (S 3 ).
- the specified dictionary e.g. English-Japanese Dictionary A
- the explanation information d 1 of the search word e.g. parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, meanings, a usage, and example sentences
- the “Listen” icon I 1 When the user touches the “Listen” icon I 1 (S 5 : YES), the “Listen” icon I 1 is displayed by way of monochrome inversion (not shown) and is displayed in an active state, and pronunciation data (“aplai”) corresponding to the search word is extracted from the specified dictionary, and the extracted pronunciation data is output from the speaker 18 a (S 6 ). Consequently, the user can learn the pronunciation of the search word by listening to the pronunciation data of the search word. Further, when the output of the pronunciation data is finished, the monochrome inversion of the “Listen” icon I 1 returns to the original state and is displayed in a non-active state, and the processing moves to step S 11 .
- pronunciation output guidance information d 2 is displayed below the explanation information d 1 to encourage the user to record a pronunciation.
- the uttered pronunciation data is registered in the user pronunciation registration area 12 g (S 9 ). Further, the registered pronunciation data is output from the speaker 18 a (S 10 ). Consequently, the user can listen to and compare correct pronunciation data included in the dictionary and pronunciation data uttered by the user.
- monochrome inversion of the “Listen/Compare” icon I 2 returns to the original state and is displayed in a non-activate state, and the processing moves to step S 11 .
- step S 7 when the user does not listen to and compare the pronunciations in step S 7 (S 7 : No), the processing moves to another processing (which will not be described in detail) of causing the speaker 18 a to read the explanation information d 1 by specifying the “Read” icon I 3 or causing the speaker 18 a to output pronunciation data of an example sentence by specifying the “Example Sentence” icon I 4 .
- step S 11 When pronunciation learning is finished in step S 11 (S 11 : Yes), the processing moves to step S 20 and, when the pronunciation learning is not finished (S 11 : No), the processing returns to S 5 .
- step S 20 whether or not the search word searched in step S 1 has already been registered in the word registration area 12 e is determined. This determination is made by cross-checking the search word searched in step S 1 and words registered in the word registration area 12 e . Further, in case where the search word has not been registered (S 20 : Yes), the search word searched in step S 1 is registered in the word registration area 12 e (S 21 ), and the processing moves to step S 22 . Meanwhile, in case where the search word has been registered (S 20 : No), the processing returns to step S 1 .
- step S 22 pronunciation-associated example sentences including the word registered in the word registration area 12 e and all pronunciation-associated example sentences stored in the example sentence pronunciation storage area 12 d are sequentially cross-checked.
- pronunciation-associated example sentences including only the pronunciation data of the words registered in the word registration area 12 e are extracted from the example sentence pronunciation storage area 12 d (S 23 : Yes).
- the pronunciation-associated example sentences extracted in step S 23 are registered in the pronunciation-associated example sentence registration area 12 f , and counted up (S 25 ).
- step S 26 when cross-checking of all pronunciation-associated example sentences stored in the example sentence pronunciation storage area 12 d is finished, and there is no pronunciation-associated example sentence which needs to be cross-checked in the example sentence pronunciation storage area 12 d (S 26 : Yes), the number counted up in step S 25 is displayed on the main screen 16 (S 27 ) and the processing returns to step S 1 .
- the processing returns to step S 22 .
- FIG. 10 is a view showing an example where counted-up number display areas d 3 and d 4 are displayed per source on the main screen 16 in step S 27 .
- step S 27 that three pronunciation-associated example sentences deriving from English phrases and six pronunciation-associated example sentences deriving from the dictionary are registered from the example sentence pronunciation storage area 12 d to the pronunciation-associated example sentence registration area 12 f is displayed on the count-up number display area d 3 and the count-up number display area d 4 .
- step S 22 in case where a pronunciation-associated example sentence including only the pronunciation data of the words registered in the word registration area 12 e has not been extracted as a result of the cross-check performed in step S 22 (S 23 : No), or when, even though pronunciation-associated example sentences are extracted in step S 23 , the extracted pronunciation-associated example sentences have already been registered in the pronunciation-associated example sentence registration area 12 f (S 24 : No), the processing returns to step S 1 .
- the pronunciation learning device 10 can extract pronunciation-associated example sentences including only the pronunciation data of the words registered in the word registration area 12 e as described above. Consequently, the user can efficiently accumulate pronunciation-associated example sentences for which only words the user has learned are used, i.e., only pronunciation-associated example sentences which the user needs to learn.
- step S 1 when search is not performed in the dictionary in step S 1 (S 1 : No), the user can select whether or not to perform registered pronunciation example sentence list processing (S 31 to S 39 ).
- step S 31 When the registered pronunciation example sentence list processing is selected in step S 31 (S 31 : Yes), a list of pronunciation-associated example sentences registered in the pronunciation-associated example sentence registration area 12 f in step S 25 is displayed from an example sentence display area d 7 secured on the main screen 16 as shown in FIG. 11 . Further, as the function selection icons I, a “ ⁇ ” icon I 5 and a “ ⁇ ” icon I 6 for scrolling the display screen are displayed in addition to the “Listen” icon I 1 (S 32 ).
- an icon indicating a corresponding variation of a pronunciation-associated example sentence including a variation word and pronunciation changing words among a list of pronunciation-associated example sentences displayed in the example sentence display area d 7 is displayed in a variation display field d 6 secured on the main screen 16 (S 34 ).
- [V] (d 6 V) among icons displayed in the variation display field d 6 in FIG. 11 means that a variation is used (e.g. “apply” is not in an original form like “applying” or “applied”).
- [C] (d 6 C) means that there are phonetically connected words when an example sentence is vocally output.
- [D] (d 6 D) means that there are phonetically disappearing words when an example sentence is vocally output.
- [E] (d 6 E) means that there are phonetically changing words when an example sentence is vocally output.
- Such a variation word and pronunciation changing words are recognized by a known technique such as waveform analysis; therefore, a detailed description thereof will be omitted.
- a head pronunciation-associated example sentence (“Why don't you apply?” in case of FIG. 11 ) among a displayed list of pronunciation-associated example sentences is targeted (S 35 ), and meanings, example sentences and a pronunciation are displayed in a preview area d 8 secured below the main screen 16 (S 36 ).
- the pronunciation-associated example sentence displayed in the preview area d 8 includes the pronunciation changing words, the relevant portion is underlined and displayed.
- FIG. 11 shows an example where “don't you” including pronunciation changing words of “Why don't you apply?” which is a pronunciation-associated example sentence displayed in the preview area d 8 is underlined.
- step S 39 whether or not to select another pronunciation-associated example sentence from the pronunciation-associated example sentences displayed in the example sentence display area d 7 is selected.
- another pronunciation-associated example sentence is selected (S 39 : Yes)
- FIG. 11 another pronunciation-associated example sentence is specified by touching one of the pronunciation-attached example sentences displayed as the list in the example sentence display area d 7 by the finger or the touch pen 20 , and the processing returns to step S 36 .
- FIG. 11 shows that only the eleven pronunciation-associated example sentences indicated in [A] to [K] are displayed in the example sentence display area d 7 , and simply shows an example where the pronunciation-associated example sentences are displayed at a time in the example sentence display area d 7 .
- the number of pronunciation-associated example sentences is not limited to this.
- the “ ⁇ ” icon I 5 or “ ⁇ ” icon I 6 are touched to scroll the screen, so that desired pronunciation-associated example sentences can be displayed in the example sentence display area d 7 .
- step S 31 when the registered pronunciation example sentence list processing is not selected in step S 31 (S 31 : No), the user can select whether or not to perform the learning word list processing (S 41 to S 49 ).
- the pronunciation learning device can flexibly select a pronunciation-associated example sentence which the user wants to vocally output from registered pronunciation-associated example sentences when the user learns a pronunciation of an example sentence.
- step S 41 When the learning word list processing is selected in step S 41 (S 41 : Yes), as shown in FIG. 12 , a list of words registered in the word registration area 12 e is displayed in a word registration display area d 9 secured on the main screen 16 (S 42 ).
- a display source display field d 9 a is further secured in the word registration display area d 9 , and an abbreviated name (e.g. “A” indicating English-Japanese Dictionary A) indicating a source is displayed per word in this source display field d 9 a.
- an abbreviated name e.g. “A” indicating English-Japanese Dictionary A
- a check field d 9 b is further secured in the word registration display area d 9 .
- the user selects an arbitrary number of words among the words registered in the word registration area 12 e , and processing of searching for pronunciation-associated example sentences including only the pronunciation data of the words checked on as selection targets is performed.
- the check field d 9 b is a field which is necessary for this processing, and explicitly indicates the words checked on by the user as the selection targets from the words displayed in the word registration display area d 9 .
- “check” marks of all words registered in the word registration area 12 e are applied to the check field d 9 b , and all words registered in the word registration area 12 e are checked on and explicitly indicated as the selection targets.
- a pronunciation-associated example sentence display area d 10 is also secured on the main screen 16 , and pronunciation-associated example sentences including only the pronunciation data of the words to which “check” marks are applied to the check field d 9 b are displayed.
- the user can uncheck one of words. More specifically, when the user touches the words displayed in the word registration display area d 9 by the finger or the touch pen 20 and further touches a “check” icon I 7 in a state where the words are in the active state, the words are checked off, the “check” marks are removed from the check field d 9 b of the words and this removal is explicitly indicated.
- the user can check on the words which have been checked off once and whose “check” marks in the check field d 9 b have been removed, by touching the words by the finger or the touch pen 20 and by touching the “check” icon I 7 in a state where the words are placed in the active state, and the “check” marks are applied to the check field d 9 b.
- FIG. 12 shows only an example where the number of words in [A] to [M] displayed in the word registration display area d 9 is thirteen.
- the number of words is not limited to this.
- step S 44 the check change processing is performed and then the processing returns to step S 42 .
- This check change processing will be described below with reference to the flowchart in FIG. 8 .
- a preview display area d 11 is secured on the main screen 16 and a translation, a pronunciation, related sentences and a usage are displayed in addition to this pronunciation-associated example sentence (S 46 ).
- this pronunciation-associated example sentence includes pronunciation changing words, the relevant portion is underlined.
- icons e.g. a note icon I 10 , a search icon I 11 , a marker icon I 12 and a post-it icon I 13 ) which realize various editor functions, and a vocabulary notebook icon I 14 which retrieve words registered in the word registration area 12 e are optionally provided as the function selection icons I.
- the preview display area d 11 displays an example where a pronunciation-associated example sentence “Please consider it.” is specified in step S 45 , a translation of this pronunciation-associated example sentence “Yoroshikuonegaiitashimasu”, a pronunciation (pl ⁇ :z k ns ⁇ d r ⁇ t), a related sentence (I hope we can give you good news.) and a usage are displayed in the preview display area d 11 in response to this specification, and, further, pronunciation changing words (consider it) of the pronunciation-associated example sentence (Please consider it.) are underlined and displayed.
- a pronunciation output icon I 9 is also provided.
- the user wants to vocally output the pronunciation-associated example sentence displayed in step S 46 (S 47 : Yes)
- the user touches the pronunciation output icon I 9 by the finger or the touch pen 20 .
- pronunciation data of the pronunciation-associated example sentence displayed in the preview display area d 11 is output from the speaker 18 a (S 48 ), and then the processing moves to step S 49 .
- step S 47 when the user does not want to vocally output the pronunciation-associated example sentence displayed in step S 46 (S 47 ), the processing directly moves to step S 49 .
- step S 49 whether or not to continue the learning word list processing is determined.
- this processing is finished (S 49 : Yes)
- the processing returns to S 1 and, when this processing is continued (S 49 : No), the processing returns to step S 45 .
- step S 41 when the user does not want the learning word list processing in step S 41 (S 41 : No) or when the user does not specify any example sentence in step S 45 (S 45 : No), for example, the processing moves to another processing (which will not be described in detail) of using the pronunciation learning device 10 according to the present embodiment as a normal electronic dictionary.
- step S 44 check change processing performed in step S 44 will be described with reference to the flowchart in FIG. 8 .
- steps S 44 b to S 44 c are performed per word. Hence, when there is a plurality of words subjected to the check-off operation in step S 43 , whether or not the processing in steps S 44 b to S 44 c has been performed on all words is determined in step S 44 d to repeat the processing in steps S 44 b to S 44 c on all of these words.
- step S 44 b to S 44 c has been performed on all words checked off in step S 43 (S 44 d : Yes ⁇ S 44 e ). Further, in case where it is determined that processing in steps S 44 b to S 44 c has been performed on all words checked off in step S 43 (S 44 d : Yes ⁇ S 44 e ), the processing returns to step S 44 shown in FIG. 7 .
- FIG. 14 shows the display example of the main screen 16 after this check change processing is performed.
- step S 43 as a result that a check-off operation has been performed on words such as “brake” and “why”, “check” marks are removed from check field d 9 b of “brake” and “why” in FIG. 14 .
- step S 44 c as a result that pronunciation-associated example sentences including “brake” and “why” are not placed in a non-display state, only pronunciation-associated example sentences other than the pronunciation-associated example sentences including “brake” or “why” are left in the pronunciation-associated example sentence display area d 10 as shown in FIG. 12 .
- the pronunciation learning device 10 when used, the user can flexibly select pronunciation-associated example sentences which the user wants to vocally output by performing a check-on/off operation of words.
- the pronunciation learning device 10 can associate pronunciation data of words and a pronunciation data of an example sentence including these words to provide to the user. Consequently, the user can efficiently learn the pronunciation data of the words and the pronunciation data of the example sentence including these words.
- a method of each processing and a database of the pronunciation learning device 10 according to the present embodiment i.e., each method of processing (part 1 to part 5 ) shown in the flowcharts in FIGS. 4 to 8 and the dictionary database 12 b can be stored and distributed as a program which can be executed by the computer, in the external recording medium 13 such as memory cards (e.g. a ROM card and a RAM card), magnetic disks (floppy disks and hard disks), optical disks (CD-ROMs and DVDs) and semiconductor memories.
- the external recording medium 13 such as memory cards (e.g. a ROM card and a RAM card), magnetic disks (floppy disks and hard disks), optical disks (CD-ROMs and DVDs) and semiconductor memories.
- a computer of an electronic device having the main screen 16 and/or, the sub screen 17 can realize processing described with reference to the flowcharts of FIGS. 4 to 8 in the present embodiment by reading to the memory 12 the program stored in this external recording medium 13 and having this read program control
- a pronunciation learning device according to the second embodiment of the present invention will be described.
- a pronunciation learning device 10 is realized as a so-called single electronic device such as an electronic dictionary device 10 D, a tablet terminal 10 T, a mobile telephone, an electronic book and a mobile game has been described.
- a pronunciation learning device 30 includes a terminal 34 and an external server 36 which are connected through a communication network 32 such as the Internet.
- such a network configuration is configured by a LAN such as the Ethernet (registered trademark) or a WAN to which a plurality of LANs is connected through a public line or a dedicated line.
- the LAN is configured by multiple subnets connected through a router when necessary.
- a WAN optionally includes a firewall which connects to a public line yet will not be shown and described in detail.
- the terminal 34 includes a CPU 11 , a recording medium reading unit 14 , a key input unit 15 , a main screen 16 , a sub screen 17 , a speaker 18 a , a microphone 18 b and a communication unit 38 which are connected with each other through a communication bus 19 . That is, the terminal 34 is just configured to include the communication unit 38 which communicates with the communication network 32 such as the Internet in place of the memory 12 for the pronunciation learning device 10 shown in FIG. 1 .
- This terminal 34 is realized as a so-called single electronic device such as a personal computer, a tablet terminal, a mobile telephone, an electronic book and a mobile game machine.
- the external server 36 includes a memory 12 shown in FIG. 1 .
- the terminal 34 causes the communication unit 38 to access the external server 36 through the communication network 32 , activates a pronunciation learning processing control program 12 a stored in the memory 12 provided in the external server 36 , and performs writing/reading operations on a dictionary database 12 b and various storage (registration) areas 12 c to 12 g under control of the pronunciation learning processing control program 12 a to provide the same functions as those of the pronunciation learning device 10 according to the first embodiment to users of the terminal 34 .
- the user can obtain an effect of the pronunciation learning device 10 according to the first embodiment by using a communication terminal which the user is used to using without purchasing a dedicated device. Consequently, it is possible to enhance user friendliness.
- the pronunciation learning processing control program 12 a and the dictionary database 12 b are provided in the external server 36 , so that, even when the pronunciation learning processing control program 12 a is updated (upgraded) or a new dictionary is introduced, it is possible to immediately enjoy a benefit of the update and the introduction without buying a terminal or installing a new application or dictionary.
- each embodiment includes inventions of various stages, and various inventions can be extracted by optional combinations of a plurality of disclosed components. For example, even when some components are removed from all components described in each embodiment or some components are combined in different forms, it is possible to solve the problem described in SUMMARY. When an effect described in paragraph [0010] is obtained, a configuration obtained by removing or combining these components can be extracted as an invention.
- part of the memory 12 may be optionally provided to the terminal 34 instead of the external server 36 .
- part of the memory 12 may be optionally provided to the terminal 34 instead of the external server 36 .
- providing only the pronunciation learning processing control program 12 a and the dictionary database 12 b to the memory 12 of the external server 36 , and providing other storage (registration) areas 12 c to 12 g to the memory of the terminal 34 (not shown) are understood to be part of the present invention.
Abstract
A pronunciation learning device includes: an example sentence text storage area in which a plurality of example sentence texts is stored; an example sentence pronunciation storage area in which each of the example sentence texts stored in the example sentence text storage area is associated with pronunciation data and stored as a pronunciation-associated example sentence; a pronunciation learning processing control program configured to vocally output pronunciation data of a word specified by a user's operation; and a pronunciation-associated example sentence registration area in which a pronunciation-associated example sentence including the pronunciation data of the word is extracted from the example sentence pronunciation storage area and is registered, and the pronunciation learning processing control program reads pronunciation data of any one of registered pronunciation-associated example sentences, from the example sentence pronunciation storage area, and vocally outputs the read pronunciation data.
Description
- 1. Technical Field
- The present invention relates to a pronunciation learning device and a control program thereof. More particularly, the present invention relates to a pronunciation learning device which has a function which allows a user to efficiently learn pronunciations, and a control program thereof.
- 2. Related Art
- Recent electronic dictionaries have, for example, a function of outputting a pronunciation of a specific example sentence as disclosed in JP 2013-37251 A or a function of searching for a word in an example sentence, vocally outputting the example sentence and displaying a translation for learning a pronunciation of words as disclosed in JP 2006-268501 A. Therefore, some electronic dictionaries are used not only for dictionaries but also for pronunciation learning devices.
- However, such a conventional pronunciation learning device has the following problem.
- That is, as described above, the conventional pronunciation learning device can not only output pronunciations of words but also output pronunciations of example sentences including the words. However, such a conventional pronunciation learning device cannot associate and provide pronunciations of words and a pronunciation of an example sentence including these words.
- Hence, when learning pronunciations of words included in an example sentence vocally output from a conventional pronunciation learning device, a user needs to operate a pronunciation learning device to output a pronunciation per individual word included this example sentence.
- Therefore, even when the user wants to learn both of pronunciations of new words and a pronunciation of an example sentence including these words, a conventional pronunciation learning device does not have a function of associating pronunciations of words and a pronunciation of an example sentence including these words. Therefore, since there is a problem that it is bothersome for a user to operate the pronunciation learning device to output a pronunciation per individual word included in the example sentence, the user cannot efficiently learn the pronunciations.
- The present invention has been made in light of such a situation. It is therefore an object of the present invention to provide a pronunciation learning device which can associate and provide pronunciations of words and a pronunciation of an example sentence including these words and, consequently, provide a function of enabling a user to efficiently learn pronunciations.
- A pronunciation learning device according to the present invention includes: an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words; an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentence text storage unit with pronunciation data as a pronunciation-associated example sentence; a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation; a pronunciation-associated example sentence registering unit configured to extract from the example sentence pronunciation storage unit a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence; and an example sentence pronunciation output unit configured to read from the example sentence pronunciation storage unit pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering unit, and to vocally output the read pronunciation data.
- The present invention can realize a pronunciation learning device which can associate and provide pronunciations of words and a pronunciation of an example sentence including these words, and which allows a user to efficiently learn pronunciations.
-
FIG. 1 is a block diagram showing a configuration of an electronic circuit of a pronunciation learning device according to a first embodiment of the present invention; -
FIG. 2 is a front view showing an external appearance configuration in case where the pronunciation learning device according to the first embodiment of the present invention is realized as an electronic dictionary device; -
FIG. 3 is a front view showing an external appearance configuration in case where the pronunciation learning device according to the first embodiment of the present invention is realized as a tablet terminal; -
FIG. 4 is a flowchart showing processing (part 1) of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 5 is a flowchart showing processing (part 2) of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 6 is a flowchart showing processing (part 3) of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 7 is a flowchart showing processing (part 4) of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 8 is a flowchart showing processing (part 5) of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 9 is a view showing an example (part 1) of a display screen of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 10 is a view showing an example (part 2) of the display screen of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 11 is a view showing an example (part 3) of the display screen of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 12 is a view showing an example (part 4) of the display screen of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 13 is a view showing an example (part 5) of the display screen of the pronunciation learning device according to the first embodiment of the present invention; -
FIG. 14 is a view showing an example (part 6) of the display screen of the pronunciation learning device according to the first embodiment of the present invention; and -
FIG. 15 is a conceptual diagram showing a configuration example of a pronunciation learning device according to a second embodiment of the present invention. - Each embodiment of the present invention will be described below with reference to the drawings.
- A pronunciation learning device according to a first embodiment of the present invention will be described.
-
FIG. 1 is a block diagram showing an electronic circuit of apronunciation learning device 10 according to the first embodiment of the present invention. - This
pronunciation learning device 10 is a device which is preferable to learn pronunciations of foreign languages in particular, and includes aCPU 11, amemory 12, a recordingmedium reading unit 14, akey input unit 15, amain screen 16, asub screen 17, aspeaker 18 a and amicrophone 18 b which are connected with each other through acommunication bus 19. - The
CPU 11 controls an operation of thepronunciation learning device 10 according to a pronunciation learningprocessing control program 12 a stored in advance in thememory 12, the pronunciation learningprocessing control program 12 a read to thememory 12 from anexternal recording medium 13 such as a ROM card through the recordingmedium reading unit 14, or the pronunciation learningprocessing control program 12 a downloaded from a web server (a program server in this case) on the Internet and read to thememory 12. - The pronunciation learning
processing control program 12 a also includes a communication program for performing data communication with each web server on the Internet or a user PC (Personal Computer) externally connected to thepronunciation learning device 10. Further, the pronunciation learningprocessing control program 12 a is activated according to an input signal corresponding to a user's operation from thekey input unit 15, an input signal corresponding to a user's operation from themain screen 16 or thesub screen 17 having a touch panel color display function, a communication signal for the web server on the externally connected Internet or a connection communication signal for therecording medium 13 such as an EEPROM, a RAM or a ROM externally connected through the recordingmedium reading unit 14. - The
memory 12 includes adictionary database 12 b, an example sentencetext storage area 12 c, an example sentencepronunciation storage area 12 d, aword registration area 12 e, a pronunciation-associated examplesentence registration area 12 f, and a userpronunciation registration area 12 g. - In the
dictionary database 12 b, dictionaries (an English-Japanese dictionary, a Japanese-English dictionary, an English-English dictionary, a Chinese-Japanese dictionary, a Japanese-Chinese dictionary, Chinese phrases and a Chinese-Chinese dictionary) and phrases of learning target foreign languages such as English or Chinese are stored. A dictionary stores pieces of general dictionary information such as parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, word pronunciation data, meanings, a usage, example sentences, and an example sentence pronunciation data per word. As phrases, information such as example sentences, meanings and pronunciations are stored per situation such as a travel, business, life and cooking. In addition, the numbers of dictionaries and phrases are not limited to singular forms. Thedictionary database 12 b may store a plurality of dictionaries of similar types like a plurality of English-Japanese dictionaries. - The example sentence
text storage area 12 c is a storage area in which example sentence texts are extracted from the dictionaries and the phrases stored in thedictionary database 12 b under control of the pronunciation learningprocessing control program 12 a, and are stored together with names of sources (e.g. English-Japanese Dictionary A). In the other words, the example sentencetext storage area 12 c configures an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words. - The example sentence
pronunciation storage area 12 d is a storage area in which each example sentence text stored in the example sentencetext storage area 12 c is associated with the pronunciation data and stored as a pronunciation-associated example sentence under control of the pronunciation learningprocessing control program 12 a. In the other words, the example sentencepronunciation storage area 12 d configures an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentencetext storage area 12 c with pronunciation data as a pronunciation-associated example sentence. - The
word registration area 12 e is a registration area in which a word of pronunciation data vocally output from thespeaker 18 a is registered under control of the pronunciation learningprocessing control program 12 a. In addition, there are also words registered in advance in theword registration area 12 e. The words registered in advance are basic words for which the user does not need to practice, and correspond to, for example, “I”, “to” and “the” in case of English. In the other words, theword registration area 12 e configures a word registering unit configured to register the word corresponding to the pronunciation data vocally output by thespeaker 18 a. And, thespeaker 18 a configures a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation. - The pronunciation-associated example
sentence registration area 12 f is a registration area in which a pronunciation-associated example sentence including the pronunciation data of the word registered in theword registration area 12 e is extracted from the example sentencepronunciation storage area 12 d and the extracted pronunciation-associated example sentence is registered under control of the pronunciation learningprocessing control program 12 a. One of items of pronunciation data of pronunciation-associated example sentences registered in the pronunciation-associated examplesentence registration area 12 f is read from the example sentencepronunciation storage area 12 d and is vocally output from thespeaker 18 a under control of the pronunciation learningprocessing control program 12 a performed by a user's operation. In the other words, the pronunciation-associated examplesentence registration area 12 f configures a pronunciation-associated example sentence registering unit configured to extract from the example sentencepronunciation storage area 12 d a pronunciation-associated example sentence including the pronunciation data of the word registered in theword registering area 12 e, and to register the extracted pronunciation-associated example sentence. And, thespeaker 18 a also configures an example sentence pronunciation output unit configured to read from the example sentencepronunciation storage area 12 d pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated examplesentence registering area 12 f, and to vocally output the read pronunciation data. - The user
pronunciation registration area 12 g is a storage area in which pronunciation data obtained by themicrophone 18 b and pronounced by the user is stored. - The example sentence
text storage area 12 c, the example sentencepronunciation storage area 12 d, theword registration area 12 e and the pronunciation-associated examplesentence registration area 12 f are preferably provided per language. - The pronunciation learning
processing control program 12 a which is applied to thepronunciation learning device 10 according to the first embodiment of the present invention controls an operation performed by a conventional electronic dictionary or pronunciation learning device, and, the following operation by using thesedictionary database 12 b, example sentencetext storage area 12 c, the example sentencepronunciation storage area 12 d, theword registration area 12 e and the pronunciation-associated examplesentence registration area 12 f. - The pronunciation learning
processing control program 12 a causes a new example sentence text to be stored in the example sentencetext storage area 12 c. - The pronunciation learning
processing control program 12 a causes each example sentence text stored in the example sentencetext storage area 12 c to be associated as a pronunciation-associated example sentence with pronunciation data and stored in the example sentencepronunciation storage area 12 d. - The pronunciation learning
processing control program 12 a causes thespeaker 18 a to vocally output pronunciation data of a word specified by a user's operation by using thekey input unit 15. - The word corresponding to pronunciation data vocally output from the
speaker 18 a is registered in theword registration area 12 e. - The pronunciation learning
processing control program 12 a causes a pronunciation-associated example sentence including the pronunciation data of the word registered in theword registration area 12 e to be extracted from the example sentencepronunciation storage area 12 d, and causes the extracted pronunciation-associated example sentence to be registered in the pronunciation-associated examplesentence registration area 12 f. - The pronunciation learning
processing control program 12 a causes pronunciation data specified by a user's operation among pronunciation-associated example sentences registered in the pronunciation-associated examplesentence registration area 12 f, to be read from the example sentencepronunciation storage area 12 d, and causes thespeaker 18 a to vocally output the pronunciation data. - The word specified by the user's operation using the
key input unit 15 and corresponding to the pronunciation data vocally output from thespeaker 18 a is registered in theword registration area 12 e. - In response to registration of the word in the
word registration area 12 e, the pronunciation learningprocessing control program 12 a causes themain screen 16 or thesub screen 17 to display a list of the pronunciation-associated example sentences registered in the pronunciation-associated examplesentence registration area 12 f. - The pronunciation learning
processing control program 12 a causes themain screen 16 or thesub screen 17 to display a list of words registered in theword registration area 12 e. - The pronunciation learning
processing control program 12 a causes a pronunciation-associated example sentence including the pronunciation data of the word specified and selected by the user among a list of the words displayed on themain screen 16 or thesub screen 17, to be extracted from the example sentencepronunciation storage area 12 d, and causes themain screen 16 or thesub screen 17 to display a list of the pronunciation-associated example sentences. - When causing the
main screen 16 or thesub screen 17 to display the list of the pronunciation-associated example sentences, the pronunciation learningprocessing control program 12 a causes themain screen 16 or thesub screen 17 to display the pronunciation-associated example sentences including the word whose pronunciation changes in a state where the word is identifiable among other words. - In addition, the pronunciation learning
processing control program 12 a also controls an operation performed by the conventional electronic device or pronunciation learning device, in addition to these operations. However, the operation in these conventional techniques will not be described in this description. -
FIG. 2 is a front view showing an external appearance configuration in case where thepronunciation learning device 10 according to the first embodiment of the present invention is realized as anelectronic dictionary device 10D. - In case of the
electronic dictionary device 10D inFIG. 2 , theCPU 11, thememory 12 and the recordingmedium reading unit 14 are built in a lower stage side of a device main body which is opened and closed, thekey input unit 15, thesub screen 17, thespeaker 18 a and themicrophone 18 b are provided, and themain screen 16 is provided at an upper stage side. - The
key input unit 15 further includescharacter input keys 15 a, variousdictionary specifying keys 15 b, a [Translation/Enter]key 15 c and a [Back/List] key 15 d. - The
main screen 16 of theelectronic dictionary device 10D inFIG. 2 shows an example where the user specifies the English-Japanese dictionary (e.g. English-Japanese Dictionary A) by using thedictionary specifying key 15 b, and, when a search word “apply” is input from thekey input unit 15, explanation information d1 (e.g. parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, meanings, a usage, and example sentences) of the search word obtained from English-Japanese dictionary (e.g. Genius English-Japanese Dictionary) data stored in thedictionary database 12 b is displayed. - On a leftmost side of the
main screen 16, various function selection icons I are displayed vertically in a row. In the example inFIG. 2 , a “Listen” icon I1, a “Listen/Compare” icon I2, a “Read” icon I3 and an “Example Sentence” icon I4 are displayed as examples of these function selection icons I. However, arbitrary icons can also be additionally provided. When the user touches one of these icons I1 to I4 by the finger or a touch pen, the pronunciation learningprocessing control program 12 a is activated to execute predetermined processing. -
FIG. 3 is a front view showing an external appearance configuration in case where thepronunciation learning device 10 according to the first embodiment of the present invention is realized as atablet terminal 10T. - In case of the
tablet terminal 10T inFIG. 3 , theCPU 11, thememory 12 and the recordingmedium reading unit 14 are built in a terminal main body, and various icons and a software keyboard displayed on themain screen 16 when necessary function as thekey input unit 15. Consequently, it is possible to realize the same function as theelectronic dictionary device 10D shown inFIG. 2 . - The
pronunciation learning device 10 according to the first embodiment of the present invention can be realized as not only a mobile device for an electronic dictionary shown inFIG. 2 (electronic dictionary device 10D) and thetablet terminal 10T having the dictionary function shown inFIG. 3 , but also so-called electronic devices such as mobile telephones, electronic books and mobile game machines. - Next, an example of various types of processing performed when the pronunciation learning
processing control program 12 a operates will be described with reference to flowcharts shown inFIGS. 4 to 8 and display screen examples shown inFIGS. 9 to 14 . - As shown in the flowchart in
FIG. 4 , when thepronunciation learning device 10 according to the first embodiment of the present invention is used to search in a dictionary (S1: Yes), the user specifies the English-Japanese dictionary (e.g. English-Japanese Dictionary A) by using thedictionary specifying key 15 b, and inputs a search word “apply” by using the key input unit 15 (S2). - Then, the search word (“apply”) is searched in the specified dictionary (e.g. English-Japanese Dictionary A) stored in the
dictionary database 12 b, and the explanation information d1 of the search word (e.g. parts of speech, an etymology, a spelling, conjugated forms, a phonetic symbol, synonyms, meanings, a usage, and example sentences) is displayed on the main screen 16 (S3). - Further, when the user wants to learn a pronunciation of this search word (S4: Yes), the user touches the “Listen” icon I1 or the “Listen/Compare” icon I2 by the finger or the touch pen (S5 or S7). Meanwhile, when the user does not want to learn the pronunciation of the search word (S4: No), for example, the processing moves to another processing (which will not be described in detail) of using the
pronunciation learning device 10 according to the present embodiment as a normal electronic dictionary. - When the user touches the “Listen” icon I1 (S5: YES), the “Listen” icon I1 is displayed by way of monochrome inversion (not shown) and is displayed in an active state, and pronunciation data (“aplai”) corresponding to the search word is extracted from the specified dictionary, and the extracted pronunciation data is output from the
speaker 18 a (S6). Consequently, the user can learn the pronunciation of the search word by listening to the pronunciation data of the search word. Further, when the output of the pronunciation data is finished, the monochrome inversion of the “Listen” icon I1 returns to the original state and is displayed in a non-active state, and the processing moves to step S11. - Meanwhile, as shown in
FIG. 9 , when the user touches the “Listen/Compare” icon I2 by the finger or the touch pen 20 (S7: Yes) instead of the “Listen” icon I1 (S5: No), the “Listen/Compare” icon I2 is displayed by way of monochrome inversion and displayed in an activate state, pronunciation data (“aplai”) corresponding to the search word “apply” is extracted from the specified dictionary, and the extracted pronunciation data is output from thespeaker 18 a (S8). - Further, pronunciation output guidance information d2 is displayed below the explanation information d1 to encourage the user to record a pronunciation. When the user utters a pronunciation (“aplai”) toward the
microphone 18 b according to this pronunciation output guidance information d2, the uttered pronunciation data is registered in the userpronunciation registration area 12 g (S9). Further, the registered pronunciation data is output from thespeaker 18 a (S10). Consequently, the user can listen to and compare correct pronunciation data included in the dictionary and pronunciation data uttered by the user. When the user finishes listening to and comparing the pronunciations, monochrome inversion of the “Listen/Compare” icon I2 returns to the original state and is displayed in a non-activate state, and the processing moves to step S11. - In addition, when the user does not listen to and compare the pronunciations in step S7 (S7: No), the processing moves to another processing (which will not be described in detail) of causing the
speaker 18 a to read the explanation information d1 by specifying the “Read” icon I3 or causing thespeaker 18 a to output pronunciation data of an example sentence by specifying the “Example Sentence” icon I4. - When pronunciation learning is finished in step S11 (S11: Yes), the processing moves to step S20 and, when the pronunciation learning is not finished (S11: No), the processing returns to S5.
- In step S20 shown in the flowchart in
FIG. 5 , whether or not the search word searched in step S1 has already been registered in theword registration area 12 e is determined. This determination is made by cross-checking the search word searched in step S1 and words registered in theword registration area 12 e. Further, in case where the search word has not been registered (S20: Yes), the search word searched in step S1 is registered in theword registration area 12 e (S21), and the processing moves to step S22. Meanwhile, in case where the search word has been registered (S20: No), the processing returns to step S1. - In step S22, pronunciation-associated example sentences including the word registered in the
word registration area 12 e and all pronunciation-associated example sentences stored in the example sentencepronunciation storage area 12 d are sequentially cross-checked. According to this cross-check processing, pronunciation-associated example sentences including only the pronunciation data of the words registered in theword registration area 12 e are extracted from the example sentencepronunciation storage area 12 d (S23: Yes). Further, when there are not the extracted pronunciation-associated example sentences in the pronunciation-associated examplesentence registration area 12 f (S24: Yes), the pronunciation-associated example sentences extracted in step S23 are registered in the pronunciation-associated examplesentence registration area 12 f, and counted up (S25). - Further, when cross-checking of all pronunciation-associated example sentences stored in the example sentence
pronunciation storage area 12 d is finished, and there is no pronunciation-associated example sentence which needs to be cross-checked in the example sentencepronunciation storage area 12 d (S26: Yes), the number counted up in step S25 is displayed on the main screen 16 (S27) and the processing returns to step S1. When there are pronunciation-associated example sentences which need to be cross-checked (S26: No), the processing returns to step S22. -
FIG. 10 is a view showing an example where counted-up number display areas d3 and d4 are displayed per source on themain screen 16 in step S27. In this example, in step S27, that three pronunciation-associated example sentences deriving from English phrases and six pronunciation-associated example sentences deriving from the dictionary are registered from the example sentencepronunciation storage area 12 d to the pronunciation-associated examplesentence registration area 12 f is displayed on the count-up number display area d3 and the count-up number display area d4. - Meanwhile, in case where a pronunciation-associated example sentence including only the pronunciation data of the words registered in the
word registration area 12 e has not been extracted as a result of the cross-check performed in step S22 (S23: No), or when, even though pronunciation-associated example sentences are extracted in step S23, the extracted pronunciation-associated example sentences have already been registered in the pronunciation-associated examplesentence registration area 12 f (S24: No), the processing returns to step S1. - The
pronunciation learning device 10 according to the present embodiment can extract pronunciation-associated example sentences including only the pronunciation data of the words registered in theword registration area 12 e as described above. Consequently, the user can efficiently accumulate pronunciation-associated example sentences for which only words the user has learned are used, i.e., only pronunciation-associated example sentences which the user needs to learn. - Meanwhile, when search is not performed in the dictionary in step S1 (S1: No), the user can select whether or not to perform registered pronunciation example sentence list processing (S31 to S39).
- An example of the registered pronunciation example sentence list processing (S31 to S39) will be described with reference to the flowchart in
FIG. 6 and the display screen example inFIG. 11 . - When the registered pronunciation example sentence list processing is selected in step S31 (S31: Yes), a list of pronunciation-associated example sentences registered in the pronunciation-associated example
sentence registration area 12 f in step S25 is displayed from an example sentence display area d7 secured on themain screen 16 as shown inFIG. 11 . Further, as the function selection icons I, a “↑” icon I5 and a “↓” icon I6 for scrolling the display screen are displayed in addition to the “Listen” icon I1 (S32). - In the example sentence display area d7, “apply” which is the lastly learned word is identified and explicitly indicated in each pronunciation-associated example sentence. Further, an abbreviated name of a source (e.g. the English-Japanese Dictionary A) of each pronunciation-associated example sentence is displayed in a source display field d5 secured on the main screen 16 (S33).
- Further, an icon indicating a corresponding variation of a pronunciation-associated example sentence including a variation word and pronunciation changing words among a list of pronunciation-associated example sentences displayed in the example sentence display area d7 is displayed in a variation display field d6 secured on the main screen 16 (S34). [V] (d6V) among icons displayed in the variation display field d6 in
FIG. 11 means that a variation is used (e.g. “apply” is not in an original form like “applying” or “applied”). [C] (d6C) means that there are phonetically connected words when an example sentence is vocally output. [D] (d6D) means that there are phonetically disappearing words when an example sentence is vocally output. [E] (d6E) means that there are phonetically changing words when an example sentence is vocally output. Such a variation word and pronunciation changing words are recognized by a known technique such as waveform analysis; therefore, a detailed description thereof will be omitted. - In the example sentence display area d7, a head pronunciation-associated example sentence (“Why don't you apply?” in case of
FIG. 11 ) among a displayed list of pronunciation-associated example sentences is targeted (S35), and meanings, example sentences and a pronunciation are displayed in a preview area d8 secured below the main screen 16 (S36). In addition, the pronunciation-associated example sentence displayed in the preview area d8 includes the pronunciation changing words, the relevant portion is underlined and displayed.FIG. 11 shows an example where “don't you” including pronunciation changing words of “Why don't you apply?” which is a pronunciation-associated example sentence displayed in the preview area d8 is underlined. - When the “Listen” icon I1 is touched in this state, pronunciation data of the pronunciation-associated example sentence displayed in the preview area d8 is output from the
speaker 18 a (S37: Yes→S38). - Thus, even when the “Listen” icon I1 is touched and a pronunciation of an example sentence is output from the
speaker 18 a (S37: Yes→S38), or when the “Listen” icon I1 is not touched (S37: No), subsequent processing moves to step S39 either way. - In step S39, whether or not to select another pronunciation-associated example sentence from the pronunciation-associated example sentences displayed in the example sentence display area d7 is selected. When another pronunciation-associated example sentence is selected (S39: Yes), as shown in
FIG. 11 , another pronunciation-associated example sentence is specified by touching one of the pronunciation-attached example sentences displayed as the list in the example sentence display area d7 by the finger or thetouch pen 20, and the processing returns to step S36.FIG. 11 shows that only the eleven pronunciation-associated example sentences indicated in [A] to [K] are displayed in the example sentence display area d7, and simply shows an example where the pronunciation-associated example sentences are displayed at a time in the example sentence display area d7. However, the number of pronunciation-associated example sentences is not limited to this. When there are pronunciation-associated example sentences which cannot be displayed at a time in the example sentence display area d7, the “↑” icon I5 or “↓” icon I6 are touched to scroll the screen, so that desired pronunciation-associated example sentences can be displayed in the example sentence display area d7. - Meanwhile, when no more pronunciation-associated example sentence is selected and the registered pronunciation example sentence list processing is finished (S39: No), the [Back/List] key 15 d shown in
FIG. 2 is pushed to return to dictionary search processing S1 and finish registered pronunciation example sentence list processing S31 (S40). - Meanwhile, when the registered pronunciation example sentence list processing is not selected in step S31 (S31: No), the user can select whether or not to perform the learning word list processing (S41 to S49).
- As described above, the pronunciation learning device according to the present embodiment can flexibly select a pronunciation-associated example sentence which the user wants to vocally output from registered pronunciation-associated example sentences when the user learns a pronunciation of an example sentence.
- Next, an example of learning word list processing (S41 to S49) will be described with reference to the flowcharts in
FIGS. 7 and 8 and the display screen examples inFIGS. 12 to 14 . - When the learning word list processing is selected in step S41 (S41: Yes), as shown in
FIG. 12 , a list of words registered in theword registration area 12 e is displayed in a word registration display area d9 secured on the main screen 16 (S42). - A display source display field d9 a is further secured in the word registration display area d9, and an abbreviated name (e.g. “A” indicating English-Japanese Dictionary A) indicating a source is displayed per word in this source display field d9 a.
- A check field d9 b is further secured in the word registration display area d9. In the learning word list processing shown in the flowcharts in
FIGS. 7 and 8 , the user selects an arbitrary number of words among the words registered in theword registration area 12 e, and processing of searching for pronunciation-associated example sentences including only the pronunciation data of the words checked on as selection targets is performed. The check field d9 b is a field which is necessary for this processing, and explicitly indicates the words checked on by the user as the selection targets from the words displayed in the word registration display area d9. In a default state, as shown inFIG. 12 , “check” marks of all words registered in theword registration area 12 e are applied to the check field d9 b, and all words registered in theword registration area 12 e are checked on and explicitly indicated as the selection targets. - Further, a pronunciation-associated example sentence display area d10 is also secured on the
main screen 16, and pronunciation-associated example sentences including only the pronunciation data of the words to which “check” marks are applied to the check field d9 b are displayed. - By contrast with this, the user can uncheck one of words. More specifically, when the user touches the words displayed in the word registration display area d9 by the finger or the
touch pen 20 and further touches a “check” icon I7 in a state where the words are in the active state, the words are checked off, the “check” marks are removed from the check field d9 b of the words and this removal is explicitly indicated. - In addition, the user can check on the words which have been checked off once and whose “check” marks in the check field d9 b have been removed, by touching the words by the finger or the
touch pen 20 and by touching the “check” icon I7 in a state where the words are placed in the active state, and the “check” marks are applied to the check field d9 b. - In addition, the “check” icon I7, the “↑” icon I5 and the “↓” icon I6 are also secured as the function selection icons I on the
main screen 16. Hence,FIG. 12 shows only an example where the number of words in [A] to [M] displayed in the word registration display area d9 is thirteen. The number of words is not limited to this. When there are words which cannot be displayed at a time in the registered word display area d9, it is possible to display other words by touching the “↑” icon I5 or the “↓” icon I6 and scrolling the screen, and perform the above check-on/off operation on the displayed words likewise. Further, it is also possible to display other pronunciation-associated example sentences of the pronunciation-associated example sentences displayed in the pronunciation-associated example sentence display area d10 by touching the “↑” icon I5 or the “↓” icon I6 and scrolling the screen. - In case where the check-on/off operation has been performed (S43: Yes), the processing moves to step S44, the check change processing is performed and then the processing returns to step S42. This check change processing will be described below with reference to the flowchart in
FIG. 8 . - Meanwhile, in case where a check-on/off operation has not been performed (S43: No), one of the pronunciation-associated example sentences displayed in the pronunciation-associated example sentence display area d10 is touched by the user's finger or
touch pen 20 and specified (S45). - When the pronunciation-associated example sentence is selected in this way, as shown in the display screen example in
FIG. 13 , a preview display area d11 is secured on themain screen 16 and a translation, a pronunciation, related sentences and a usage are displayed in addition to this pronunciation-associated example sentence (S46). When this pronunciation-associated example sentence includes pronunciation changing words, the relevant portion is underlined. Further, when icons (e.g. a note icon I10, a search icon I11, a marker icon I12 and a post-it icon I13) which realize various editor functions, and a vocabulary notebook icon I14 which retrieve words registered in theword registration area 12 e are optionally provided as the function selection icons I. - The preview display area d11 displays an example where a pronunciation-associated example sentence “Please consider it.” is specified in step S45, a translation of this pronunciation-associated example sentence “Yoroshikuonegaiitashimasu”, a pronunciation (plí:z knsídr ít), a related sentence (I hope we can give you good news.) and a usage are displayed in the preview display area d11 in response to this specification, and, further, pronunciation changing words (consider it) of the pronunciation-associated example sentence (Please consider it.) are underlined and displayed.
- In the preview display area d11, a pronunciation output icon I9 is also provided. When the user wants to vocally output the pronunciation-associated example sentence displayed in step S46 (S47: Yes), the user touches the pronunciation output icon I9 by the finger or the
touch pen 20. Thus, pronunciation data of the pronunciation-associated example sentence displayed in the preview display area d11 is output from thespeaker 18 a (S48), and then the processing moves to step S49. - Meanwhile, when the user does not want to vocally output the pronunciation-associated example sentence displayed in step S46 (S47), the processing directly moves to step S49.
- In step S49, whether or not to continue the learning word list processing is determined. When this processing is finished (S49: Yes), the processing returns to S1 and, when this processing is continued (S49: No), the processing returns to step S45.
- Meanwhile, when the user does not want the learning word list processing in step S41 (S41: No) or when the user does not specify any example sentence in step S45 (S45: No), for example, the processing moves to another processing (which will not be described in detail) of using the
pronunciation learning device 10 according to the present embodiment as a normal electronic dictionary. - Next, check change processing performed in step S44 will be described with reference to the flowchart in
FIG. 8 . - First, all pronunciation-associated example sentences registered in the pronunciation-associated example
sentence registration area 12 f are targeted (S44 a), and, when there are words checked off in step S43 (S44 b: Yes), pronunciation-associated example sentences including the checked-off words are not displayed in the pronunciation-associated example sentence display area d10 (S44 c). - The processing in steps S44 b to S44 c is performed per word. Hence, when there is a plurality of words subjected to the check-off operation in step S43, whether or not the processing in steps S44 b to S44 c has been performed on all words is determined in step S44 d to repeat the processing in steps S44 b to S44 c on all of these words.
- Further, in case where it is determined that processing in steps S44 b to S44 c has been performed on all words checked off in step S43 (S44 d: Yes→S44 e), the processing returns to step S44 shown in
FIG. 7 . -
FIG. 14 shows the display example of themain screen 16 after this check change processing is performed. In step S43, as a result that a check-off operation has been performed on words such as “brake” and “why”, “check” marks are removed from check field d9 b of “brake” and “why” inFIG. 14 . Further, in response to this removal, in step S44 c, as a result that pronunciation-associated example sentences including “brake” and “why” are not placed in a non-display state, only pronunciation-associated example sentences other than the pronunciation-associated example sentences including “brake” or “why” are left in the pronunciation-associated example sentence display area d10 as shown inFIG. 12 . - Consequently, when the
pronunciation learning device 10 according to the present embodiment is used, the user can flexibly select pronunciation-associated example sentences which the user wants to vocally output by performing a check-on/off operation of words. - As described above, the
pronunciation learning device 10 according to the present embodiment can associate pronunciation data of words and a pronunciation data of an example sentence including these words to provide to the user. Consequently, the user can efficiently learn the pronunciation data of the words and the pronunciation data of the example sentence including these words. - In addition, a method of each processing and a database of the
pronunciation learning device 10 according to the present embodiment, i.e., each method of processing (part 1 to part 5) shown in the flowcharts inFIGS. 4 to 8 and thedictionary database 12 b can be stored and distributed as a program which can be executed by the computer, in theexternal recording medium 13 such as memory cards (e.g. a ROM card and a RAM card), magnetic disks (floppy disks and hard disks), optical disks (CD-ROMs and DVDs) and semiconductor memories. Further, a computer of an electronic device having themain screen 16 and/or, thesub screen 17 can realize processing described with reference to the flowcharts ofFIGS. 4 to 8 in the present embodiment by reading to thememory 12 the program stored in thisexternal recording medium 13 and having this read program control the operation. - Further, it is possible to transmit program data for realizing each processing on a communication network in a format of a program code. Further, it is also possible to realize each processing by installing this program data in the computer of the electronic device having the
main screen 16 connected to the communication network and/or thesub screen 17 by way of communication. - A pronunciation learning device according to the second embodiment of the present invention will be described.
- In addition, only different components from those of the first embodiment will be described in the present embodiment, and overlapping description will be omitted. Hence, the same elements as those in the first embodiment will be assigned the same reference numerals below.
- In the first embodiment, a case where a
pronunciation learning device 10 is realized as a so-called single electronic device such as anelectronic dictionary device 10D, atablet terminal 10T, a mobile telephone, an electronic book and a mobile game has been described. - By contrast with this, as shown in
FIG. 15 , a pronunciation learning device 30 according to the second embodiment includes a terminal 34 and anexternal server 36 which are connected through acommunication network 32 such as the Internet. - In addition, such a network configuration is configured by a LAN such as the Ethernet (registered trademark) or a WAN to which a plurality of LANs is connected through a public line or a dedicated line. The LAN is configured by multiple subnets connected through a router when necessary. Further, a WAN optionally includes a firewall which connects to a public line yet will not be shown and described in detail.
- The terminal 34 includes a
CPU 11, a recordingmedium reading unit 14, akey input unit 15, amain screen 16, asub screen 17, aspeaker 18 a, amicrophone 18 b and acommunication unit 38 which are connected with each other through acommunication bus 19. That is, the terminal 34 is just configured to include thecommunication unit 38 which communicates with thecommunication network 32 such as the Internet in place of thememory 12 for thepronunciation learning device 10 shown inFIG. 1 . - This terminal 34 is realized as a so-called single electronic device such as a personal computer, a tablet terminal, a mobile telephone, an electronic book and a mobile game machine.
- Meanwhile, the
external server 36 includes amemory 12 shown inFIG. 1 . - Further, the terminal 34 causes the
communication unit 38 to access theexternal server 36 through thecommunication network 32, activates a pronunciation learningprocessing control program 12 a stored in thememory 12 provided in theexternal server 36, and performs writing/reading operations on adictionary database 12 b and various storage (registration)areas 12 c to 12 g under control of the pronunciation learningprocessing control program 12 a to provide the same functions as those of thepronunciation learning device 10 according to the first embodiment to users of the terminal 34. - According to this configuration, the user can obtain an effect of the
pronunciation learning device 10 according to the first embodiment by using a communication terminal which the user is used to using without purchasing a dedicated device. Consequently, it is possible to enhance user friendliness. Further, the pronunciation learningprocessing control program 12 a and thedictionary database 12 b are provided in theexternal server 36, so that, even when the pronunciation learningprocessing control program 12 a is updated (upgraded) or a new dictionary is introduced, it is possible to immediately enjoy a benefit of the update and the introduction without buying a terminal or installing a new application or dictionary. - The present invention is not limited to each of these embodiments and can be variously deformed without departing from the spirit of the present invention at a stage at which the present invention is carried out. Further, each embodiment includes inventions of various stages, and various inventions can be extracted by optional combinations of a plurality of disclosed components. For example, even when some components are removed from all components described in each embodiment or some components are combined in different forms, it is possible to solve the problem described in SUMMARY. When an effect described in paragraph [0010] is obtained, a configuration obtained by removing or combining these components can be extracted as an invention.
- For example, although not shown, in the second embodiment, part of the
memory 12 may be optionally provided to the terminal 34 instead of theexternal server 36. For example, providing only the pronunciation learningprocessing control program 12 a and thedictionary database 12 b to thememory 12 of theexternal server 36, and providing other storage (registration)areas 12 c to 12 g to the memory of the terminal 34 (not shown) are understood to be part of the present invention.
Claims (12)
1. A pronunciation learning device comprising:
an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words;
an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentence text storage unit with pronunciation data as a pronunciation-associated example sentence;
a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation;
a pronunciation-associated example sentence registering unit configured to extract from the example sentence pronunciation storage unit a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence; and
an example sentence pronunciation output unit configured to read from the example sentence pronunciation storage unit the pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering unit, and to vocally output the read pronunciation data.
2. The pronunciation learning device according to claim 1 , further comprising
a word registering unit configured to register the word corresponding to the pronunciation data vocally output by the word pronunciation output unit; and
a first example sentence display unit configured to display a list of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering unit, in response to the registration of the word in the word registering unit.
3. The pronunciation learning device according to claim 1 , further comprising:
a word registering unit configured to register the word corresponding to the pronunciation data vocally output by the word pronunciation output unit;
a word display unit configured to display a list of the words registered in the word registering unit; and
a second example sentence display unit configured to extract, from the example sentence pronunciation storage unit, pronunciation-associated example sentences including the pronunciation data of the word specified and selected by the user from the words displayed as the list by the word display unit, and to display a list of the extracted pronunciation-associated example sentences.
4. The pronunciation learning device according to claim 2 , further comprising:
a word registering unit configured to register the word corresponding to the pronunciation data vocally output by the word pronunciation output unit;
a word display unit configured to display a list of the words registered in the word registering unit; and
a second example sentence display unit configured to extract, from the example sentence pronunciation storage unit, pronunciation-associated example sentences including the pronunciation data of the word specified and selected by the user from the words displayed as the list by the word display unit, and to display a list of the extracted pronunciation-associated example sentences.
5. The pronunciation learning device according to claim 1 , further comprising a third example sentence display unit configured to, when there is a word wherein the pronunciation data output from the example sentence pronunciation output unit is different from the pronunciation data output from the word pronunciation output unit, display a pronunciation-associated example sentence corresponding to the word in a state where the word is identifiable among other words.
6. The pronunciation learning device according to claim 2 , further comprising a third example sentence display unit configured to, there is a word wherein the pronunciation data output from the example sentence pronunciation output unit is different from the pronunciation data output from the word pronunciation output unit, display a pronunciation-associated example sentence corresponding to the word in a state where the word is identifiable among other words.
7. The pronunciation learning device according to claim 3 , further comprising a third example sentence display unit configured to, there is a word wherein the pronunciation data output from the example sentence pronunciation output unit is different from the pronunciation data output from the word pronunciation output unit, display a pronunciation-associated example sentence corresponding to the word in a state where the word is identifiable among other words.
8. The pronunciation learning device according to claim 4 , further comprising a third example sentence display unit configured to, there is a word wherein the pronunciation data output from the example sentence pronunciation output unit is different from the pronunciation data output from the word pronunciation output unit, display a pronunciation-associated example sentence corresponding to the word in a state where the word is identifiable among other words.
9. The pronunciation learning device according to claim 2 , wherein there is a word registered in advance in the word registering unit.
10. A pronunciation learning device which outputs pronunciation data by transmitting and receiving necessary data to and from an external server configured to store pronunciation data of a word, an example sentence text including a plurality of words, and a pronunciation-associated example sentence obtained by associating the pronunciation data with the example sentence text, the pronunciation learning device comprising:
a word pronunciation output unit configured to obtain pronunciation data of a word specified by a user's operation, from the external server, and to vocally output the obtained pronunciation data;
a pronunciation-associated example sentence registering unit configured to extract from the external server a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence in the external server; and
an example sentence pronunciation output unit configured to read from the external server the pronunciation data of any one of the pronunciation-associated example sentences registered in the external server, and to vocally output the pronunciation data.
11. A program for controlling a computer of an electronic device, the program causing the computer to function as:
an example sentence text storage unit configured to store a plurality of example sentence texts each of which includes a plurality of words;
an example sentence pronunciation storage unit configured to associate and store each of the example sentence texts stored in the example sentence text storage unit with pronunciation data as a pronunciation-associated example sentence;
a word pronunciation output unit configured to vocally output pronunciation data of a word specified by a user's operation;
a pronunciation-associated example sentence registering unit configured to extract from the example sentence pronunciation storage unit a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence; and
an example sentence pronunciation output unit configured to read from the example sentence pronunciation storage unit the pronunciation data of any one of the pronunciation-associated example sentences registered in the pronunciation-associated example sentence registering unit, and to vocally output the pronunciation data.
12. A program for controlling a computer of an electronic device configured to output pronunciation data by transmitting and receiving necessary data to and from an external server configured to store pronunciation data of a word, an example sentence text including a plurality of words, and a pronunciation-associated example sentence obtained by associating the pronunciation data with the example sentence text, the program causing the computer to function as:
a word pronunciation output unit configured to obtain pronunciation data of a word specified by a user's operation, from the external server, and to vocally output the obtained pronunciation data;
a pronunciation-associated example sentence registering unit configured to extract from the external server a pronunciation-associated example sentence including the pronunciation data of the word, and to register the extracted pronunciation-associated example sentence in the external server; and
an example sentence pronunciation output unit configured to read from the external server the pronunciation data of any one of the pronunciation-associated example sentences registered in the external server, and to vocally output the read pronunciation data.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014188131A JP6535998B2 (en) | 2014-09-16 | 2014-09-16 | Voice learning device and control program |
JP2014-188131 | 2014-09-16 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160180741A1 true US20160180741A1 (en) | 2016-06-23 |
Family
ID=55505859
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/841,565 Abandoned US20160180741A1 (en) | 2014-09-16 | 2015-08-31 | Pronunciation learning device, pronunciation learning method and recording medium storing control program for pronunciation learning |
Country Status (3)
Country | Link |
---|---|
US (1) | US20160180741A1 (en) |
JP (1) | JP6535998B2 (en) |
CN (1) | CN105427686A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111475708A (en) * | 2019-01-24 | 2020-07-31 | 上海流利说信息技术有限公司 | Push method, medium, device and computing equipment for follow-up reading content |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105825730B (en) * | 2016-06-06 | 2018-08-28 | 袁亚蒙 | The bilingual play system of foreign language learning |
CN107967293B (en) * | 2016-10-20 | 2021-09-28 | 卡西欧计算机株式会社 | Learning support device, learning support method, and recording medium |
JP6957994B2 (en) * | 2017-06-05 | 2021-11-02 | カシオ計算機株式会社 | Audio output control device, audio output control method and program |
CN107240311A (en) * | 2017-08-10 | 2017-10-10 | 李霞芬 | Intelligent english teaching aid |
CN107230400A (en) * | 2017-08-10 | 2017-10-03 | 李霞芬 | Enforce one's memory English teaching aid |
CN110472254A (en) * | 2019-08-16 | 2019-11-19 | 深圳传音控股股份有限公司 | Voice translation method, communication terminal and computer readable storage medium |
JP7131518B2 (en) * | 2019-09-20 | 2022-09-06 | カシオ計算機株式会社 | Electronic device, pronunciation learning method, server device, pronunciation learning processing system and program |
JP7353130B2 (en) * | 2019-10-24 | 2023-09-29 | 東京瓦斯株式会社 | Audio playback systems and programs |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4615680A (en) * | 1983-05-20 | 1986-10-07 | Tomatis Alfred A A | Apparatus and method for practicing pronunciation of words by comparing the user's pronunciation with the stored pronunciation |
US5885081A (en) * | 1994-12-02 | 1999-03-23 | Nec Corporation | System and method for conversion between linguistically significant symbol sequences with display of support information |
US20020051955A1 (en) * | 2000-03-31 | 2002-05-02 | Yasuo Okutani | Speech signal processing apparatus and method, and storage medium |
US20040219497A1 (en) * | 2003-04-29 | 2004-11-04 | Say-Ling Wen | Typewriting sentence learning system and method with hint profolio |
US20050158696A1 (en) * | 2004-01-20 | 2005-07-21 | Jia-Lin Shen | [interactive computer-assisted language learning method and system thereof] |
US20060247920A1 (en) * | 2005-04-28 | 2006-11-02 | Casio Computer Co., Ltd. | Speech output control device and recording medium recorded with speech output control programs |
US20080172226A1 (en) * | 2007-01-11 | 2008-07-17 | Casio Computer Co., Ltd. | Voice output device and voice output program |
US20120060093A1 (en) * | 2009-05-13 | 2012-03-08 | Doohan Lee | Multimedia file playing method and multimedia player |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0660063A (en) * | 1992-08-05 | 1994-03-04 | Matsushita Electric Ind Co Ltd | Chinese character converting device |
JPH0877176A (en) * | 1994-09-07 | 1996-03-22 | Hitachi Ltd | Foreign language translating device |
KR100490367B1 (en) * | 2001-08-03 | 2005-05-17 | 정택 | The portable apparatus of word studying and method of word studying using the same |
CN1396541A (en) * | 2002-05-24 | 2003-02-12 | 北京南山高科技有限公司 | Method and device based on text speech library for inquiring and reproducing phrases |
JP2006267881A (en) * | 2005-03-25 | 2006-10-05 | Sharp Corp | Electronic learning device |
GB2446427A (en) * | 2007-02-07 | 2008-08-13 | Sharp Kk | Computer-implemented learning method and apparatus |
CN101266600A (en) * | 2008-05-07 | 2008-09-17 | 陈光火 | Multimedia multi- language interactive synchronous translation method |
JP4983943B2 (en) * | 2010-03-05 | 2012-07-25 | カシオ計算機株式会社 | Text display device and program |
JP5664978B2 (en) * | 2011-08-22 | 2015-02-04 | 日立コンシューマエレクトロニクス株式会社 | Learning support system and learning support method |
US20140170612A1 (en) * | 2012-12-13 | 2014-06-19 | Vladimir Yudavin | Computer Program Method for Teaching Languages which Includes an Algorithm for Generating Sentences and Texts |
-
2014
- 2014-09-16 JP JP2014188131A patent/JP6535998B2/en active Active
-
2015
- 2015-08-31 US US14/841,565 patent/US20160180741A1/en not_active Abandoned
- 2015-09-16 CN CN201510589695.6A patent/CN105427686A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4615680A (en) * | 1983-05-20 | 1986-10-07 | Tomatis Alfred A A | Apparatus and method for practicing pronunciation of words by comparing the user's pronunciation with the stored pronunciation |
US5885081A (en) * | 1994-12-02 | 1999-03-23 | Nec Corporation | System and method for conversion between linguistically significant symbol sequences with display of support information |
US20020051955A1 (en) * | 2000-03-31 | 2002-05-02 | Yasuo Okutani | Speech signal processing apparatus and method, and storage medium |
US20040219497A1 (en) * | 2003-04-29 | 2004-11-04 | Say-Ling Wen | Typewriting sentence learning system and method with hint profolio |
US20050158696A1 (en) * | 2004-01-20 | 2005-07-21 | Jia-Lin Shen | [interactive computer-assisted language learning method and system thereof] |
US20060247920A1 (en) * | 2005-04-28 | 2006-11-02 | Casio Computer Co., Ltd. | Speech output control device and recording medium recorded with speech output control programs |
US20080172226A1 (en) * | 2007-01-11 | 2008-07-17 | Casio Computer Co., Ltd. | Voice output device and voice output program |
US20120060093A1 (en) * | 2009-05-13 | 2012-03-08 | Doohan Lee | Multimedia file playing method and multimedia player |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111475708A (en) * | 2019-01-24 | 2020-07-31 | 上海流利说信息技术有限公司 | Push method, medium, device and computing equipment for follow-up reading content |
Also Published As
Publication number | Publication date |
---|---|
JP2016061855A (en) | 2016-04-25 |
JP6535998B2 (en) | 2019-07-03 |
CN105427686A (en) | 2016-03-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160180741A1 (en) | Pronunciation learning device, pronunciation learning method and recording medium storing control program for pronunciation learning | |
US10210154B2 (en) | Input method editor having a secondary language mode | |
US20150169537A1 (en) | Using statistical language models to improve text input | |
KR20120026468A (en) | Electronic illustrated dictionary device, illustrated dictionary display method, and storage medium storing program for performing illustrated dictionary display control | |
JP2010198241A (en) | Chinese input device and program | |
JP7477006B2 (en) | Electronic dictionary device, search support method and program | |
JP2008186376A (en) | Voice output device and voice output program | |
US8489389B2 (en) | Electronic apparatus with dictionary function and computer-readable medium | |
JP5810814B2 (en) | Electronic device having dictionary function, compound word search method, and program | |
JP6676093B2 (en) | Interlingual communication support device and system | |
KR20100024566A (en) | Input apparatus and method for the korean alphabet for handy terminal | |
JP5673215B2 (en) | Russian language search device and program | |
JP5472378B2 (en) | Mobile devices and programs | |
JP2012168696A (en) | Dictionary information display device and program | |
US20140081622A1 (en) | Information display control apparatus, information display control method, information display control system, and recording medium on which information display control program is recorded | |
CN112541071A (en) | Electronic dictionary, learning word judgment method, and recording medium | |
JP2010282507A (en) | Electronic apparatus including dictionary function, and program | |
JP2005292303A (en) | Information display controller, information display control processing program, and information management server | |
JP5338252B2 (en) | Electronic device with dictionary function | |
JP2008299431A (en) | Handwritten character input device and control program therefor | |
JP2023046232A (en) | Electronic equipment, learning support system, learning processing method, and program | |
JP6417754B2 (en) | Combination word registration device and program | |
JP6446801B2 (en) | Display control apparatus and program | |
JP6451153B2 (en) | Information display control device and program | |
JP2015166905A (en) | Electronic apparatus with dictionary display function, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CASIO COMPUTER CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAMAMOTO, ATSUSHI;REEL/FRAME:036462/0685 Effective date: 20150831 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |