CN115132007A - Server device, learning support method, and recording medium - Google Patents

Server device, learning support method, and recording medium Download PDF

Info

Publication number
CN115132007A
CN115132007A CN202210249179.9A CN202210249179A CN115132007A CN 115132007 A CN115132007 A CN 115132007A CN 202210249179 A CN202210249179 A CN 202210249179A CN 115132007 A CN115132007 A CN 115132007A
Authority
CN
China
Prior art keywords
word
related image
user
control unit
wes
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210249179.9A
Other languages
Chinese (zh)
Inventor
麻生麻实
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Casio Computer Co Ltd
Original Assignee
Casio Computer Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Casio Computer Co Ltd filed Critical Casio Computer Co Ltd
Publication of CN115132007A publication Critical patent/CN115132007A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/08Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations
    • G09B5/12Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations different stations being capable of presenting different information simultaneously
    • G09B5/125Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations different stations being capable of presenting different information simultaneously the stations being mobile
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • G06F16/535Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • G06F16/538Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K17/00Methods or arrangements for effecting co-operative working between equipments covered by two or more of main groups G06K1/00 - G06K15/00, e.g. automatic card files incorporating conveying and reading operations
    • G06K17/0022Methods or arrangements for effecting co-operative working between equipments covered by two or more of main groups G06K1/00 - G06K15/00, e.g. automatic card files incorporating conveying and reading operations arrangements or provisious for transferring data to distant stations, e.g. from a sensing device
    • G06K17/0025Methods or arrangements for effecting co-operative working between equipments covered by two or more of main groups G06K1/00 - G06K15/00, e.g. automatic card files incorporating conveying and reading operations arrangements or provisious for transferring data to distant stations, e.g. from a sensing device the arrangement consisting of a wireless interrogation device in combination with a device for optically marking the record carrier
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/06Foreign languages

Abstract

The invention provides a server device, a learning support method and a recording medium. The WES server (30) generates a word card for each word in the search history, which acquires the word, explanatory information including word translation, and related images from dictionary data in a dictionary Database (DB) and associates the word with the dictionary data, automatically generates a user-specific word book, and stores the word book in the word book Database (DB) in association with a dictionary < user > ID. The user acquires the generated word book at the user terminal 20, and displays each word card of the word book while switching among the english word page (TCa), the related image page (TCb), and the word translation page (TCc). When a word of a word card is designated, a list of related images of word cards of other users including the designated word is received from the WES server (30), and when any related image is designated, the related image of the word card is updated to the related image of the designated other user.

Description

Server device, learning support method, and recording medium
Technical Field
The present invention relates to a server device, a learning support method, and a recording medium for effectively supporting language learning.
Background
In the past, there has been a word book as a prop to learn, for example, english words. In this example, a word card is prepared by binding one word card, in which an english word is described on the front surface and a word translation is described on the back surface, for each word. After the user answers his/her word translation with the english word on the word card, he/she turns over the word card to determine whether the answer word translation is a correct answer, thereby learning the english word.
In recent years, electronic dictionaries have been widely used, and when a user searches for an arbitrary word from dictionary data and displays explanatory information including the word and its word translation, the electronic dictionary stores a history of the searched word as a search history. Further, the system has a function of a word book, and when a user designates an arbitrary word from a list of words stored as a search history, the system operates so as to read and display explanatory information of the designated word from dictionary data (see, for example, japanese patent laid-open No. 2005-292303).
Further, the functions of: in a word learning card capable of improving the learning effect of words, a learning target word is displayed together with patterns conceptually associated with the word, thereby assisting memory (see, for example, japanese patent laid-open No. 2003-241631).
However, since it is relatively more impressive for the user to select a picture or a pattern associated with a word by himself/herself, it is desirable to perform word learning more effectively by switching to a picture or a pattern which is selected from the WEB and is easy to remember.
In learning words using a word book, it is not limited to pictures and patterns associated with words selected by a user, but it is also desirable to perform substitution or addition more efficiently and to perform word learning more quickly and efficiently.
Disclosure of Invention
The invention aims to provide a server device, a learning support method and a recording medium which can effectively and efficiently memorize the language of a learning object for learning.
The server device according to the present invention includes a communication unit; a control unit; and a storage unit configured to store word books of a plurality of users in association with each other for each of the plurality of users, wherein the control unit, upon receiving a specified word or phrase specified by a user via the communication unit, extracts a related image related to the specified word or phrase from a word book other than the user stored in the storage unit, and transmits the related image to the communication device of the user via the communication unit.
Drawings
Fig. 1 is a diagram showing the overall configuration of a learning support system 1 according to an embodiment of a server device, a learning support method, and a recording medium of the present invention.
Fig. 2 is a block diagram showing the configuration of an electronic circuit of the electronic dictionary (electronic device) 10.
Fig. 3 is a diagram showing the contents of the search history data stored in the search history data storage area 12d of the electronic dictionary 10.
Fig. 4 is a block diagram showing the configuration of an electronic circuit of the user terminal (communication device) 20.
Fig. 5 is a block diagram showing a configuration of an electronic circuit of the WES server (server device) 30.
Fig. 6 is a diagram showing the contents of dictionary data stored in the dictionary database storage area 32c of the WES server 30.
Fig. 7 is a diagram showing the contents of the user management data stored in the user management database storage area 32d of the WES server 30.
Fig. 8 is a diagram showing the contents of the word book data stored in the word book database storage area 32e of the WES server 30.
Fig. 9 is a flowchart showing the WES server synchronization process of the electronic dictionary 10.
Fig. 10 is a flowchart showing a WES utilization process of the user terminal 20.
Fig. 11 is a flowchart showing the wordbook creation process of the WES server 30.
Fig. 12 is a flowchart showing the wordbook utilization process of the user terminal 20.
Fig. 13 is a flowchart showing (one of) the related image update processing of the user terminal 20.
Fig. 14 is a flowchart showing the related image update processing (second process) of the user terminal 20.
Fig. 15 is a flowchart showing (one of) the wordbook processing of the WES server 30.
Fig. 16 is a flowchart showing the wordbook processing (second) of the WES server 30.
Fig. 17 is a flowchart showing a Web related image list generation process included in the wordbook process of the WES server 30.
Fig. 18 is a flowchart showing a WES related image list generation process included in the wordbook process of the WES server 30.
Fig. 19 is a diagram showing a menu screen GW for WES use, which follows the WES use process of the user terminal 20.
Fig. 20 is a diagram showing an english word page TCa corresponding to the word card "archive" with the display of the vocabulary following the vocabulary use process of the user terminal 20.
Fig. 21 is a diagram showing a switching state of the english word page TCa, the related image page TCb, and the word translation page TCc corresponding to the word card "archive" accompanying the display of the word book of the user terminal 20.
Fig. 22 is a diagram showing an example of display of the Web related image list RL1 and the WES related image list RL2 corresponding to the word card "archive" following the related image update processing of the user terminal 20.
Fig. 23 is a diagram showing an operation of non-display setting of the irrelevant related image RaN in accordance with the Web related image list generation processing of the WES server 30.
Fig. 24 is a diagram showing an example of division display of the Web related image list RL1 and the Web related image list SLa corresponding to the word card "road" following the related image update process of the user terminal 20.
Fig. 25 is a diagram showing an example of operation of filtering related images in the display of the we related image list RL2 in accordance with the related image update process of the user terminal 20.
Fig. 26 is a diagram showing a display state of an english word page TCa, a related image page TCb, and a word translation page TCc of another user, which are formed by specifying an arbitrary we-related image Rb2 from the we-related image list RL2 corresponding to the word card "archive" in fig. 25.
Detailed Description
Embodiments of the present invention are described below with reference to the drawings.
(structure of the embodiment)
Fig. 1 is a diagram showing the overall configuration of a learning support system 1 according to an embodiment of a server device, a learning support method, and a recording medium of the present invention.
The learning support system 1 includes an electronic dictionary (electronic device) 10, a user terminal (communication device) 20, and a WES (world Education Service) server (server device) 30 provided on a communication network N such as the internet. Here, the WES (world Education Service) server means a dedicated server that connects an electronic dictionary, an electronic dictionary application, and a cloud Service.
The electronic dictionary (electronic device) 10 may be configured as a smartphone, a tablet terminal, a PC (personal computer), a mobile phone, an electronic book, a portable game machine, or the like having a dictionary function, in addition to the dedicated electronic dictionary 10 described below.
The user terminal (communication device) 20 can be configured as a smartphone, a tablet terminal, an electronic dictionary, a PC, a mobile phone, an electronic book, a portable game machine, or the like having a communication function.
The user terminal 20 is provided with an application program (WES use application here) using the WES server 30.
The learning support system 1 shown in fig. 1 has at least the following functions (10a), (20b), (20c), (20e), (30a), (30b), and (30 c).
In the electronic dictionary 10, electronic dictionary information including a dictionary ID (dictionary identification information, which may also be used as user identification information: dictionary < user > ID) and a search history of a word is encoded into a two-dimensional code QR and displayed (function (10 a)).
The user terminal 20 reads the two-dimensional code QR displayed on the electronic dictionary 10 in accordance with the WES use application, decodes the two-dimensional code QR into electronic dictionary information (including the dictionary ID and the search history) (function (20a)), and transmits data obtained by adding user information including user identification information to the decoded electronic dictionary information to the WES server 30 (function (20 b)).
The WES server 30 generates a word book in which the word translation read from the dictionary data and explanatory information thereof are associated with the associated image of the word for each word in the search history based on the electronic dictionary information (including the dictionary ID and the search history) received from the user terminal 20, and stores the generated word book in association with the dictionary < user > ID (function (30 a)).
The user terminal 20 receives and displays the data of the wordbook stored in the WES server 30 and corresponding to the dictionary < user > ID of the user terminal (function (20c)), specifies the related image of an arbitrary word included in the displayed wordbook, instructs the WES server 30 to update the related image (function (20d)), adds arbitrary memo data to the explanatory information of the arbitrary word included in the displayed wordbook, and instructs the WES server 30 to update the wordbook (function (20 e)).
The WES server 30 updates the related image or (and) the additional word book including the memo data in accordance with an instruction to update the related image of the specified word from the user terminal 20 or (and) an instruction to update the word book including the memo data in the description information of the specified word (function (30b)), and updates the additional word book including the similar meaning word and the related image thereof (similar meaning word related image) when the similar meaning word is present in the specified word at the time of updating the related image (function (30 c)).
According to the learning support system 1, it is possible to automatically generate a word book corresponding to the user based on the search history of words obtained from the electronic dictionary 10. Further, since the word book is configured by associating not only the words and the explanatory information thereof but also the related images of the words and/or the arbitrary memo data, the user can easily understand the meaning of the words intuitively by looking at the related images of the words, or easily memorize the words intuitively by looking at the consonant of the pronunciation of the words added as the memo data, for example, and can effectively and efficiently memorize the words to learn.
When the electronic dictionary 10 and the user terminal 20 have a short-range wireless communication function such as Bluetooth (registered trademark), the electronic dictionary information (including the dictionary ID and the search history) of the electronic dictionary 10 is transmitted to the user terminal 20 by short-range wireless communication.
In addition, when the electronic dictionary 10 has a communication function with the communication network N such as Wi-Fi (registered trademark), a WES-use application may be installed in the electronic dictionary 10, so that the electronic dictionary 10 may have the above-described functions (20a), (20b), (20c), and (20e) of the user terminal 20 in a lump.
When the user terminal 20 has a dictionary function, the user terminal 20 itself may generate and store a search history of words. When the user terminal 20 does not have the dictionary function and searches dictionary data by using the dictionary function of the WES server 30, the WES server 30 may generate and store a search history of words.
Fig. 2 is a block diagram showing the configuration of an electronic circuit of the electronic dictionary (electronic device) 10.
The electronic dictionary (electronic device) 10 includes a computer control Unit (CPU) 11.
The control unit 11 controls the operations of the circuit units in accordance with a dictionary control program (12a) stored in the storage unit (memory) 12. The dictionary control program (12a) may be any one of: the program is stored in the storage unit 12 in advance, is read from an external recording medium (storage)13 such as a memory card by the recording medium reading unit 14, and is stored in the storage unit 12, or is downloaded from a Web server (here, a program server) 40 on the communication network N via the communication unit 15, and is stored in the storage unit 12.
The dictionary control program (12a) includes, in addition to a system program for controlling the whole of the electronic dictionary 10, a program for searching for dictionary data (12c) in accordance with the operation of a user, a program for generating search history data (12d), and a program for executing the above-described function (10 a).
The control unit 11 is connected to a key input unit 16 including a character input key and a [ translate/determine ] key, a touch panel display unit (touch panel type display)17, and the like, in addition to the storage unit 12, the recording medium reading unit 14, and the communication unit 15, via a system and a data bus.
In addition to a program storage area 12a for storing a dictionary control program (12a), a dictionary ID storage area 12b for storing a dictionary ID (dictionary identification information) (12b), a dictionary data storage area 12c for storing dictionary data (12c), a search history data storage area 12d for storing search history data (12d), and the like are secured in the storage unit 12.
The dictionary ID storage area 12b stores dictionary identification information unique to the electronic dictionary 10 as a dictionary ID. Note that the dictionary ID may also serve as user identification information (in this case, "dictionary < user > ID").
The dictionary data storage area 12c stores various dictionary data such as english-japanese dictionary, japanese-english dictionary, and japanese dictionary as dictionary data associating words as terms and explanatory information including word translations, example sentences, and example sentence translations corresponding to the terms.
Fig. 3 is a diagram showing the contents of the search history data stored in the search history data storage area 12d of the electronic dictionary 10.
As shown in fig. 3, the entry (word) of the user search dictionary data (12c) is stored in the search history data storage area 12d in association with the search date and time for each type of dictionary data (12 c). Fig. 3 is an example of search history data of the english-japanese dictionary.
The electronic dictionary 10 thus configured realizes various functions as described in the operation explanation below by the control unit 11 operating in cooperation with software and hardware by controlling the operations of each part of the circuit in accordance with the commands described in the dictionary control program (12 a).
Fig. 4 is a block diagram showing the configuration of an electronic circuit of the user terminal (communication device) 20.
The user terminal (communication device) 20 includes a control unit (CPU)21 as a computer.
The control unit 21 controls the operations of the circuit units in accordance with a terminal control program (22a) and a WES application (WES application) (22b) stored in the storage unit 22. The terminal control program (22a) may be any one of: the program is stored in advance in the storage unit 22, is read from an external recording medium 23 such as a memory card by the recording medium reading unit 24, and is stored in the storage unit 22, or is downloaded from a Web server (here, a program server) 40 on the communication network N via the communication unit 25, and is stored in the storage unit 22.
The WES-use application (22b) is downloaded from a Web server (here, a program server of an application store) 40 on the communication network N via the communication unit 25 and stored in the storage unit 22.
The terminal control program (22a) includes, in addition to a system program for controlling the entire user terminal 20: a program for performing communication connection with a communication device outside the WES server 30 and the Web server 40 on the communication network N at any time in cooperation with various application programs stored in the storage unit 22.
The WES-use application (22b) includes a program for executing the aforementioned functions (20a), (20b), (20c), and (20 e).
The control unit 21 is connected to the storage unit 22, the recording medium reading unit 24, and the communication unit 25 via a system and a data bus, and also connected to a key input unit 26 including a power key and a volume adjustment key, a touch panel display unit 27, an image pickup unit 28, and the like.
In addition to the program storage areas 22a and 22b for storing the terminal control program (22a) and the WES use application (22b), the storage unit 22 also stores a terminal data storage area 22c for storing terminal data (22c), a dictionary information storage area 22d for storing dictionary information (22d), and the like.
The terminal data storage area 22c stores, as terminal data unique to the user terminal 20, data such as a telephone number and a mail address in addition to a terminal number for communication connection with an external communication device including the WES server 30 and the Web server 40 on the communication network N.
The dictionary information storage area 22d stores, as dictionary information, data including a dictionary < user > ID obtained by QR decoding the two-dimensional code read from the electronic dictionary 10 and a search history.
The user terminal 20 configured as described above operates by the control unit 21 in cooperation with software and hardware in accordance with the operations of each unit of the circuit controlled by commands described in the terminal control program (22a) and the WES use application (22b), and realizes various functions as described in the operation description below.
Fig. 5 is a block diagram showing the configuration of an electronic circuit of the WES server (server device) 30.
The WES server (server device) 30 includes a control unit (CPU)31 as a computer.
The control unit 31 controls the operations of the circuit units in accordance with a server control program (32a) and an image recognition program (including an AI learning function) (32b) stored in the storage unit 32.
The server control program (32a) and (or) the image recognition program (32b) may be either: stored in advance in the storage unit 32; or read from an external recording medium 33 such as a CD-ROM by a recording medium reading unit 34 and stored in the storage unit 32; or downloaded from a Web server (here, a program server) 40 on the communication network N via the communication section 35 and stored in the storage section 32.
The server control program (32a) includes a program for performing communication connection with an external communication device including the user terminal 20 as needed, and a program for executing the above-described functions (30a), (30b), and (30c) in cooperation with the image recognition program (32b) as needed, in addition to a system program for controlling the entire WES server 30.
The image recognizing program (32b) includes, for example, a program for: whether or not the image is a related image having a meaning of a word is recognized for 1 or more related images of the word obtained by performing a Web search (search by matching the text of the word, recognizing the image, or the like from a Web server 40 on a communication network N) in correspondence with the entry (word) of dictionary data located in a dictionary database (32 c).
The image recognition program (32b) has, for example, a feature image database storing, for each entry (word) included in dictionary data, a feature image having a high semantic relation with the word, and recognizes a related image having a high degree of relation with the word into a probability of "0" to "1" depending on how much the feature images of the word match with each other, for example, a probability of "0.8" or more, as a related image having a semantic relation with the word, and a probability of "0.7" or less, as a related image having no relation with the semantic relation with the word.
The control unit 31 is connected to a pointing device, an input unit 36 including a keyboard, a display unit 37, and the like, in addition to the storage unit 32, the recording medium reading unit 34, and the communication unit 35, via a system and a data bus.
The storage unit 32 has a dictionary database storage area 32c, a user management database storage area 32d, a word book database storage area 32e, a work data storage area 32f, and the like, in addition to program storage areas 32a and 32b for storing a server control program (32a) and an image recognition program (32 b).
Fig. 6 is a diagram showing the contents of dictionary data stored in the dictionary database storage area 32c of the WES server 30.
In the dictionary database storage area 32c, entries (words), word translations including the words, explanatory information of example sentences and example sentence translations, related images of the words, and similar words of the words are stored in association with each other in advance for each type of dictionary data.
Fig. 7 is a diagram showing the contents of the user management data stored in the user management database storage area 32d of the WES server 30.
In the user management database storage area 32d, the following information of a user who downloads a WES application (22b) and uses the WES application is stored in association with: a dictionary < user > ID of the electronic dictionary 10, a terminal device number of the user terminal 20; user attributes such as the occupation of the user (for example, school and school year in the case of a student, and the kind of occupation in the case of a person who has already attended work); features of images that the user likes as related images of words (related image features: pictorial images, photographic images, harmonic images of words, etc.); and a search history of the electronic dictionary 10 of the user.
The dictionary < user > ID, terminal device number, user attribute, and related image feature of the user management database storage area 32d can be input from the user terminal 20 and stored, for example, along with initial setting of the user terminal 20 for the we use application (22 b). The user uses the user terminal 20 at any time and transmits the result to the WES server 30, thereby storing the search history of the electronic dictionary 10.
The related image feature may be input from the user terminal 20 by the user himself or may be stored, or the feature of the related image included in the word book data of the user stored in the word book database storage area 32e may be acquired by the image recognition processing and stored.
Fig. 8 is a diagram showing the contents of the word book data stored in the word book database storage area 32e of the WES server 30.
In the word book database storage area 32e, words located in the search history, explanatory information (word translation, example sentence translation) of the words, related images (including update date and time, sharable/not) of the words, memo data (including sharable/not) inputted by the user in association with the words, similar words of the words, and related images (similar word related images) of the similar words are stored in association with each other as word book data for each dictionary < user > ID of each user stored in the user management database storage area 32 d.
The job data storage area 32f temporarily stores various data generated or acquired in accordance with the control of the operation of each unit by the control unit 31, as necessary.
The WES server 30 configured as described above controls the operations of each circuit unit by the control unit 31 in accordance with commands described in the server control program (32a) and the image recognition program (32b), and realizes various functions described in the operation description below by operating software and hardware in cooperation with each other.
(operation of the embodiment)
Next, the operation of the learning support system 1 according to the embodiment will be described.
< Process of synchronizing search History of electronic dictionary 10 with WES Server 30 (FIGS. 9 and 10) >
Fig. 9 is a flowchart showing the WES server synchronization process of the electronic dictionary 10.
Fig. 10 is a flowchart showing a WES utilization process of the user terminal 20.
Fig. 19 is a diagram showing a menu screen GW for WES use, which follows the WES use process of the user terminal 20.
When the user terminal 20 starts the WES use application (22b) (yes in step T1), the control unit 21 displays the menu screen GW for use of the WES on the touch panel display unit 20 as shown in fig. 19.
On the menu screen GW for WES use, for example, an item "synchronize dictionary and WES server" M1 for sending the search history data (12d) of the electronic dictionary 10 to the WES server 30 and synchronizing them, and an item "use word book" M2 for using the word book generated by the WES server 30 are displayed.
When the user designates an item "synchronize dictionary and WES server" M1 on the menu screen GW (yes at step T2), the control unit 21 causes the touch panel display unit 27 to display guidance for acquiring the search history data (12d) from the electronic dictionary 10 as, for example, "please read the two-dimensional code of the search history from the electronic dictionary".
In the electronic dictionary 10, when an item for synchronizing the electronic dictionary 10 and the WES server 30 is specified by a user on a menu screen displayed on the touch panel display unit 17 in response to, for example, a pressing operation of a [ menu ] key of the key input unit 16 (yes in step D1), the control unit 11 encodes the dictionary ID (dictionary < user > ID) stored in the dictionary ID storage area 12b and the search history data stored in the search history data storage area 12D, and generates the two-dimensional code QR (step D2). The control unit 21 causes the touch panel display unit 17 to display the generated two-dimensional code QR (step D3).
When the two-dimensional code QR displayed on the electronic dictionary 10 is imaged by the imaging unit 28 in the user terminal 20 (step T3), the control unit 21 decodes the two-dimensional code QR and transmits the decoded dictionary < user > ID and the search history data to the WES server 30 (step T4).
When the screen displaying the two-dimensional code QR is closed in the electronic dictionary 10 (yes in step D4), the control unit 11 ends the WES server synchronization process in the electronic dictionary 10.
When the user terminal 20 closes the screen on which the two-dimensional code QR is photographed (yes in step T5), the control unit 21 ends the WES use process in the user terminal 20.
< processing for generating a word book by the WES server 30 (fig. 11) >
Fig. 11 is a flowchart showing the wordbook creation process of the WES server 30.
Upon receiving the dictionary < user > ID and the search history data transmitted from the user terminal 20 (yes in step S1), the control unit 31 stores the received search history data in the user management database storage area 32d in association with the received dictionary < user > ID (see fig. 7), and sequentially acquires the words of the search history included in the search history data from the beginning (step S2).
The control unit 31 acquires explanatory information and related images of word translations, example sentences and example sentence translations including the word from dictionary data (for example, english word, english dictionary: refer to fig. 6) corresponding to the acquired word stored in the dictionary database storage area 32c, and generates a word card in which the related image and explanatory information are associated with the word (step S3).
The control unit 31 determines whether or not all words included in the search history data have been acquired (step S4), and if it is determined that there are any words not acquired (no in step S4), acquires the next word from the search history data and generates a word card (steps S2 and S3).
When it is determined that all words included in the search history data have been acquired and that word cards corresponding to all the words have been generated (yes at step S4), the control unit 31 stores the data of each generated word card in the word book database storage area 32e (see fig. 8) in association with the received dictionary < user > ID as a word book (step S5).
This makes it possible to automatically generate a word book in which explanatory information of a word and a related image are associated with each other, for all words in a search history of the word unique to a user.
Note that the word card of the word book that has just been automatically generated from the search history data in accordance with the word book generation processing (fig. 11) of the WES server 30 does not include the data of the memo, the similar meaning word, and the similar meaning word-related image.
< processing of Using the word book by the user terminal 20 (FIGS. 10, 12, and 15) >
Following the WES utilization process of the user terminal 20 (fig. 10), as shown in fig. 19, when the user designates the item "use vocabulary" M2 on the menu screen GW of the WES utilization application (22b) displayed on the touch panel display unit 27 (yes at step T6), the control unit 21 shifts to the vocabulary utilization process (fig. 12) (step TA).
Fig. 12 is a flowchart showing the wordbook utilization process of the user terminal 20.
Fig. 15 is a flowchart showing (one of) the wordbook processing of the WES server 30.
Fig. 16 is a flowchart showing the wordbook processing (second) of the WES server 30.
Fig. 20 is a diagram showing an english word page TCa corresponding to the word card "archive" with the display of the vocabulary following the vocabulary use process of the user terminal 20.
Fig. 21 is a diagram showing a switching state of the english word page TCa, the related image page TCb, and the word translation page TCc corresponding to the word card "archive" accompanying the display of the word book of the user terminal 20.
The control unit 21 communicates with the WES server 30 based on the dictionary < user > ID, and transmits a signal requesting a word book (step TA 1).
Upon receiving the signal requesting the vocabulary book transmitted from the user terminal 20 together with the dictionary < user > ID (yes in step S6), the control unit 30 transmits the data of the vocabulary book corresponding to the dictionary < user > ID stored in the vocabulary book database storage area 32e to the user terminal 20 that is the request source of the vocabulary book based on the terminal device number stored in the user management database storage area 32d (step S7).
Upon receiving the data of the wordbook transmitted from the WES server 30 at the user terminal 20 (yes at step TA 2), the control unit 21 causes the touch panel display unit 27 to display an english word page TCa of an arbitrary wordbard (here, "archive" is designated) included in the received wordbook, as shown in fig. 20, for example (step TA 3).
Here, for example, when the screen is slid and the page is turned on the touch panel display 27 displaying the english word page TCa of the designated word, the control unit 21 switches and displays the english word page TCa, the related image page TCb, and the word translation page TCc of the designated word card of the received word book as shown in (a), (B), and (C) of fig. 21 (step TA 3).
An english word "archive" of a specified word card is displayed on the english word page TCa, an associated image (here, a storage of a file) corresponding to the english word "archive" is displayed on the associated image page TCb, and explanatory information including japanese translation Ya, example sentence Re, and example sentence translation Ry and memo data Me corresponding to the english word "archive" are displayed on the word translation page TCc.
The memo data "い part は deposited under the term of japanese-opera-tional preservation で." Me "of the word translation page TCc corresponding to the english word" archive "shown in fig. 21 (C) is added by, for example, inputting the word book as a consonant of the english word" archive "during the use of the word book by the user as described later.
Thus, the user can learn the word by seeing the related image of the word in the word book to make it easy to intuitively understand the meaning of the word, or by seeing the consonant of the pronunciation of the word added as memo data to make it easy to intuitively memorize the word, and can effectively and efficiently memorize the word.
In the display example of the word card of the word book shown in fig. 21, a display example in which the similar meaning word and the similar meaning word related image are not associated with each other is shown.
< processing for updating the word book in accordance with the user terminal 20 (FIGS. 12 to 18) >
Fig. 22 is a diagram showing an example of display of the Web related image list RL1 and the WES related image list RL2 corresponding to the word card "archive" following the related image update processing of the user terminal 20.
For example, as shown in fig. 21B, in a state where the touch panel display unit 27 of the user terminal 20 is caused to display the related image page TCb of the designated word card "archive" (step TA3), when the [ Web ] button BT1 displayed on the margin of the related image page TCb is pressed as shown in fig. 22a (yes) (step TA 4) in order to update the displayed current related image to the related image preferred by the user by searching from the image on the Web, for example, the control unit 21 transmits pressing signals of the designated word "archive" and the [ Web ] button BT1 to the WES server 30 (step TA 5).
Upon receiving the designated word "archive" and the press signal of the [ Web ] button BT1 transmitted from the user terminal 20 (yes in step S8) at the WES server 30, the control unit 31 shifts to the Web-related-image list generation processing (fig. 17) (step SE).
Fig. 17 is a flowchart showing a Web related image list generation process included in the wordbook process of the WES server 30.
The control unit 31 stores the specified word "archive" as an initial value in the search target word list secured in the job data storage area 32f (step SE1), and determines whether or not the search target word "archive" stored in the search target word list has a synonym based on the dictionary data stored in the dictionary database storage area 32c (see fig. 6) (step SE 2).
For example, when it is determined that the search target word "archive" has no similar meaning word (no at step SE2), the control unit 31 performs a search for a predetermined number of pieces (20 pieces in this case) from the Web server 40 on the communication network N, as shown in fig. 22B, and acquires related images Ra1, Ra2, Ra3, which are related to the search target word "archive" (step SE 5).
The control unit 31 takes each related image acquired in step SE5 as a target, and digitalizes the height of the relationship with the search target word "archive" to a probability of "0" to "1" by following the image recognition program (32b), and determines whether or not the probability is "0.7" or less, that is, whether or not the image is a related image having no relationship with the meaning of the search target word (step SE 6).
Fig. 23 is a diagram showing an operation of non-display setting of the irrelevant related image RaN in accordance with the Web related image list generation processing of the WES server 30.
Here, if it is determined that the related image has no relation to the search target word "archive" (yes at step SE6), the related image RaN having no relation is set to be not displayed as shown in fig. 23, for example (step SE 7).
When the related images RaN that do not have a relation with the search target word "archive" are determined and set to be not displayed for each related image (20 pieces in this case) acquired in step SE5 (yes in step SE8), the control unit 31 determines whether or not there is a search target word for which a related image search from the Web server 40 was not performed in the search target word list secured in the job data storage area 32f (step SE 9).
Here, since only the word "archive" specified in the wordbook being displayed on the user terminal 20 is stored as the search target word in the search target word list, it is determined that there is no search target word for which a search from the related image of the Web server 40 was not performed (step SE9 (no)).
The control unit 31 determines whether or not the search target word is 2 or more (step SE10), and determines that the search target word is not 2 or more (step SE10 (no)).
Then, the control section 31 transmits the data of the list of related images Ra1, Ra2, Ra3,. corresponding to the designated word "archive" to the user terminal 20 as the Web related image list data (step SE 11).
When the user terminal 20 receives the Web related image list data corresponding to the designated word "archive" transmitted from the WES server 30 (yes at step TA 6), the control unit 21 proceeds to the related image update processing (fig. 13) (step TB).
Fig. 13 is a flowchart showing (one of) the related image update processing of the user terminal 20.
Fig. 14 is a flowchart showing the related image update processing (second process) of the user terminal 20.
When receiving the Web-related-image list data from the WES server 30 (yes in step TB 1), the control unit 21 determines whether or not the Web-related-image list data includes a similar meaning word of the specified word and a similar meaning word-related image (step TB 2).
Here, when it is determined that the similar meaning word and the similar meaning word related image of the specified word are not included (no in step TB2), the control unit 21 causes the touch panel display unit 27 to display a list of related images Ra1, Ra2, Ra3,. of the specified word "archive" received from the WES server 30 as a Web related image list RL1, as shown in fig. 22B (step TB 3).
When the list of related images Ra1, Ra2, Ra3, and/or.. for the designated word "archive" includes a related image RaN unrelated to the designated word "archive" and set to be not displayed (step SE7) in accordance with the Web-related image list generation processing (fig. 17) in the WES server 30, the control unit 21 causes the touch panel type display unit 27 to display a Web-related image list RL1 in which the association flag Nd is added to the related image RaN set to be not displayed, as shown in fig. 23 (step TB 3). The related image RaN set to be not displayed may be set to be not displayed.
Here, for example, when the related image Ra3 preferred by the user H is specified in the Web related image list RL1 corresponding to the specified word "archive" displayed as shown in fig. 23 (yes in step TB 4), the control unit 21 causes the specified frame W to be displayed in the specified related image Ra3, and transmits the data of the specified related image Ra3 to the WES server 30 (step TB 5).
In the Web related image list RL1 displayed on the user terminal 20, when the related image Ra3 designated by the user H can share the wordbook given to another user, the [ sharable ] button BCy displayed together with the Web related image list RL1 is pressed to designate the related image Ra3, and when sharing is not desired, the [ unshared ] button BCn is pressed to designate the related image Ra 3.
Thus, the data of the related image Ra3 specified from the Web related image list RL1 in the user terminal 20 is transmitted to the WES server 30 together with sharable or non-sharable information (step TB 5). The data of the designated related image Ra3 transmitted to the WES server 30 may be identification information (such as an index) for identifying the image data, instead of the image data itself.
Upon receiving the data (including sharable/non-sharable information) of the designated related image Ra3 transmitted from the user terminal 20 (yes in step S9), the control unit 31 transmits the related image corresponding to the word card "archive" of the corresponding word book in the word book database storage area 32e (see fig. 8) as shown in fig. 22a to the user terminal 20 and causes the related image to be displayed, including the update date and time and sharable/non-sharable information, to be updated to the received related image Ra3 (step S10).
Thus, the user can specify the related image Ra3 preferred by the user H from the Web related image list RL1 retrieved from the Web server 40 on the communication network N, and update the related image Ra 3526 to the specified related image Ra3, with respect to the related image generated by the WES server 30 and associated with an arbitrary word of the own word book.
< Process of adding a similar meaning term and a similar meaning term related image to a word book and updating >
In the wordbook utilization process (fig. 12) of the user terminal 20, for example, in a state where the related image page TCb (refer to fig. 22a) displaying the related image corresponding to the designated word "road" is displayed (step TA3), when the [ Web ] button BT1 is pressed to transmit the pressing signals of the designated word "road" and the [ Web ] button BT1 to the WES server 30 (steps TA4 and TA5), the wordbook processing of the WES server 30 executes the Web related image generation process including the synonym "street" of the designated word "road" and the synonym related image as follows (step S8 yes → SE).
Fig. 24 is a diagram showing an example of the division display of the Web related image list RL1 and the Web related image list Sla corresponding to the word card "road" in accordance with the related image update processing of the user terminal 20.
The control unit 31 stores the specified word "road" as an initial value in the search target word list secured in the job data storage area 32f (step SE1), and determines whether or not the search target word "road" stored in the search target word list has a synonym based on the dictionary data stored in the dictionary database storage area 32c (see fig. 6) (step SE 2).
For example, when it is determined that the search target word "road" has the similar meaning word "street" (yes in step SE2), the control unit 31 determines whether or not the similar meaning word "street" is included in the search history stored in the user management database storage area 32d (see fig. 7) in association with the dictionary < user > ID of the user terminal 20, which is the transmission source of the specified word (search target word) "road" (step SE 3).
Here, when it is determined that the similar meaning word "street" of the search target word "road" is included in the search history of the corresponding user (yes at step SE3), the control unit 31 adds the similar meaning word "street" to the search target word list in the work data storage area 32f and stores the same (step SE 4).
Then, the control unit 31 performs a search for a predetermined number of pieces (20 pieces in this case) from the Web server 40 on the communication network N, for example, as shown in fig. 24 a, and acquires related images Ra1, Ra2, Ra3, which are related to the search target word (designated word) "road" at the head of the search target word list (step SE 5).
Then, the control unit 31 determines whether or not there is a related image having no relation to the word sense of the search target word "road" for 20 related images Ra1, Ra2, Ra3,. that were acquired in step SE5 (step SE6), and if it is determined that there is a related image (yes in step SE6), sets the unrelated related image RaN to be not displayed (steps SE7, SE 8).
The control unit 31 determines that there is a next search target word (synonym) "street" in which the search is not performed for the related image from the Web server 40.. in the search target word list (yes at step SE9), and, as in the same manner as described above, searches are performed for a predetermined number of pieces (20 pieces in this case) from the Web server 40 on the communication network N, as shown in fig. 24 a, for example, to acquire related images (synonym related images) SR1, SR2, SR3,. related to the search target word "street" (step SE 5).
Then, the control unit 31 determines whether or not there is a related image having no relation to the word meaning of the search target word "street" for 20 related images SR1, SR2, SR3,. cndot.obtained in step SE5 (step SE6), and if it is determined that there is a related image (step SE6 (yes)), sets the unrelated related image RaN to be not displayed (steps SE7, SE 8).
Then, if it is determined that there are no search target words in the search target word list for which a search from the Web server 40. was not performed for the related images (no) (step SE9), and if it is determined that there are 2 or more search target words in the search target words from "road" as the specified word and "street" as the synonym (yes at step SE10), the control unit 31 transmits data of a list of related images Ra1, Ra2, Ra3,. corresponding to the specified word "road" and data of a list of related images SR1, SR2, SR3,. corresponding to the synonym "street" to the user terminal 20 as Web related image list data (step SE 12).
When the user terminal 20 receives the Web related image list data corresponding to the designated word "road" and the synonym word "street" transmitted from the WES server 30 (yes at step TA 6), the control unit 21 shifts to the related image update processing (fig. 13) (step TB).
The control unit 21 receives the Web related image list data from the WES server 30 (yes in step TB 1), and determines that the synonym "street" and the synonym-related image of the designated word "road" are included in the Web related image list data (yes in step TB 2).
As shown in fig. 24 a, the control unit 21 causes the touch panel display unit 27 to display a list of related images Ra1, Ra2, Ra3,. and a list of related images SR1, SR2, SR3,. and. of the specified word "road" and a list of related images SR1, SR2, SR3,. and. of the synonym "street" received from the WES server 30 as a related image list screen GL in which the screen is divided into a Web related image list RL1 and a Web synonym related image list Sla (step TB 6).
Here, when the related image Ra3 preferred by the user H is specified from the Web related image list RL1 corresponding to the specified word "road", and the similar meaning term related image SR1 preferred by the user H is specified from the Web similar meaning term related image list Sla corresponding to the similar meaning term "street", and the [ image creation ] button BM is pressed (yes in step TB 7), the control section 21 causes the specified frame W to be displayed on the specified related image Ra3 and the similar meaning term related image SR1, and transmits the data of the related images Ra3 and SR1 to the WES server 30 (step TB 8).
Here, as described above, in the Web related image list RL1 and the Web related image list Sla displayed on the user terminal 20, when the related image Ra3(SR1) designated by the user H can be shared with the wordbook of another user, as shown in fig. 23, the related image Ra3 is designated (SR1) after the [ sharable ] button BCy is pressed, and when sharing is not desired, the related image Ra3 is designated (SR1) after the [ unshared ] button BCn is pressed.
In this way, the data of the related image Ra3 (related close sense word image SR1) designated from the Web related image list RL1(Web related close sense word image list SLa) in the user terminal 20 is transmitted to the WES server 30 together with information that can be shared or not shared (step TB 8). The data of the designated related image Ra3 transmitted to the WES server 30 may be identification information (such as an index) for identifying the image data, instead of the image data itself.
When the data (including sharable/non-sharable) of the designated related image Ra3 and the related image SR1 transmitted from the user terminal 20 and the press signal of the [ image creation ] button BM are received together in the WES server 30 (yes in step S11), the control unit 31 updates the related image corresponding to the word card "road" of the corresponding word book in the word book database storage area 32e (see fig. 8) transmitted and displayed to the user terminal 20 in step S7, including the update date and time and sharable/non-sharable information, to the received related image Ra 3. Further, the received synonym "street" and its synonym associated image SR1 (containing the update date and time and sharable/non-sharable information) are added to update the word card "road" of the word book (step S12).
Thus, the user specifies the related image Ra3 preferred by the user H from the Web related image list RL1 retrieved from the Web server 40 on the communication network N, and updates the related image generated in the WES server 30 and associated with an arbitrary word (here, "road") of the own word book to the specified related image Ra 3. When the similar meaning word "street" of the word "road" is included in the search history of the user, the word card of the word "road" is updated by adding the similar meaning word related image SR1 preferred by the user H specified from the similar meaning word "street" and the Web similar meaning word related image list SLa.
In this case, as shown in fig. 21, when the user terminal 20 displays the word card of the designated word (here, "road") of the own word book received from the WES server 30 (steps TA1 to TA3), the related image page TCb displays the related image Ra3 designated by the user of the word "road" and the related image SR1 designated by the user of the synonym "street" in parallel so as to enable comparison, as shown in fig. 24 (B), for example.
Therefore, the user can understand the difference between the word "road" and the word meaning of the similar word "street" (for example, "road" is a road through which a vehicle continuously travels from a village to a village/"street" is a road in a city where buildings and street trees are erected side by side) deeply and easily, and the word "road" can be effectively and efficiently remembered together with the similar word "street".
< processing for adding memo data Me to a word book and updating >
In the wordbook utilization process (fig. 12) of the user terminal 20, for example, as shown in fig. 22C, in a state where the word translation page TCc corresponding to the designated word "archive" is displayed (step TA3), when the field of the memo data Me of the word translation page TCc is designated (yes in step TA 10), the control unit 21 causes a soft keyboard, not shown, to display and sets the designated field of the memo data Me in a character input enabled state (step TA 11).
Here, an arbitrary memo (here, the word "archive" and "erythro い part は japanese depository で.", which is a harmonic of japanese translation) is input by an operation of the user, displayed, and a sharable or non-sharable [ sharable ] button BCy or [ non-sharable ] button BCn (see fig. 23) for specifying the memo is pressed.
The control unit 21 transmits the designated word "archive" and the input memo data "erythro い part は japanese Japanese raisin bureau で." Me "to the WES server 30 together with information sharable or not sharable (step TA 12).
Upon receiving the specified word "archive" and memo data "erythro い transmitted from the user terminal 20, part は of the WES server 30 has preceded rabbeting the phrase で." Me (including sharable/non-sharable) (yes in step S20), the control portion 31 stores the received memo data Me in association with the word "archive" of the corresponding vocabulary in the vocabulary database storage area 32e, including sharable or non-sharable information, and updates the memo data Me (step S21).
Thus, the user can effectively and efficiently learn a word by adding any memo data Me which is thought to be easy to remember the word to the word in association with any word in his/her word book.
When the user terminal 20 performs an operation to close the screen being displayed (yes in step TA 7), the control unit 21 ends the series of wordbook use processing.
< processing for retrieving and updating related images of word book from the WES server 30 >
For example, as shown in fig. 21B, in a state where the touch panel display unit 27 of the user terminal 20 is caused to display the related image page TCb of the designated word card "archive" (step TA3), the control unit 21 searches for a word book of another user in the WES server 30 and updates the displayed current related image to the related image preferred by the user, for example, and when the [ WES ] button BT2 of the related image page TCb is pressed (yes in step TA 8), as shown in fig. 22a, transmits pressing signals of the designated words "archive" and the [ WES ] button BT2 to the WES server 30 (step TA 9).
Upon receiving the designated word "archive" and the press signal of the [ WES ] button BT2 transmitted from the user terminal 20 at the WES server 30 (yes in step S13), the control unit 31 goes to the WES-related image list generation process (fig. 18) (step SF).
Fig. 18 is a flowchart showing a we related image list generation process included in the wordbook process of the WES server 30.
The control unit 31 performs a predetermined number of searches as shown in fig. 22 (C), for example, from the data of the other users' word books stored in the word book database storage area 32e to acquire related images Rb1, Rb2, Rb3, etc. corresponding to the designated word "archive" (step SF 1).
The control unit 31 transmits the data of the list of the related images Rb1, Rb2, Rb3,. corresponding to the acquired designated word "archive" to the user terminal 20 as the WES related image list data (step SF 2).
Upon receiving the we-related-image list data corresponding to the designated word "archive" transmitted from the we server 30 in the user terminal 20 (yes at step TA 6), the control unit 21 proceeds to the related-image updating process (fig. 13) (step TB).
Upon receiving the we-related-image list data from the WES server 30 (yes at step TB 9), the control unit 21 causes the touch panel display unit 27 to display a list of the related images Rb1, Rb2, Rb3,. of the designated word "archive" received from the WES server 30 as a WES-related-image list RL2 as shown in fig. 22C (step TB 10). By displaying the we related image list RL2, the user can know which related image is set by another user with respect to the word designated by the user.
Fig. 25 is a diagram showing an operation example of filtering related images in the display of the we related image list RL2 following the related image update process of the user terminal 20.
In the WES related image list RL2, for example, as shown in fig. 25, when the [ filter ] button PD for filtering related images Rb1, Rb2, Rb3,. is pressed, the control unit 21 displays, as pull-down menus, a [ recent ] button BX for searching related images retrieved from the word book database (32e) in the order of the update date and time from the new to the old, [ personal preference ] button BY for searching related images retrieved from word books of other users having the same user attribute as the user, and a [ recommendation ] button BZ for searching related images retrieved from word books of other users having the same related image feature as the user.
In response to the displayed [ recent ] button BX, [ personal preference ] button BY, or [ recommended ] button BZ being pressed, control unit 21 transmits a pressing signal of button BX, BY, or button BZ to WES server 30 (step TB 11).
Upon receiving the press signal of [ recent ] button BX from user terminal 20 (yes at step SF 3), control unit 31 retrieves a related image of the specified word "archive" from the data of the vocabulary of the other users in the vocabulary database (32e) in the order of the latest date and time (posting order) (step SF4), and transmits the data of the list of the retrieved related images as "recent" we-related-image list data to user terminal 20 (step SF 5).
Further, upon receiving the press signal of the [ personal atmosphere ] button BY from the user terminal 20 (yes at step SF 6), the control unit 31 acquires a user attribute (for example, senior three students) associated with the dictionary < user > ID of the user from the user management database (32d) (step SF 7).
Then, the related image of the designated word "archive" is retrieved and acquired from the word book of another user in the word book database (32e) having the same user attribute (senior three students) as the acquired user attribute (step SF8), and the data of the list of the acquired related images is transmitted to the user terminal 20 as the we related image list data of "popularity" (step SF 9).
Further, upon receiving a pressing signal of the [ recommended ] button BZ from the user terminal 20 (yes at step SF 10), the control unit 31 acquires an associated image feature (for example, a harmonic image of a word) associated with the dictionary < user > ID of the user from the user management database (32d) (step SF 11).
Then, a related image of the designated word "archive" is retrieved and acquired from a word book of another user in a word book database (32e) having the same related image feature (harmonic image of the word) as the acquired related image feature (step SF12), and data of a list of the acquired related images is transmitted to the user terminal 20 as "recommended" WES related image list data (step SF 13).
Upon receiving the we-related-image list data of "recent", "popularity", or "recommended" transmitted from the WES server 30 in the user terminal 20 (yes in step TB 9), the control unit 21 displays a list of the received filtered related images Rb1, Rb2, Rb3,. or. of the designated word "archive" on the touch panel display unit 27 as the we-related-image list RL2, for example, as shown in fig. 22 (C) (step TB 12).
When the related image Rb2 preferred by the user H (here, the related image in which the word "archive" and the japanese translation harmonic sound "red いブタを are stored") is specified in the we related image list RL2 of the specified word "archive" displayed in step TB10 or TB12, the control unit 21 causes the specified related image Rb2 to display the specified frame W and transmits the data of the specified related image Rb2 to the WES server 30 (step TB 13). The data of the designated related image Rb2 transmitted to the WES server 30 may be identification information (such as an index) for identifying the image data, instead of the image data itself.
Upon receiving the designated related image Rb2 (here, the related image stored in "red いブタを" as the harmonic of the word "archive") transmitted from the user terminal 20 (yes in step S14), the control unit 31 transmits, to the user terminal 20, data of a word card (including the word "archive", the related image Rb2, the description information, and the memo) in which the related image Rb2 and the memo are associated with the related image Rb2 of the word book of the other user, which is the search source of the designated related image Rb2, together with sharable/unusable information (step S15).
In the user terminal 20, the control unit 21 receives the data of the word cards of the wordbooks of other users corresponding to the designated related image Rb2 transmitted from the WES server 30 (the related image Rb2 and the memo also include sharable/unusable information). Then, as shown in fig. 26 (a), (B), and (C), for example, the english word page TCa, the related image page TCb, and the word translation page TCc of the received word card "archive" of the other user are switchably displayed on the touch panel display unit 27 (step TB 14).
In the processing up to now, the word designated by the user can be viewed not only the related image set by another user but also the word book of another user in which the similar meaning word and the memo are written.
At this time, when sharable information is added to the received related image Rb2 of the token card of another user, a download flag DL indicating sharability is additionally displayed on the related image page TCb as shown in fig. 26 (B). When sharable information is added to the received memo data of the word card of another user, a copy flag CP indicating sharability is additionally displayed on the word translation page TCc as shown in fig. 26 (C).
Here, as shown in fig. 26B, when a related image of the related image page TCb on which the download flag DL is displayed (a related image stored in "red いブタを" as a harmonic of the word "archive") is designated (yes in step TB 15), the control unit 21 transmits data of the designated related image (sharable) (the data of the designated related image may be identification information of the related image) to the WES server 30 (step TB 16).
Upon receiving the data of the designated related image (sharable) transmitted from the user terminal 20 (yes at step S16), the control unit 31 updates the related image corresponding to the word card "archive" of the corresponding word book in the word book database (32e) transmitted to the user terminal 20 and displayed, as shown in fig. 22a, to the related image Rb2 (related image stored in "red いブタを" as a harmonic of the word "archive") of the received other user' S word card (step S17) including the update date and time and sharable information.
Further, the WES server 30 may be configured as follows: when receiving the designated related image Rb2 transmitted from the user terminal 20 (here, a related image in which the harmonic sound of the word "archive" is stored in red いブタを ") (yes in step S14), data of the corresponding word card (including the word" archive "and the related image Rb2) is transmitted to the user terminal 20, the related image Rb2 of the word book of the other user, which is the retrieval source of the designated related image Rb2, is created and displayed (step S15/TB14), the user is not subjected to the process of confirming and designating the related image (steps TB15, TB16/S16), and the related image corresponding to the word card" archive "of the corresponding word book transmitted to the user terminal 20 and displayed in the word book database (32e) is updated to the related image Rb2 of the word card of the other user received in step S14.
Thus, the user can specify the related image Rb2 that the user H prefers to remember words more easily than the previous related image in the word book of the user H, for example, from the WES related image list RL2 retrieved from the data of the word books of other users in the WES server 30, and update the related image generated in the WES server 30 and associated with an arbitrary word in the own word book to the specified related image Rb 2.
In the we related image list RL2 searched for "recent", "popularity", or "recommendation", the user specifies the related image Rb2 preferred by the user H as the related image related to the word of the user H and updates the related image related to the word of the user's own word book by targeting the related image updated most recently in the case of "recent", targeting the related image of the word book of the other user having the same attribute as the user in the case of "popularity", and targeting the related image of the word book of the other user having the same feature as the feature of the related image of the word book of the user in the case of "recommendation".
In addition, the press information of "recent", "popularity", and "recommendation" is an example of the filtering information.
< processing for updating memo data Me of a word book with word books of other users >
Further, following the processing of step TB14 of the user terminal 20, when memo data (here, the harmonic "red いブタを, storage すゐ" of the word "archive") Me of the word translation page TCc displayed with the copy flag CP added as shown in fig. 26 (C) among the english word page TCa, related image page TCb, and word translation page TCc of the word card "archive" of another user switchably displayed as shown in fig. 26 (a), (B), and (C) is specified (step TB17 (yes)), the control unit 21 transmits the specified memo data (sharable) Me to the WES server 30 (step TB 18).
Upon receiving the specified memo data (sharable) Me transmitted from the user terminal 20 (yes in step S18), the control unit 31 transmits the memo data "red い part は" Me corresponding to the word card "archive" of the corresponding word book of the word book database (32e) to the user terminal 20 as shown in fig. 21 (C) and causes the displayed memo data to be updated to the received memo data "red いブタを, storage する" Me of the word cards of the other users including sharable information (step S59 19).
Thus, the user can specify memo data Me of the word book of another user having a related image of his/her own preference in the WES server 30 (a memo in which a word is more likely to be memorized than the memo located in the word book of his/her own person), and update the memo data Me generated in the WES server 30 and associated with an arbitrary word of his/her own word book to the specified memo data Me.
In step TB14 in the related image updating process (fig. 13) of the user terminal 20, for example, as shown in fig. 26, when a [ back ] button (not shown) is pressed in a state where the english word page TCa, the related image page TCb, and the word translation page TCc of the vocabulary book of the other user received from the WES server 30 are switchably displayed (step TB19 (yes)), if the filtering operation is performed in step TB11, the control unit 21 returns to the process of step TB12 to display the filtered WES related image list RL2 (step TB20 (yes) → TB12), and if the filtering operation is not performed, the control unit 21 returns to the process of step TB10 to display the non-filtered WES related image list RL2 (step TB20 (no) → TB 10).
(summary of the embodiments)
According to the learning support system 1 of the embodiment, when data of a search history of words searched by the electronic dictionary 10 is transmitted to the WES server 30 via the user terminal 20 as a communication device, the WES server 30 generates a word card in which the word, explanatory information including word translation, and related images of the word are acquired from dictionary data of the dictionary database (32c) and associated with each word included in the data of the search history, automatically generates a word book corresponding to the user, and stores the word book in the word book database (32e) in association with the dictionary < user > ID.
The user acquires the word book automatically generated by the WES server 30 at the user terminal 20, and the user terminal 20 displays each word card of the word book while switching between the english word page TCa, the related image page TCb, and the word translation page TCc, thereby performing learning.
At this time, when an arbitrary memo data Me (such as a reading of a word and a memo that becomes a harmonic of the word translation) that is considered to be easy to remember is input in the memo data Me column of the word translation page TCc, for example, the memo data Me is added to the word card of the corresponding word book of the WES server 30 and updated, and thereafter, the memo data Me is added to the word translation page TCc displayed on the user terminal 20.
Thus, the user can easily understand the meaning of a word intuitively by looking at the related image of the word in the word book, and can easily memorize the word intuitively by looking at the consonant of the pronunciation of the word added as memo data, thereby effectively and efficiently memorizing the word and performing learning.
Further, according to the learning support system 1 of the embodiment, when the [ Web ] button BT1 is pressed on the associated image page TCb of an arbitrary wordbard displayed on the user terminal 20, a list of associated images related to words of the wordbard retrieved from the Web by the WES server 30 is displayed as the Web associated image list RL 1. When any related image is specified from the Web related image list RL1, the WES server 30 updates the related image of the word card in the corresponding word book in the word book database (32e) to the related image specified by the user.
Thus, the user can specify the related image Ra3 preferred by the user H in the Web related image list RL1 retrieved from the Web and update the related image associated with an arbitrary word in the own word book.
When a related image related to a word of an arbitrary word card displayed on the user terminal 20 is searched from the Web by the WES server 30, if a near word of the word is located in the dictionary database (32c) and the near word is included in the search history of the user, a related image of the near word is also searched from the Web, and a list of the searched related images of the near word is displayed on the user terminal 20 as a Web near word related image list Sla and a Web related image list RL1 in a screen-divided manner. When any related image is specified from each of the Web related image list RL1 and the Web similar meaning term related image list Sla, the WES server 30 updates the related image of the word card of the corresponding word book located in the word book database (32e) to the related image specified by the user, and adds the similar meaning term and the similar meaning term related image specified by the user to update the word card.
Thus, the user can understand the difference between the word of the word card displayed on the user terminal 20 and the word meaning of the similar meaning word with high precision and ease from the difference between the content of the related image of the word and the content of the related image of the similar meaning word, and the similar meaning word can be learned effectively and efficiently.
Further, according to the learning support system 1 of the embodiment, when the [ WES ] button BT2 is pressed on the [ WES ] image page TCb of an arbitrary word displayed on the user terminal 20, a list of related images related to the word retrieved from the word book of another user located in the word book database (32e) by the WES server 30 is displayed as the WES related image list RL 2. When any related image is specified from the WES related image list RL2, the WES server 30 updates the related image of the wordboard of the user terminal 20 located in the wordbook database (32e) to the specified related image.
Thus, the user can specify the related image Rb2 that is found to be easier to remember than the previous related image of the own word book, for example, in the we related image list RL2 retrieved from the word book of another user in the WES server 30, and update the related image associated with an arbitrary word of the own word book to the specified related image Rb 2.
Further, BY the WES server 30 searching for and displaying the WES related image list RL2 from the wordbooks of other users on the user terminal 20, BY operating the [ recent ] button BX, the [ personal preference ] button BY, or the [ recommendation ] button BZ, it is possible to screen a list of related images in the order from new to old at the update date and time, a list of related images searched for from the wordbooks of other users having the same user attributes as the user of the user terminal 20, or a list of related images searched for from the wordbooks of other users having the same related image features as the user. The user can update the related image associated with the word of the own word book by specifying the more appropriate related image Rb2 of the word book of the other user which is close to the learning feeling preferred by the user.
In the learning support system 1 according to the embodiment, when the user designates any related image Rb2 of the we related image list RL2 corresponding to the designated word displayed on the user terminal 20, the WES server 30 transmits the data of the word card of the word book of the other user, which is the search source of the designated related image Rb2, to the user terminal 20, and the words are displayed so as to be switchable to the english word page TCa, the related image page TCb, and the word translation page TCc. When the user designates the memo data Me of the word translation page TCc, the WES server 30 updates the current memo data Me corresponding to the designated word of the user's word book to the memo data Me of the designated other user.
Thus, the user can designate the memo data Me of the word book of another user having the related image preferred by the user in the WES server 30 as the memo data Me which is thought to be easier to remember the word, with respect to the current memo data Me associated with an arbitrary word of the word book of the user, and update the memo data Me to the designated better memo data Me.
(other embodiments)
In the WES related image list generation process of the WES server 30, the process of sorting the related images corresponding to the [ personal preference ] button BY stores the selection time until the user selects and specifies the related image from the display of the Web related image list RL1 or the WES related image list RL2 in association with the data of the related image of the word book. Then, a method of selecting a related image whose selection time is shorter than a predetermined time may be assumed as a method of filtering related images that are easy for the user to remember.
Further, the whole related image with clear contrast, the related image with clear background of the feature portion corresponding to the word meaning of the word, the related image with the feature portion having a size larger than a predetermined size, and the like may be subjected to image recognition and analysis, and the related image may be subjected to the filtering corresponding to the [ personal ] button BY.
Further, the filtering process of the related image corresponding to the [ recommended ] button BZ may be as follows: when a related image in the past is updated, related images located in the word book of another user who has specified and updated a related image preferred by the user based on the WES related image list RL2 are screened as objects.
In the Web related image list generation process, the URL on the Web of the related image for which downloading from the Web is prohibited may be acquired and displayed in place of the related image for which downloading is prohibited in the Web related image list RL 1.
In addition, in the WES server 30, when generating a word book based on the data of the search history, the related image associated with the word is the related image of the word in the dictionary data in the dictionary database (32c), but the present invention is not limited thereto, and may be a related image searched from the Web.
Note that the words of the search history described in the embodiment may be various words such as idioms, compound words, and abbreviations that exist as entries of dictionary data, for example.
Therefore, the learning support system 1 according to the embodiment can effectively and efficiently memorize the language of the learning object and perform learning.
The methods of the processes of the learning support system 1 described in the above embodiments, that is, the WES server synchronization process in the electronic dictionary 10 shown in the flowchart of fig. 9, the WES utilization process in the user terminal 20 shown in the flowchart of fig. 10, the wordbook generation process in the WES server 30 shown in the flowchart of fig. 11, the wordbook utilization process in the user terminal 20 shown in the flowchart of fig. 12, the related image update process in the user terminal 20 shown in the flowcharts of fig. 13 and 14, the wordbook processing in the WES server 30 shown in the flowcharts of fig. 15 and 16, the Web related image list generation process in the WES server 30 shown in the flowchart of fig. 17, and the WES related image list generation process in the WES server 30 shown in the flowchart of fig. 18, can be stored as programs that can be executed by a computer in a memory card (ROM, card, or the like), RAM cards and the like), magnetic disks (floppy (registered trademark) disks, hard disks and the like), optical disks (CD-ROMs, DVDs and the like), semiconductor memories and the like. The control unit (CPU) of the electronic device reads a program recorded on a medium of the external recording device into the storage device, and controls the operation using the read program to realize the various functions described in the embodiments, thereby performing the same processing as the above-described method.
Further, data of a program for realizing each method can be transmitted as program codes over a communication network (N), and the data of the program is fetched from a computer device (program server) connected to the communication network (N) into an electronic device and stored in a storage device, thereby realizing the above-described various functions.
The present invention is not limited to the embodiments, and various modifications can be made in the embodiments without departing from the scope of the invention. Further, the embodiments can be combined and carried out as appropriate, and in this case, combined effects can be obtained. Further, the embodiments include various inventions, and various inventions can be extracted by combinations selected from a plurality of disclosed constituent elements. For example, in the case where the problem can be solved and the effect can be obtained even if several constituent elements are deleted from the entire constituent elements shown in the embodiments, the configuration in which the constituent elements are deleted can be extracted as an invention.

Claims (13)

1. A server device is characterized by comprising:
a communication unit;
a control unit; and
a storage unit for storing word books of a plurality of users in association with each of the plurality of users,
the control unit, upon receiving a specified word or phrase specified by a user via the communication unit, extracts a related image related to the specified word or phrase from a word book other than the user stored in the storage unit, and transmits the related image to the communication device of the user via the communication unit.
2. The server device according to claim 1, configured to,
upon receiving the related image screening information, the control unit extracts a related image based on the related image screening information.
3. The server device according to claim 1 or 2, configured to,
upon receiving information on 1 related image selected by a user from the transmitted related images, the control unit updates the related image of the designated sentence recorded in the user's word book to an image corresponding to the received information on the related image.
4. The server device according to claim 1 or 2, configured to,
upon receiving selection information, which is information of 1 related image selected by the user from the transmitted related images, the control unit transmits word card data corresponding to the selection information recorded in the word book other than the user to the user.
5. The server device according to claim 4, configured to,
upon receiving a related image update request, the control unit updates the related image of the designated word or phrase recorded in the user's word book to the related image of the transmitted word card data.
6. The server device according to claim 4 or 5, configured to,
the control unit updates memo data of the specified word or phrase recorded in the user's word book to the transmitted memo data of the word card data upon receiving a request for updating memo data associated with the specified word or phrase.
7. A learning support method executed by a control unit of a server device, the server device comprising:
a communication unit;
the control section; and
a storage unit for storing word books of a plurality of users in association with each of the plurality of users,
when a specified word or phrase specified by a user is received via the communication unit, the control unit extracts a related image associated with the specified word or phrase from a word book other than the user stored in the storage unit, and transmits the related image to the communication device of the user via the communication unit.
8. The learning support method according to claim 7, wherein the learning support method further comprises the step of,
upon receiving the related image screening information, the control unit extracts a related image based on the related image screening information.
9. The learning support method according to claim 7 or 8, wherein the learning support method further comprises the step of determining the learning support target value,
upon receiving information on 1 related image selected by a user from the transmitted related images, the control unit updates the related image of the designated sentence recorded in the user's word book to an image corresponding to the received information on the related image.
10. The learning support method according to claim 7 or 8, wherein the learning support method further comprises the step of determining the learning support target value,
the control unit, upon receiving selection information, which is information of 1 related image selected by the user from the transmitted related images, transmits word card data corresponding to the selection information recorded in the word book other than the user to the user.
11. The learning support method according to claim 10, wherein the learning support method further comprises the step of,
upon receiving a related image update request, the control unit updates the related image of the designated word or phrase recorded in the user's word book to the related image of the transmitted word card data.
12. The learning support method according to claim 10 or 11, wherein the learning support method further comprises the step of determining the learning support target value,
the control unit updates memo data of the specified word or phrase recorded in the user's word book to the transmitted memo data of the word card data upon receiving a request for updating memo data associated with the specified word or phrase.
13. A non-transitory computer-readable recording medium having a program recorded thereon for causing a control unit of a server device to function,
the server device includes:
a communication unit;
the control section; and
a storage unit for storing word books of a plurality of users in association with each of the plurality of users,
the program causes the control unit to function as follows:
when a specified word or phrase specified by a user is received via the communication unit, a related image associated with the specified word or phrase is extracted from a word book other than the user stored in the storage unit, and the related image is transmitted to the communication device of the user via the communication unit.
CN202210249179.9A 2021-03-24 2022-03-14 Server device, learning support method, and recording medium Pending CN115132007A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021050488A JP7298638B2 (en) 2021-03-24 2021-03-24 SERVER DEVICE, LEARNING SUPPORT METHOD AND PROGRAM
JP2021-050488 2021-03-24

Publications (1)

Publication Number Publication Date
CN115132007A true CN115132007A (en) 2022-09-30

Family

ID=83375983

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210249179.9A Pending CN115132007A (en) 2021-03-24 2022-03-14 Server device, learning support method, and recording medium

Country Status (2)

Country Link
JP (1) JP7298638B2 (en)
CN (1) CN115132007A (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7422443B1 (en) 2023-08-28 2024-01-26 良明 片野 Memory support processing device and memory support program

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101458704A (en) * 2007-12-14 2009-06-17 卡西欧计算机株式会社 Electronic device
CN101855660A (en) * 2007-11-14 2010-10-06 夏普株式会社 Electronic device, control program, recording medium, and control method
CN102165437A (en) * 2008-07-25 2011-08-24 夏普株式会社 Information processing device and information processing method
JP2012008738A (en) * 2010-06-23 2012-01-12 Casio Comput Co Ltd Dictionary retrieval device and program
JP2015179335A (en) * 2014-03-18 2015-10-08 カシオ計算機株式会社 Electronic equipment provided with dictionary function, electronic equipment provided with word book function, and program
CN105280086A (en) * 2014-06-05 2016-01-27 卡西欧计算机株式会社 Learning support apparatus, and data output method in learning support apparatus
JP2016118650A (en) * 2014-12-19 2016-06-30 カシオ計算機株式会社 Electronic apparatus having dictionary function, electronic dictionary system and control program of electronic apparatus having dictionary function
JP2020160382A (en) * 2019-03-28 2020-10-01 シャープ株式会社 Learning support system, learning support method, and learning support program

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6513129B2 (en) * 2017-05-20 2019-05-15 武史 藤田 Document retrieval learning system

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101855660A (en) * 2007-11-14 2010-10-06 夏普株式会社 Electronic device, control program, recording medium, and control method
CN101458704A (en) * 2007-12-14 2009-06-17 卡西欧计算机株式会社 Electronic device
CN102165437A (en) * 2008-07-25 2011-08-24 夏普株式会社 Information processing device and information processing method
JP2012008738A (en) * 2010-06-23 2012-01-12 Casio Comput Co Ltd Dictionary retrieval device and program
JP2015179335A (en) * 2014-03-18 2015-10-08 カシオ計算機株式会社 Electronic equipment provided with dictionary function, electronic equipment provided with word book function, and program
CN105280086A (en) * 2014-06-05 2016-01-27 卡西欧计算机株式会社 Learning support apparatus, and data output method in learning support apparatus
JP2016118650A (en) * 2014-12-19 2016-06-30 カシオ計算機株式会社 Electronic apparatus having dictionary function, electronic dictionary system and control program of electronic apparatus having dictionary function
JP2020160382A (en) * 2019-03-28 2020-10-01 シャープ株式会社 Learning support system, learning support method, and learning support program

Also Published As

Publication number Publication date
JP2022148704A (en) 2022-10-06
JP7298638B2 (en) 2023-06-27

Similar Documents

Publication Publication Date Title
US6192332B1 (en) Adaptive electronic phrase book
JP6413256B2 (en) CONFERENCE SUPPORT DEVICE, CONFERENCE SUPPORT DEVICE CONTROL METHOD, AND PROGRAM
CN108846037B (en) Method and device for prompting search terms
US9262399B2 (en) Electronic device, character conversion method, and storage medium
CN110738996B (en) Method for controlling printer printing through voice and printing terminal
US20060190240A1 (en) Method and system for locating language expressions using context information
EP3242220A1 (en) Intelligent prediction input method and system
CN115132007A (en) Server device, learning support method, and recording medium
WO2017168241A1 (en) Menu information providing system, menu information providing method, and non-transitory computer-readable storage medium storing menu information providing program
US10546512B2 (en) Learning support apparatus, data output method in learning support apparatus, and storage medium
CN111078982A (en) Electronic page retrieval method, electronic device and storage medium
JP5780208B2 (en) Electronic device having communication function with electronic dictionary and recording medium recording electronic device control program
CN102033907A (en) Electronic apparatus with dictionary function
JP2008021235A (en) Reading and registration system, and reading and registration program
KR20130076852A (en) Method for creating educational contents for foreign languages and terminal therefor
JP2022148702A (en) Server device, learning support method and program
JP2016091020A (en) Program, processing method, and information processing apparatus
JP2005292303A (en) Information display controller, information display control processing program, and information management server
JP2016118650A (en) Electronic apparatus having dictionary function, electronic dictionary system and control program of electronic apparatus having dictionary function
JP7036483B1 (en) Programs, information processing equipment and information processing methods
JP2019021337A (en) Conference support device, conference support device control method, and program
JP7448049B2 (en) Information processing equipment, learning support methods and programs
JPWO2005059769A1 (en) Text creation device with modification information
WO2019038635A1 (en) Device to teach an alphabet
JP7420293B2 (en) Learning support systems, learning support methods and programs

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination