WO2017217288A1 - Electronic apparatus, head-mounted display, processing method of electronic apparatus, and program therefor - Google Patents

Electronic apparatus, head-mounted display, processing method of electronic apparatus, and program therefor Download PDF

Info

Publication number
WO2017217288A1
WO2017217288A1 PCT/JP2017/021067 JP2017021067W WO2017217288A1 WO 2017217288 A1 WO2017217288 A1 WO 2017217288A1 JP 2017021067 W JP2017021067 W JP 2017021067W WO 2017217288 A1 WO2017217288 A1 WO 2017217288A1
Authority
WO
WIPO (PCT)
Prior art keywords
character
string
voice
number string
password
Prior art date
Application number
PCT/JP2017/021067
Other languages
French (fr)
Japanese (ja)
Inventor
軌行 石井
Original Assignee
コニカミノルタ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by コニカミノルタ株式会社 filed Critical コニカミノルタ株式会社
Publication of WO2017217288A1 publication Critical patent/WO2017217288A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/36User authentication by graphic or iconic representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • G10L17/22Interactive procedures; Man-machine interfaces
    • G10L17/24Interactive procedures; Man-machine interfaces the user being prompted to utter a password or a predefined phrase
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/64Constructional details of receivers, e.g. cabinets or dust covers

Definitions

  • the present invention relates to an electronic device having a voice recognition function, a head mounted display, a processing method of the electronic device, and a program thereof.
  • a typical mobile terminal has a touch panel screen that serves both as an image display and a user interface. By touching this screen, the user can make necessary inputs to display a desired image or input information. Can be performed.
  • biometric authentication is performed in which a biometric pattern such as a user's fingerprint, voiceprint, vein, retina, etc. is read for authentication.
  • biometric authentication requires a sensor for reading a biometric pattern and dedicated software for performing pattern matching processing, resulting in a problem that the system becomes complicated and costs increase.
  • the biometric patterns of the plurality of people must be stored, which is inconvenient.
  • some portable terminals have a voice recognition function to realize hands-free. Therefore, there is an idea that the mobile terminal can be unlocked without performing a button operation by using the voice recognition function.
  • voice recognition the words spoken by the user must first be collected with a microphone in order to convert them into character strings, etc. There is a risk that the password will be known.
  • the same problem also occurs when it is desired to input information such as a telephone number to be concealed into a portable terminal or the like using a voice recognition function.
  • Patent Document 1 a pair of keywords randomly selected from a plurality of pairs of keywords and passwords registered in advance is displayed on the display unit.
  • Patent Document 1 mentions the “correct answer rate”, which is a proof that the user is assumed to make a mistake, and there is a concern that there is difficulty in terms of usability.
  • the present invention has been made in view of the above circumstances, and is capable of suppressing leakage of information to a third party who has heard a user's utterance while using a voice recognition function, a head mounted display, and an electronic device. It is an object of the present invention to provide a device processing method and a program thereof.
  • an electronic device reflecting one aspect of the present invention.
  • a storage device for storing patterns;
  • a display device for displaying a plurality of images;
  • a voice recognition device that acquires voice spoken by a user according to the image and converts the voice into a corresponding character / number string;
  • the plurality of images displayed on the display device are connected in the order of the character / number string converted by the voice recognition device, and the locus matches the pattern stored in the storage device
  • a processing device that permits a predetermined function.
  • a storage device for storing a password consisting of a character / number string
  • a display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship
  • a voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string
  • a conversion device for converting the password into a second character / number string in accordance with the predetermined relationship
  • a processing device that permits a predetermined function when the first character / numeric string converted by the voice recognition device matches the second character / numeric string converted by the conversion device; is there.
  • a storage device for storing a password consisting of a character / number string
  • a display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship
  • a voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a third letter / number string
  • Processing for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password
  • a device for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password
  • a device for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric
  • still another electronic device reflecting one aspect of the present invention is: A display device that displays letters / numbers constituting a prescribed letter / number string in association with a plurality of images according to a predetermined relationship; A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string; A conversion device that converts the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship; And an input device for inputting the second character / number string as the prescribed character / number string.
  • still another electronic device reflecting one aspect of the present invention is: A display device for displaying a plurality of images; With a microphone, A voice recognition device that analyzes voice acquired through the microphone and recognizes a character / number string represented by the voice; A processing unit that specifies an image corresponding to the character / number string among the plurality of images and performs processing based on the specified image.
  • a processing method of an electronic device reflecting one aspect of the present invention is as follows.
  • the pattern Display multiple images, Acquire the voice spoken by the user according to the image, convert it to the corresponding character / number string,
  • the plurality of displayed images are connected in the order of the converted character / numerical string, when the locus coincides with the pattern stored in the storage device, predetermined authentication is performed. is there.
  • another electronic device processing method includes: Memorize passwords consisting of letters and numbers, Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship; Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string; Converting the password into a second character / number string according to the predetermined relationship; When the first character / numeric string matches the second character / numeric string, predetermined authentication is performed.
  • still another electronic device processing method reflecting one aspect of the present invention is as follows.
  • Memorize passwords consisting of letters and numbers, Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
  • the third character / number string is converted into a fourth character / number string according to the predetermined relationship, and when the fourth character / number string matches the password, predetermined authentication is performed. is there.
  • still another electronic device processing method reflecting one aspect of the present invention is as follows.
  • the letters / numbers that make up the prescribed letter / number string are displayed in association with a plurality of images according to a predetermined relationship, Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string; Converting the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
  • the second character / number string is input as the prescribed character / number string.
  • an electronic device a head-mounted display, a method for processing an electronic device, and a program thereof that can prevent information from leaking to a third party who has heard a user's speech while using a voice recognition function. Can do.
  • FIG. 1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state in which a user is wearing it.
  • 3 is a schematic cross-sectional view showing a configuration of a display unit 104.
  • FIG. It is a block diagram of HMD100 concerning this embodiment. It is a figure which shows the pattern PT memorize
  • FIG. It is a figure which shows the message displayed in order to confirm whether speech recognition was performed appropriately. It is a figure which shows the example of the character string CA produced
  • FIG. 1 is a figure which shows the arrangement
  • A is a figure which shows the numerical sequence (here "4 * 9 * 1 * 3 * 5") memorize
  • FIG. (B) is a figure which shows the combination image NG of the number and image produced
  • FIG. 1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state where a user wears it.
  • HMD head mounted display
  • the right side and the left side of the HMD 100 refer to the right side and the left side for the user wearing the HMD 100.
  • the frame 101 to be mounted on the head of the user US has two spectacle lenses 102 arranged in front of the user US.
  • a cylindrical main body 103 is fixed on the upper part of the spectacle lens 102 on the right side (which may be on the left side according to the user's dominant eye).
  • the main body 103 is provided with a display unit 104.
  • a display drive control unit 104DR (see FIG. 3 described later) that controls display control of the display unit 104 is disposed in the main body unit 103. If necessary, a display unit may be arranged in front of both eyes.
  • FIG. 2 is a schematic cross-sectional view showing the configuration of the display unit 104.
  • the display unit 104 as a display device includes an image forming unit 104A and an image display unit 104B.
  • the image forming unit 104A is incorporated in the main body unit 103, and includes a light source 104a, a unidirectional diffuser 104b, a condenser lens 104c, and a display element 104d.
  • the image display unit 104B which is a so-called see-through type display member, is disposed on the entire plate so as to extend downward from the main body unit 103 and extend in parallel to one eyeglass lens 102 (see FIG. 1).
  • the eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h is disposed on the entire plate so as to extend downward from the main body unit 103 and extend in parallel to one eyeglass lens 102 (see FIG. 1).
  • the light emitted from the light source 104a is diffused by the unidirectional diffusion plate 104b, condensed by the condenser lens 104c, and enters the display element 104d.
  • the light incident on the display element 104d is modulated for each pixel based on the image data input from the display drive control unit 104DR, and is emitted as image light. As a result, a color image is displayed on the display element 104d.
  • Image light from the display element 104d enters the eyepiece prism 104f from its base end face PL1, is totally reflected a plurality of times by the inner side face PL2 and the outer side face PL3, and enters the hologram optical element 104h.
  • the light incident on the hologram optical element 104h is reflected there, passes through the inner side surface PL2, and reaches the pupil B.
  • the user can observe an enlarged virtual image of the image displayed on the display element 104d, and can visually recognize it as a screen formed on the image display unit 104B.
  • the hologram optical element 104h constitutes a screen, or it can be considered that a screen is formed on the inner surface PL2.
  • “screen” may refer to an image to be displayed.
  • the eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h transmit almost all of the external light, the user can observe an external field image (real image) through them. Therefore, the virtual image of the image displayed on the display element 104d is observed so as to overlap with a part of the external image. In this manner, the user of the HMD 100 can simultaneously observe the image provided from the display element 104d and the external image via the hologram optical element 104h. Note that when the display unit 104 is in the non-display state, the image display unit 104B is transparent, and only the external image can be observed.
  • a display unit is configured by combining a light source, a liquid crystal display element, and an optical system.
  • a self-luminous display element for example, an organic EL display
  • Element for example, an organic EL display
  • a transmissive organic EL display panel having transparency in a non-light emitting state may be used.
  • FIG. 3 is a block diagram of the HMD 100 according to the present embodiment, which is shown together with the user US.
  • the HMD 100 collects the above-described display unit (display unit) 104, the microphone 105 that collects the voice spoken by the user US and converts it into a signal, and the voice that processes the signal output from the microphone 105 and outputs it as a voice signal.
  • the processing unit 106 and the control unit 110 that receives the audio signal output from the audio processing unit 106 are included.
  • the control unit 110 receives the voice signal output from the voice processing unit 106, recognizes a character / number string represented by the voice, and converts the character / number string to the corresponding character / number, and the voice recognition unit 113.
  • a processing unit (processing device) 112 that processes characters / numbers output from the display unit, a display drive control unit 104DR that receives signals from the processing unit 112 and drives and controls the display unit 104, and an authentication code (here, a pattern) Or an authentication code storage unit (storage device) 114 that stores a password.
  • the microphone 105, the voice processing unit 106, and the voice recognition unit 113 constitute a voice recognition device.
  • letters / numbers when referring to letters / numbers, it means at least one of letters and numbers, and when referring to letters / number strings, a plurality of letters or numbers are arranged (including only letters or numbers only).
  • An image includes letters / numbers.
  • a word includes a single character.
  • FIG. 4 is a diagram showing a pattern PT stored as authentication code information in the authentication code storage unit 114.
  • the pattern PT is formed by connecting a horizontal line from left to right, a vertical line from top to bottom, and a horizontal line from left to right in this order.
  • the pattern PT is assumed to be stored in advance by the user US.
  • the “pattern” refers to a traced geometric shape, and corresponds to, for example, a one-stroke stroke.
  • FIG. 5 is a diagram schematically showing a numeric string NA generated by the processing unit 112.
  • the arrow AR is drawn so as to be easily understood, but it is not actually displayed.
  • the numerical string NA is, for example, a random number generated by the processing unit 112 assigned to 3 rows and 3 columns.
  • the processing unit 112 transmits information on the numeric string NA to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the numeric string NA into an image signal and transmits it to the display unit 104, so that the display unit 104 can display the numeric string NA shown in FIG.
  • the user US does not need to remember the arrangement order of individual elements to be displayed such as character strings, numeric strings, and images in advance. It is only necessary to utter the elements displayed along the pattern PT in order, which can be said to be highly convenient. In order to improve confidentiality, it is preferable to display individual elements at random. Elements to be displayed may be numbers, letters (alphabet, hiragana, katakana, kanji ...), colors, and patterns. These are collectively called images. Only one type (for example, only hiragana) may be displayed, or a plurality of types (such as katakana and numbers) may be mixed.
  • the reading method is registered.
  • the pattern and its reading are registered in advance in the processing unit 112 and the user US also remembers the pattern and its registered reading.
  • a plurality of readings may be registered for one pattern. However, it is desirable to avoid using multiple pictures with the same reading (bridge, chopsticks, etc.).
  • the control unit 110 performs a display (not illustrated) for requesting an authentication code to be input to the display unit 104 via the display drive control unit 104DR.
  • the numeric string NA shown in FIG. 5 is displayed.
  • the user US Since the user US stores the pattern PT as the authentication code, when the stored pattern PT is applied to the numerical string NA displayed as shown in FIG. 5, the numbers “5” are entered in the order indicated by the arrow AR. If you read “3, 6, 2, 9”, you can see that the pattern PT can be reproduced. Therefore, when the user US utters a number in the form of “Go, San, Roku, Ni, Kyu”, the microphone 105 collects the sound, and the speech recognition unit 113 via the speech processing unit 106 collects this. Is converted into a numeric string “5, 3, 6, 2, 9” and transmitted to the processing unit 112 as numeric string information.
  • the processing unit 112 causes the display unit 104 to perform a display as shown in FIG. 6 via the display drive control unit 104DR, and confirms whether the input number string is as intended by the user US. You may do it.
  • the user US speaks “No” while the display of FIG. 6 is being performed, a voice signal is input from the microphone 105 via the voice processing unit 106 to the processing unit 112, and the displayed number string is inappropriate.
  • the processing unit 112 determines that it is a thing and requests the utterance of the user US again.
  • the processing unit 112 determines that the displayed number string is appropriate, and performs the subsequent processing.
  • the processing unit 112 that has received the number string information from the voice recognition unit 113 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the number string NA stored in itself. More specifically, when the numeric strings NA are connected in the order of the numeric strings (“5, 3, 6, 2, 9”) converted by the speech recognition unit 113, the locus coincides with the pattern PT. In this case, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. After releasing the screen lock, the pattern PT stored in the authentication code storage unit 114 may be updated through similar authentication. On the other hand, if the locus does not match the pattern PT, the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested.
  • FIG. 7 is a diagram illustrating an example of a character string CA generated by the processing unit 112 instead of a numeric string.
  • the arrow AR is drawn so as to be easy to understand, but it is not actually displayed.
  • the user US stores the pattern PT as the same authentication code
  • the stored pattern PT is applied to the character string CA displayed as shown in FIG. Is read as “C, H, G, D, E”
  • the pattern PT can be reproduced. Therefore, when the user US utters the alphabet in the form of “Shi-Eichi-Ji-Die-I”, the microphone 105 collects the sound and passes through the voice processing unit 106.
  • a voice signal is input to the voice recognition unit 113, converted into a character string “C ⁇ H ⁇ G ⁇ D ⁇ E”, and transmitted to the processing unit 112 as character string information.
  • the processing unit 112 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the numeric string NA stored in the processing unit 112 to determine whether or not the authentication codes match.
  • Other configurations are the same as those in the above-described embodiment.
  • FIG. 8 is a diagram showing an arrangement GA of vegetable or fruit patterns generated by the processing unit 112 instead of a numeric string or a character string.
  • the arrow AR is drawn so as to be easily understood, but it is not actually displayed.
  • the processing unit 112 registers image names in advance in association with individual images. The sequence of images is random.
  • the user US stores the pattern PT as the same authentication code
  • the stored pattern PT is applied to the displayed image array GA as shown in FIG.
  • the pattern PT can be reproduced by connecting images of “tomato, corn, mandarin orange, persimmon, green pepper”. Therefore, when the user US utters a word of a pattern in the form of “Tomato, corn, mandarin orange, kaki, peaman”, the microphone 105 collects the sound, and the voice recognition unit 113 via the voice processing unit 106 collects the words. Is converted to a character string “Tomato, corn, mandarin orange, kaki, piman” and transmitted to the processing unit 112 as character information. In the user's utterance, it is preferable to provide a certain silent time between words because it is easy to convert to an appropriate word when performing speech recognition.
  • the processing unit 112 selects the tomato pattern TO when the character information of “Tomato” matches the name of the tomato as the display image, and the character information of “corn” matches the name of the corn as the display image
  • the corn pattern CR when the character information of “mandarin orange” matches the name of the mandarin orange as the display image, select the pattern orange of the mandarin orange, and the character information of “oyster” is displayed as the display image. If the name matches the name of the rice cake, the image PR of the rice cake is selected, and if the character information of “Piman” matches the name of the green pepper as the display image, the image of the green pepper is selected and the selected image is selected.
  • the processing unit 112 specifies and selects an image corresponding to a character / number string from a plurality of display images, and performs processing based on the selected image.
  • FIG. 9A is a diagram showing a number string (here, “4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored as authentication code information (password that is a character / number string) in the authentication code storage unit 114. is there.
  • FIG. 9B is a diagram showing a combination image NG of numbers and images generated by the processing unit 112 together with the explanation image RG.
  • the combined image NG shown in FIG. 9B is an image in which colored blocks are arranged in 3 rows and 3 columns, and the two are associated by arranging a number in the center.
  • each color is represented by hatching or vertical and horizontal lines as shown in the adjacent explanatory image RG.
  • the numbers always include those constituting the number string stored in the authentication code storage unit 114.
  • the combination of numbers and colors is random.
  • the combination of each color and the corresponding number constitutes a predetermined relationship.
  • the explanation image RG is added to indicate the color corresponding to the block diagram, and is not actually displayed.
  • the processing unit 112 transmits information on the generated combination image NG to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the combination image NG into an image signal and transmits it to the display unit 104. Therefore, the display unit 104 displays the combination image NG (excluding the explanation image RG) shown in FIG. Can be displayed.
  • the control unit 110 performs a display requesting to input an authentication code on the display unit 104 via the display drive control unit 104DR, and FIG.
  • the combined image NG shown in b) is displayed.
  • the user US Since the user US stores a numerical string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG.
  • the color of the block corresponding to “4” is yellow
  • the color of the block corresponding to the number “9” is green
  • the color of the block corresponding to the number “1” is blue
  • the color of the block is yellow
  • the color of the block corresponding to the number “5” is blue. Therefore, when the user US utters the word “Ki / Midori / Ao / Ki / Ao”, the microphone 105 collects the sound and inputs the audio signal to the audio recognition unit 113 via the audio processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) of “Ki / Midori / Ao / Ki / Ao” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number string
  • the processing unit 112 that also serves as the conversion device reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the combined image NG generated by itself. Then, the characters “ki, green, ao, ki, ao” are picked up from the corresponding image color and converted to a character string (called a second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100.
  • the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested.
  • Other configurations are the same as those in the above-described embodiment.
  • FIG. 10A to 10 (c) are diagrams showing a combination image NG of numbers and images generated by the processing unit 112 together with an explanation image RG.
  • the processing unit 112 displays the combined image NG illustrated in FIG. 10A on the display unit 104 via the display drive control unit 104DR.
  • the combination of numbers and colors is random.
  • the user US Since the user US stores a numeric string (“1, 2, 3, 4”) as an update code, when viewing the combined image NG as shown in FIG.
  • the color of the block corresponding to is blue, the color of the block corresponding to the number “2” is red, the color of the block corresponding to the number “3” is yellow, and the block corresponding to the number “4” It can be seen that the color is yellow. Therefore, when the user US utters a word of the color “Ao / Aka / Ki / Ki”, the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / number string) “Ao, Aka, Ki, Ki” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number string
  • the processing unit 112 reads the numeric string (“1, 2, 3, 4”) stored as the update code in the authentication code storage unit 114, and generates the combined image NG (FIG. 10). Applying to (a), the characters “Ao, Aka, Ki, Ki” are picked up from the color of the corresponding image, and converted into a character string (second character / number string) arranged in this order. Further, when the first character / numeric string and the second character / numeric string match, the processing unit 112 determines that the update codes match and permits the password update. The processing unit 112 that has permitted the update of the password generates a combined image NG in which the correspondence between the numbers and the colors is changed as shown in FIG. 10B, and displays the display unit 104 via the display drive control unit 104DR. To display.
  • the color of the block corresponding to the number “9” is yellow.
  • the color of the block corresponding to the number “8” is purple
  • the color of the block corresponding to the number “5” is green
  • the color of the block corresponding to the number “6” is red.
  • the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an updated character string “Ki / Murasaki / Midori / Aka” and transmitted to the processing unit 112 as updated character string information.
  • the processing unit 112 collates with the combined image NG shown in FIG. 10B to determine that the character “ki” in the updated character string is yellow and the corresponding number is “7, 9”.
  • the character “Murasaki” indicates purple, the corresponding number is determined to be “4,8”, the character “Midori” indicates green, the corresponding number is determined to be “5”, and the character “Red” indicates red, and the corresponding numbers are determined to be “1, 6”.
  • the processing unit 112 has a plurality of password candidates that the user US desires to update, which are “7 ⁇ 4 ⁇ 5 ⁇ 1”, “9 ⁇ 4 ⁇ 5 ⁇ 1”, “7 ⁇ 8 ⁇ 5 ⁇ 1”.
  • the processing unit 112 stores these eight numeric strings as password candidates. As described above, the processing unit 112 identifies a numeric string corresponding to a character in the character string, and performs processing based on the identified numeric string.
  • the processing unit 112 newly generates a combination image NG in which the correspondence between numbers and colors is changed as illustrated in FIG. 10C and displays the combined image NG again on the display unit 104 via the display drive control unit 104DR.
  • the color of the block corresponding to the number “9” is blue, and the color of the block corresponding to the number “8” is green. It can be seen that the color of the block corresponding to the number “5” is yellow and the color of the block corresponding to the number “6” is green.
  • the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an update character string “Ao / Midori / Ki / Midori” and transmitted to the processing unit 112 as update character string information.
  • the processing unit 112 collates with the combination image NG shown in FIG. 10C to determine that the character “AO” in the updated character string is blue and the corresponding number is “2, 9”.
  • the character “Midori” indicates green, and the corresponding number is determined to be “6, 8”, the character “K” indicates yellow, and the corresponding number is determined to be “1, 5”. .
  • the processing unit 112 uses the numeric string “9, 8, 5, 6” as a new password.
  • the password in the authentication code storage unit 114 is updated. If a plurality of password candidates remain in the second process, the processing unit 112 may further display a new combined image and obtain the user's US utterance.
  • the number of colors used for the combined image is limited to five so that it can be easily distinguished. Therefore, the user US has to speak a plurality of times for password input. If the number of passwords is matched one-on-one, the user's utterance can be done only once. Alternatively, the number of images and the number of passwords can be matched on a one-to-one basis by using characters and patterns as in the following embodiment without using color types as combination images. Similarly, one utterance is sufficient.
  • FIG. 11 is a diagram showing a correspondence table CT in which characters are arranged in association with a predetermined relationship instead of colored blocks in correspondence with the numbers constituting the number string stored in the authentication code storage unit 114.
  • the correspondence table CT in FIG. 11 is displayed on the display unit 104.
  • the user US since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the correspondence table CT as shown in FIG. Is the character corresponding to the number “9”, the character corresponding to the number “9” is “no”, the character corresponding to the number “1” is “no”, and the character corresponding to the number “3”. Is “ka”, and the character corresponding to the number “5” is “ma”. Therefore, when the user US speaks hiragana “ko, ta, no, ka, or ma”, the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / numerical string) of “ko, ta, no, koma” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / numerical string
  • the processing unit 112 reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 11) generated by itself.
  • the correspondence table CT (FIG. 11) generated by itself.
  • a character string arranged in this order referred to as a second character / number string.
  • the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100.
  • Other configurations are the same as those in the above-described embodiment.
  • the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. Is converted into a character string (referred to as a third character / number string) and transmitted to the processing unit 112 as character string information.
  • the processing unit 112 compares the character string “ko, ta, no, ka, or ma” with the correspondence table CT (FIG. 11) generated by the processing unit 112 to convert the character string into numbers corresponding to hiragana, respectively. It is converted to a numeric string “4 ⁇ 9 ⁇ 1, 3 ⁇ 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • FIG. 12 shows a correspondence table CT in which vegetables or fruit patterns are arranged in association with a predetermined relationship in place of colored blocks or characters, corresponding to the numbers constituting the number string stored in the authentication code storage unit 114.
  • FIG. 12 the names of vegetables or fruits are registered in association with the displayed patterns of vegetables or fruits, but this is not always necessary.
  • the correspondence table CT shown in FIG. 12 is displayed on the display unit 104.
  • the user US since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG.
  • the pattern corresponding to "" is shiitake, the pattern corresponding to the number “9” is “green pepper”, the pattern corresponding to the number “1” is “tomato”, and the pattern corresponding to the number “3” Is “cherry” and the pattern corresponding to the number “5” is “mandarin orange”. Therefore, when the user US utters the word “Shitake / Piman / Tomato / Sakurabo / Mikan”, the microphone 105 collects the sound and inputs the speech signal to the speech recognition unit 113 via the speech processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) “Shitake / Piman / Tomato / Cherry Bamboo / Mikan” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number
  • the processing unit 112 reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 12) generated by itself. By collating with each other, the characters “Shitake / Piman / Tomato / Sakurambo / Mikan” corresponding to the numbers are picked up and converted into a character string (second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • the processing unit 112 converts the character string “Shitake / Pi-Man / Tomato / Cherry Bamboo / Tangerine” into a correspondence table CT (FIG. 11) generated by itself, thereby converting the character string into a number corresponding to the design. It is converted to a numeric string “4 ⁇ 9 ⁇ 1, 3 ⁇ 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • FIG. 13 is a flowchart showing a control operation (except for step S102) of the processing unit 112 according to the above-described embodiment.
  • the processing unit 112 as an image, includes a numeric string NA (FIG. 5), a character string CA (FIG. 7), an image arrangement GA (FIG. 8), a combined image NG (FIG. 9B), and a correspondence table CT. (FIGS. 11 and 12) are displayed.
  • step S102 When the user US who has seen the display image speaks in response to this (step S102), the voice recognition unit 113 recognizes the character / numeric string represented by the voice and performs voice recognition (step S103). 112 displays the audio processing result as shown in step S104 (see FIG. 6).
  • step S105 if the voice processing result is inappropriate (NO in step S105), the flow returns to step S102, and the same processing is repeated.
  • the processing unit 112 reads the authentication code in step S106, collates the voice recognition result and the authentication code in step S107, and the two match. Process whether or not.
  • the processing unit 112 displays a message such as “authentication failed” on the display unit 104 in step S109, Continue screen lock.
  • step S108 the processing unit 112 releases the screen lock because the authentication is successful.
  • biometric authentication can identify an individual user, there is a demerit that biometric information as a release code is difficult to share with multiple people. In other words, it can be shared by registering biometric information for a plurality of people, but it takes time to register, and it is expected that a single electronic device will be shared by a plurality of people, especially when an electronic device is used in a factory. In such a case, there is a request to increase the convenience of sharing by using a common password. According to the present embodiment, there is an advantage that it can be used with almost the same user load as conventional password authentication. Speech recognition is a user interface that can be input hands-free, and has an advantage of high affinity with HMD.
  • the HMD is mounted on the user's head, the image displayed on the display unit is difficult for others to see, so there is an advantage that confidentiality can be kept high. Even if someone remembers the character string that the user utters, if a different image (or a combination of a character / number string and an image) is displayed for each display, the learned character string will be displayed at a different timing. Even if another person speaks, the electronic device is not unlocked and strong security can be secured. In addition, there is no need for the process of displaying the input characters etc. as the hidden characters “******”, which is performed when inputting the password to the conventional electronic device. If the image displayed on the display unit is not visible to others, a code reversely converted when displaying the result of voice recognition may be displayed.
  • the HMD 100 of the present embodiment can also be used as information input means.
  • the processing unit 112 when the user US tries to make a call to a telephone number “030-1234-5678”, the processing unit 112 generates a correspondence table CT as shown in FIG. 11 in response to a request from the user US. Displayed on the display unit 104.
  • the user US utters the characters corresponding to the telephone number as “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i”.
  • the processing unit 112 uses the correspondence table CT (FIG. 11) generated by itself to process the input character string “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i” It is converted into a number string (second character / number string) of “0, 3, 0, 1, 2, 3, 4, 5, 6, 7, 8”. Further, when the processing unit 112 has a telephone function, the processing unit 112 can make a call by inputting the obtained numeric string as a telephone number. In such a case, the processing unit 112 also serves as an input device. As a result, the user US can make a hands-free call without knowing the other party's telephone number. The converted telephone number may be displayed on the display unit 104 to make a call after confirming the user US.
  • the HMD 100 may be used for inputting not only a telephone number but also a My Number, a credit card number, and the like.
  • the present invention is not limited to the embodiments described in the specification, and other embodiments and modifications are included for those skilled in the art from the embodiments and technical ideas described in the present specification. it is obvious.
  • the description and the embodiments are for illustrative purposes only, and the scope of the present invention is indicated by the following claims.
  • the present invention has been described by taking the HMD as an example.
  • the present invention is not limited to the HMD and can be applied to all electronic devices such as portable terminals. Further, some or all of the functions permitted by the authentication described above may be used.
  • the screen lock is released as an example of permitting a predetermined function by inputting a password.
  • the application can be started by appropriately inputting a password as described above in a state where a login screen at the time of starting the application is displayed on the display unit 104.
  • it can also authenticate in it. In such a case, it is desirable to change to the authentication screen during a hands-free operation using the user's utterance without using an operation to turn on the switch.

Abstract

The present invention provides an electronic apparatus, a head-mounted display, a processing method of the electronic apparatus, and a program therefor with which it is possible to suppress information from leaking to a third person who heard the utterance of a user. The electronic apparatus has a storage device for storing a pattern, a display device for displaying a plurality of images, a speech recognition device for acquiring speech uttered by a user in response to the images and converting same into a corresponding character/numeric string, and a processing device for enabling a prescribed function when the plurality of images displayed on the display device are linked in order of the character/numeric string converted by the speech recognition device and a locus thereof matches the pattern stored in the storage device.

Description

電子機器、ヘッドマウントディスプレイ、電子機器の処理方法及びそのプログラムElectronic device, head mounted display, electronic device processing method and program thereof
 本発明は,音声認識機能を備えた電子機器、ヘッドマウントディスプレイ、電子機器の処理方法及びそのプログラムに関する。 The present invention relates to an electronic device having a voice recognition function, a head mounted display, a processing method of the electronic device, and a program thereof.
 近年、急速に発達したスマートフォン等の携帯端末は、ビジネスや家庭での作業補助に用いられることも多い。一般的な携帯端末では、画像表示とユーザーインターフェースを兼ねるタッチパネル式の画面を備えているので、ユーザーはこれにタッチすることで必要な入力を行って、所望の画像を表示したり、情報を入力したりするなどの操作を行うことができる。 In recent years, mobile terminals such as smartphones that have been rapidly developed are often used for business and home work assistance. A typical mobile terminal has a touch panel screen that serves both as an image display and a user interface. By touching this screen, the user can make necessary inputs to display a desired image or input information. Can be performed.
 一方、許可を得ない第三者の使用を阻止すべく、携帯端末の画面をロックすることも通常行われている。画面がロックされた携帯端末を使用する為には、まずロックを解除しなくてはならない。しかるに、ユーザーの手がふさがっている場合など、ボタン操作を行わずに携帯端末のロックを解除したい場合がある。 On the other hand, it is also common to lock the screen of a mobile terminal in order to prevent the use of a third party without permission. To use a mobile device with a locked screen, you must first unlock it. However, there are cases where the user wants to unlock the mobile terminal without performing a button operation, such as when the user's hand is occupied.
 ボタン操作を行うことなくロックを解除しようとする場合、例えばユーザーの指紋、声紋、静脈、網膜等の生体パターンを読み取って認証を行う生体認証が考えられる。しかしながら,このような生体認証は、生体パターンを読み取るためのセンサや,パターンマッチング処理を行う為に専用のソフトウェアが必要になり、システムが複雑化しコストが増大するという問題がある。又、複数人で一つの携帯端末を共用するような場合、複数人の生体パターンをそれぞれ記憶しなくてはならず,使い勝手が悪い。 When attempting to unlock without performing a button operation, for example, biometric authentication is performed in which a biometric pattern such as a user's fingerprint, voiceprint, vein, retina, etc. is read for authentication. However, such biometric authentication requires a sensor for reading a biometric pattern and dedicated software for performing pattern matching processing, resulting in a problem that the system becomes complicated and costs increase. In addition, when a single portable terminal is shared by a plurality of people, the biometric patterns of the plurality of people must be stored, which is inconvenient.
 一方、携帯端末の中には、ハンズフリーを実現すべく音声認識機能を備えたものもある。そこで、音声認識機能を用いることでボタン操作を行うことなく携帯端末のロックを解除すれば良いという考えもある。しかしながら、音声認識を利用してロック解除を行う場合、文字列等に変換するために、まずユーザーが発話したワードをマイクで集音しなくてはならないから、ユーザーが発話した時点で周囲の者にパスワードを知られてしまう恐れがある。同様の問題は、秘匿したい電話番号等の情報を、音声認識機能を用いて携帯端末等に入力したい場合にも生じる。 On the other hand, some portable terminals have a voice recognition function to realize hands-free. Therefore, there is an idea that the mobile terminal can be unlocked without performing a button operation by using the voice recognition function. However, when unlocking using voice recognition, the words spoken by the user must first be collected with a microphone in order to convert them into character strings, etc. There is a risk that the password will be known. The same problem also occurs when it is desired to input information such as a telephone number to be concealed into a portable terminal or the like using a voice recognition function.
 これに対し、特許文献1には、あらかじめ登録された複数組のキーワードとパスワードの中からランダムに選択した1組のペアのキーワードを表示部に表示し、その表示を見たユーザーに、キーワードとペアとなるパスワードを発話させ、その音声認識結果をもとに、ユーザーから発話されたパスワードと表示したキーワードと対になるパスワードとの同一性を照合して、端末使用者本人の認証を行う技術が開示されている。 On the other hand, in Patent Document 1, a pair of keywords randomly selected from a plurality of pairs of keywords and passwords registered in advance is displayed on the display unit. A technology that authenticates the terminal user himself / herself by verifying the identity of the password spoken by the user and the displayed keyword and the paired password based on the voice recognition result by uttering the password to be paired Is disclosed.
特開2002-312318号公報JP 2002-312318 A
 しかしながら、特許文献1の技術によれば、キーワードとパスワードのペアをユーザーが覚えておかなければならず、ユーザーの負担が比較的大きいという問題がある。又、キーワードとパスワードのペアを両方登録する必要があり,その手間がかかる。一方、ペア数を少なくすれば、それに応じてユーザーの負担も少なくなるが、その代わりパスワードを知られる恐れが高まるといえる。特に、特許文献1では「正答率」についての言及があり、これはユーザーが間違えることを前提にしていることの証左であり、従って使い勝手の面から難があることが懸念されている。 However, according to the technique of Patent Document 1, there is a problem that the user must remember a keyword / password pair, and the burden on the user is relatively large. Moreover, it is necessary to register both a keyword and password pair, which is troublesome. On the other hand, if the number of pairs is reduced, the burden on the user will be reduced accordingly, but the risk that the password will be known instead increases. In particular, Patent Document 1 mentions the “correct answer rate”, which is a proof that the user is assumed to make a mistake, and there is a concern that there is difficulty in terms of usability.
 本発明は、上記の事情に鑑みてなされたものであって、音声認識機能を用いながらも、ユーザーの発話を聞いた第三者に情報が洩れることを抑制できる電子機器、ヘッドマウントディスプレイ、電子機器の処理方法及びそのプログラムを提供することを目的とする。 The present invention has been made in view of the above circumstances, and is capable of suppressing leakage of information to a third party who has heard a user's utterance while using a voice recognition function, a head mounted display, and an electronic device. It is an object of the present invention to provide a device processing method and a program thereof.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映した電子機器は、
 パターンを記憶する記憶装置と、
 複数の画像を表示する表示装置と、
 前記画像に応じてユーザーが発話した音声を取得して、対応する文字/数字列に変換する音声認識装置と、
 前記音声認識装置が変換した文字/数字列の順序で、前記表示装置に表示された前記複数の画像を繋いでいったとき、その軌跡が前記記憶装置に記憶された前記パターンと一致したときは、所定の機能を許可する処理装置とを有するものである。
In order to achieve at least one of the above objects, an electronic device reflecting one aspect of the present invention is provided.
A storage device for storing patterns;
A display device for displaying a plurality of images;
A voice recognition device that acquires voice spoken by a user according to the image and converts the voice into a corresponding character / number string;
When the plurality of images displayed on the display device are connected in the order of the character / number string converted by the voice recognition device, and the locus matches the pattern stored in the storage device And a processing device that permits a predetermined function.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映した別の電子機器は、
 文字/数字列からなるパスワードを記憶する記憶装置と、
 少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換する音声認識装置と、
 前記パスワードを、前記所定の関係に従って第2の文字/数字列に変換する変換装置と、
 前記音声認識装置が変換した前記第1の文字/数字列と、前記変換装置が変換した前記第2の文字/数字列が一致したときは、所定の機能を許可する処理装置とを有するものである。
In order to achieve at least one of the objects described above, another electronic device reflecting one aspect of the present invention is:
A storage device for storing a password consisting of a character / number string;
A display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship;
A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string;
A conversion device for converting the password into a second character / number string in accordance with the predetermined relationship;
A processing device that permits a predetermined function when the first character / numeric string converted by the voice recognition device matches the second character / numeric string converted by the conversion device; is there.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映したさらにべつの電子機器は、
 文字/数字列からなるパスワードを記憶する記憶装置と、
 少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第3の文字/数字列に変換する音声認識装置と、
 前記第3の文字/数字列を、前記所定の関係に従って第4の文字/数字列に変換し、前記第4の文字/数字列と前記パスワードが一致したときは、所定の機能を許可する処理装置とを有するものである。
In order to achieve at least one of the above-described objects, another electronic device reflecting one aspect of the present invention is:
A storage device for storing a password consisting of a character / number string;
A display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship;
A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a third letter / number string;
Processing for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password And a device.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映したさらに別の電子機器は、
 規定の文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換する音声認識装置と、
 前記音声認識装置が変換した第1の文字/数字列を、前記所定の関係に従って第2の文字/数字列に変換する変換装置と、
 前記規定の文字/数字列として、前記第2の文字/数字列を入力する入力装置と、を有するものである。
In order to achieve at least one of the above objects, still another electronic device reflecting one aspect of the present invention is:
A display device that displays letters / numbers constituting a prescribed letter / number string in association with a plurality of images according to a predetermined relationship;
A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string;
A conversion device that converts the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
And an input device for inputting the second character / number string as the prescribed character / number string.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映したさらに別の電子機器は、
 複数の画像を表示する表示装置と、
 マイクと、
 前記マイクを通して取得した音声を解析し、当該音声が表す文字/数字列を認識する音声認識装置と、
 前記複数の画像の中で、前記文字/数字列に対応する画像を特定し、前記特定された画像に基づく処理を行う処理部と、を有するものである。
In order to achieve at least one of the above objects, still another electronic device reflecting one aspect of the present invention is:
A display device for displaying a plurality of images;
With a microphone,
A voice recognition device that analyzes voice acquired through the microphone and recognizes a character / number string represented by the voice;
A processing unit that specifies an image corresponding to the character / number string among the plurality of images and performs processing based on the specified image.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映した電子機器の処理方法は、
 パターンを記憶し、
 複数の画像を表示し、
 前記画像に応じてユーザーが発話した音声を取得して、対応する文字/数字列に変換し、
 変換された文字/数字列の順序で、表示された前記複数の画像を繋いでいったとき、その軌跡が前記記憶装置に記憶された前記パターンと一致したときは、所定の認証を行うものである。
In order to achieve at least one of the objects described above, a processing method of an electronic device reflecting one aspect of the present invention is as follows.
Remember the pattern,
Display multiple images,
Acquire the voice spoken by the user according to the image, convert it to the corresponding character / number string,
When the plurality of displayed images are connected in the order of the converted character / numerical string, when the locus coincides with the pattern stored in the storage device, predetermined authentication is performed. is there.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映した別の電子機器の処理方法は、
 文字/数字列からなるパスワードを記憶し、
 少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換し、
 前記パスワードを、前記所定の関係に従って第2の文字/数字列に変換し、
 前記第1の文字/数字列と前記第2の文字/数字列が一致したときは、所定の認証を行うものである。
In order to realize at least one of the above-described objects, another electronic device processing method reflecting one aspect of the present invention includes:
Memorize passwords consisting of letters and numbers,
Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string;
Converting the password into a second character / number string according to the predetermined relationship;
When the first character / numeric string matches the second character / numeric string, predetermined authentication is performed.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映したさらに別の電子機器の処理方法は、
 文字/数字列からなるパスワードを記憶し、
 少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第3の文字/数字列に変換し、
 前記第3の文字/数字列を、前記所定の関係に従って第4の文字/数字列に変換し、前記第4の文字/数字列と前記パスワードが一致したときは、所定の認証を行うものである。
In order to achieve at least one of the above objects, still another electronic device processing method reflecting one aspect of the present invention is as follows.
Memorize passwords consisting of letters and numbers,
Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
Obtaining voice spoken by the user according to the image associated with the letters / numbers, and converting the voice into a third letter / number string;
The third character / number string is converted into a fourth character / number string according to the predetermined relationship, and when the fourth character / number string matches the password, predetermined authentication is performed. is there.
 上述した目的のうち少なくとも一つを実現するために、本発明の一側面を反映したさらに別の電子機器の処理方法は、
 規定の文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
 前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換し、
 前記音声認識装置が変換した第1の文字/数字列を、前記所定の関係に従って第2の文字/数字列に変換し、
 前記規定の文字/数字列として、前記第2の文字/数字列を入力するものである。
In order to achieve at least one of the above objects, still another electronic device processing method reflecting one aspect of the present invention is as follows.
The letters / numbers that make up the prescribed letter / number string are displayed in association with a plurality of images according to a predetermined relationship,
Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string;
Converting the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
The second character / number string is input as the prescribed character / number string.
 本発明によれば、音声認識機能を用いながらも、ユーザーの発話を聞いた第三者に情報が洩れることを抑制できる電子機器、ヘッドマウントディスプレイ、電子機器の処理方法及びそのプログラムを提供することができる。 According to the present invention, it is possible to provide an electronic device, a head-mounted display, a method for processing an electronic device, and a program thereof that can prevent information from leaking to a third party who has heard a user's speech while using a voice recognition function. Can do.
本実施形態にかかる、電子機器であるヘッドマウントディスプレイ(以下、HMDという)100をユーザーが装着した状態で示す正面図である。1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state in which a user is wearing it. ディスプレイユニット104の構成を示す概略断面図である。3 is a schematic cross-sectional view showing a configuration of a display unit 104. FIG. 本実施形態にかかるHMD100のブロック図である。It is a block diagram of HMD100 concerning this embodiment. 認証コード記憶部114に認証コード情報として記憶されるパターンPTを示す図である。It is a figure which shows the pattern PT memorize | stored in the authentication code memory | storage part 114 as authentication code information. 処理部112で生成される数字列NAを模式的に示す図である。It is a figure which shows typically the numerical string NA produced | generated by the process part 112. FIG. 音声認識が適切に行われたかを確認する為に表示されるメッセージを示す図である。It is a figure which shows the message displayed in order to confirm whether speech recognition was performed appropriately. 数字列の代わりに処理部112で生成される文字列CAの例を示す図である。It is a figure which shows the example of the character string CA produced | generated by the process part 112 instead of a number string. 数字列や文字列の代わりに処理部112で生成される野菜又は果物の絵柄の並びGAを示す図である。It is a figure which shows the arrangement | sequence GA of the pattern of the vegetable or the fruit produced | generated by the process part 112 instead of a numerical string or a character string. (a)は、認証コード記憶部114に認証コード情報(パスワード)として記憶される数字列(ここでは「4・9・1・3・5」)を示す図である。(b)は、処理部112で生成される数字と画像の組み合わせ画像NGを、説明画像RGと共に示す図である。(A) is a figure which shows the numerical sequence (here "4 * 9 * 1 * 3 * 5") memorize | stored as authentication code information (password) in the authentication code memory | storage part 114. FIG. (B) is a figure which shows the combination image NG of the number and image produced | generated by the process part 112 with the description image RG. 処理部112で生成される数字と画像の組み合わせ画像NGを、説明画像RGと共に示す図である。It is a figure which shows the combination image NG of the number and image produced | generated by the process part 112 with description image RG. 認証コード記憶部114に記憶された数字列を構成する数字に対応して、有色ブロックの代わりに文字を対応づけて配置した対応表CTを示す図である。It is a figure which shows the corresponding | compatible table CT which matched and arranged the character instead of the colored block corresponding to the number which comprises the number sequence memorize | stored in the authentication code memory | storage part. 認証コード記憶部114に記憶された数字列を構成する数字に対応して、有色ブロック又は文字の代わりに野菜又は果物の絵柄を対応づけて配置した対応表CTを示す図である。It is a figure which shows the corresponding | compatible table CT arrange | positioned corresponding to the number which comprises the number sequence memorize | stored in the authentication code memory | storage part 114, and matched the pattern of vegetables or fruits instead of a colored block or a character. 本実施形態にかかる処理部112の制御動作(ただし、ステップS102を除く)を示すフローチャートである。It is a flowchart which shows the control operation (however, except step S102) of the process part 112 concerning this embodiment.
 以下に、本発明の実施形態を、図面を参照して説明する。図1は、本実施形態にかかる、電子機器であるヘッドマウントディスプレイ(以下、HMDという)100をユーザーが装着した状態で示す正面図である。以下、HMD100の右側及左側とは、HMD100を装着したユーザーにとっての右側及び左側をいうものとする。 Embodiments of the present invention will be described below with reference to the drawings. FIG. 1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state where a user wears it. Hereinafter, the right side and the left side of the HMD 100 refer to the right side and the left side for the user wearing the HMD 100.
 図1に示す本実施形態のHMD100において,ユーザーUSの頭部に装着されるフレーム101は、2つの眼鏡レンズ102をユーザーUSの眼前に配置している。右側(ユーザーの利き目などに応じて左側でもよい)の眼鏡レンズ102の上部において、円筒状の主本体部103が固定されている。主本体部103にはディスプレイユニット104が設けられている。主本体部103内には、ディスプレイユニット104の表示制御を司る表示駆動制御部104DR(後述する図3を参照)が配置されている。なお、必要であれば両眼の前にそれぞれディスプレイユニットを配置してもよい。 In the HMD 100 of the present embodiment shown in FIG. 1, the frame 101 to be mounted on the head of the user US has two spectacle lenses 102 arranged in front of the user US. A cylindrical main body 103 is fixed on the upper part of the spectacle lens 102 on the right side (which may be on the left side according to the user's dominant eye). The main body 103 is provided with a display unit 104. A display drive control unit 104DR (see FIG. 3 described later) that controls display control of the display unit 104 is disposed in the main body unit 103. If necessary, a display unit may be arranged in front of both eyes.
 図2は、ディスプレイユニット104の構成を示す概略断面図である。表示装置であるディスプレイユニット104は、画像形成部104Aと画像表示部104Bとからなる。画像形成部104Aは、主本体部103内に組み込まれており、光源104aと、一方向拡散板104bと、集光レンズ104cと、表示素子104dとを有している。一方、いわゆるシースルー型の表示部材である画像表示部104Bは、主本体部103から下方に向かい、片方の眼鏡レンズ102(図1参照)に平行に延在するように配置された全体的に板状であって、接眼プリズム104fと、偏向プリズム104gと、ホログラム光学素子104hとを有している。 FIG. 2 is a schematic cross-sectional view showing the configuration of the display unit 104. The display unit 104 as a display device includes an image forming unit 104A and an image display unit 104B. The image forming unit 104A is incorporated in the main body unit 103, and includes a light source 104a, a unidirectional diffuser 104b, a condenser lens 104c, and a display element 104d. On the other hand, the image display unit 104B, which is a so-called see-through type display member, is disposed on the entire plate so as to extend downward from the main body unit 103 and extend in parallel to one eyeglass lens 102 (see FIG. 1). The eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h.
 次に、ディスプレイユニット104の動作について説明する。光源104aから出射された光は、一方向拡散板104bにて拡散され、集光レンズ104cにて集光されて表示素子104dに入射する。表示素子104dに入射した光は、表示駆動制御部104DRから入力された画像データに基づいて画素ごとに変調され、画像光として出射される。これにより、表示素子104dにはカラー画像が表示される。 Next, the operation of the display unit 104 will be described. The light emitted from the light source 104a is diffused by the unidirectional diffusion plate 104b, condensed by the condenser lens 104c, and enters the display element 104d. The light incident on the display element 104d is modulated for each pixel based on the image data input from the display drive control unit 104DR, and is emitted as image light. As a result, a color image is displayed on the display element 104d.
 表示素子104dからの画像光は、接眼プリズム104fの内部にその基端面PL1から入射し、内側面PL2と外側面PL3で複数回全反射されて、ホログラム光学素子104hに入射する。ホログラム光学素子104hに入射した光は、そこで反射され、内側面PL2を透過して瞳孔Bに達する。瞳孔Bの位置では、ユーザーは、表示素子104dに表示された画像の拡大虚像を観察することができ、画像表示部104Bに形成される画面として視認することができる。この場合、ホログラム光学素子104hが画面を構成しているとみることもできるし、内側面PL2に画面が形成されているとみることもできる。なお、本明細書において「画面」というときは、表示される画像を指すこともある。 Image light from the display element 104d enters the eyepiece prism 104f from its base end face PL1, is totally reflected a plurality of times by the inner side face PL2 and the outer side face PL3, and enters the hologram optical element 104h. The light incident on the hologram optical element 104h is reflected there, passes through the inner side surface PL2, and reaches the pupil B. At the position of the pupil B, the user can observe an enlarged virtual image of the image displayed on the display element 104d, and can visually recognize it as a screen formed on the image display unit 104B. In this case, it can be considered that the hologram optical element 104h constitutes a screen, or it can be considered that a screen is formed on the inner surface PL2. In the present specification, “screen” may refer to an image to be displayed.
 一方、接眼プリズム104f、偏向プリズム104gおよびホログラム光学素子104hは、外光をほとんど全て透過させるので、ユーザーはこれらを介して外界像(実像)を観察することができる。したがって、表示素子104dに表示された画像の虚像は、外界像の一部に重なって観察されることになる。このようにして、HMD100のユーザーは、ホログラム光学素子104hを介して、表示素子104dから提供される画像と外界像とを同時に観察することができる。尚、ディスプレイユニット104が非表示状態のとき画像表示部104Bは素通しとなり、外界像のみを観察できる。なお、本例では、光源と液晶表示素子と光学系とを組み合わせて表示ユニットを構成しているが、光源と液晶表示素子の組合せに代えて、自発光型の表示素子(例えば、有機EL表示素子)を用いても良い。また、光源と液晶表示素子と光学系の組合せに代えて、非発光状態で透過性を有する透過型有機EL表示パネルを用いてもよい。いずれにしても、画像表示部104Bに対向するユーザーの眼の視野に入るように、好ましくは、有効視野に少なくとも一部が重なるように、画面を配置すると、ユーザーは画像を容易に視認することができる。 On the other hand, since the eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h transmit almost all of the external light, the user can observe an external field image (real image) through them. Therefore, the virtual image of the image displayed on the display element 104d is observed so as to overlap with a part of the external image. In this manner, the user of the HMD 100 can simultaneously observe the image provided from the display element 104d and the external image via the hologram optical element 104h. Note that when the display unit 104 is in the non-display state, the image display unit 104B is transparent, and only the external image can be observed. In this example, a display unit is configured by combining a light source, a liquid crystal display element, and an optical system. However, instead of a combination of a light source and a liquid crystal display element, a self-luminous display element (for example, an organic EL display) is used. Element) may be used. Further, instead of a combination of a light source, a liquid crystal display element, and an optical system, a transmissive organic EL display panel having transparency in a non-light emitting state may be used. In any case, when the screen is arranged so as to fall within the visual field of the user's eye facing the image display unit 104B, and preferably at least partially overlaps the effective visual field, the user can easily visually recognize the image. Can do.
 図3は、本実施形態にかかるHMD100のブロック図であり、ユーザーUSと共に示している。HMD100は、上述したディスプレイユニット(表示部)104と、ユーザーUSの発話した音声を集音して信号に変換するマイク105と、マイク105から出力された信号を処理して音声信号として出力する音声処理部106と、音声処理部106から出力された音声信号を受信する制御部110とを有する。 FIG. 3 is a block diagram of the HMD 100 according to the present embodiment, which is shown together with the user US. The HMD 100 collects the above-described display unit (display unit) 104, the microphone 105 that collects the voice spoken by the user US and converts it into a signal, and the voice that processes the signal output from the microphone 105 and outputs it as a voice signal. The processing unit 106 and the control unit 110 that receives the audio signal output from the audio processing unit 106 are included.
 制御部110は、音声処理部106から出力された音声信号を入力して、その音声が表す文字/数字列を認識し、対応する文字/数字に変換する音声認識部113と、音声認識部113から出力された文字/数字を処理する処理部(処理装置)112と、処理部112からの信号を受信して、ディスプレイユニット104を駆動制御する表示駆動制御部104DRと、認証コード(ここではパターン又はパスワード)を記憶する認証コード記憶部(記憶装置)114とを有する。マイク105,音声処理部106,及び音声認識部113が音声認識装置を構成する。尚、本明細書で文字/数字というときは、文字及び数字の少なくとも一方をいい、文字/数字列というときは、文字又は数字を複数個並べたもの(文字のみ又は数字のみの場合を含む)をいい、画像とは文字/数字を含むものである。ワードとは、一文字のものも含む。 The control unit 110 receives the voice signal output from the voice processing unit 106, recognizes a character / number string represented by the voice, and converts the character / number string to the corresponding character / number, and the voice recognition unit 113. A processing unit (processing device) 112 that processes characters / numbers output from the display unit, a display drive control unit 104DR that receives signals from the processing unit 112 and drives and controls the display unit 104, and an authentication code (here, a pattern) Or an authentication code storage unit (storage device) 114 that stores a password. The microphone 105, the voice processing unit 106, and the voice recognition unit 113 constitute a voice recognition device. In this specification, when referring to letters / numbers, it means at least one of letters and numbers, and when referring to letters / number strings, a plurality of letters or numbers are arranged (including only letters or numbers only). An image includes letters / numbers. A word includes a single character.
(第1の実施形態)
 次に、電子機器の処理方法にかかる本発明の実施形態を説明する。尚、以下に述べる電子機器の処理方法は全て、制御部110をコンピュータと位置づけて、それに内蔵されたプログラムを実行することで実現できる。まず第1の実施形態にかかるHMD100の動作について説明する。以下の実施形態で、ユーザーUSはパターン又はパスワードを記憶しており、ディスプレイユニット104に表示された複数の画像を表すワードを、パターン又はパスワードに沿って順に発話するというHMD100の使用ルールを、マニュアル等により予め知っているものとする。図4は、認証コード記憶部114に認証コード情報として記憶されるパターンPTを示す図である。かかるパターンPTは、左から右へ向かう水平ラインと、上から下へ向かう垂直ラインと、左から右へ向かう水平ラインとを、この順序で繋いだものである。パターンPTは、ユーザーUSが予め記憶しているものとする。ここで、「パターン」とは、幾何学的形状をトレースしたものをいい、例えば一筆書きの軌跡などが相当する。
(First embodiment)
Next, an embodiment of the present invention according to a processing method of an electronic device will be described. Note that all the processing methods of the electronic device described below can be realized by positioning the control unit 110 as a computer and executing a program incorporated therein. First, the operation of the HMD 100 according to the first embodiment will be described. In the following embodiment, the user US stores a pattern or a password, and the usage rule of the HMD 100 in which words representing a plurality of images displayed on the display unit 104 are uttered in order along the pattern or the password is manually It is assumed that the information is known beforehand. FIG. 4 is a diagram showing a pattern PT stored as authentication code information in the authentication code storage unit 114. The pattern PT is formed by connecting a horizontal line from left to right, a vertical line from top to bottom, and a horizontal line from left to right in this order. The pattern PT is assumed to be stored in advance by the user US. Here, the “pattern” refers to a traced geometric shape, and corresponds to, for example, a one-stroke stroke.
 図5は、処理部112で生成される数字列NAを模式的に示す図である。但し、図5において、矢印ARを理解しやすいように重ねて描画しているが、実際は表示されることはない。ここで数字列NAは、例えば処理部112で発生させたランダムな数字を3行3列に割り振ったものである。 FIG. 5 is a diagram schematically showing a numeric string NA generated by the processing unit 112. However, in FIG. 5, the arrow AR is drawn so as to be easily understood, but it is not actually displayed. Here, the numerical string NA is, for example, a random number generated by the processing unit 112 assigned to 3 rows and 3 columns.
 処理部112は、数字列NAの情報を表示駆動制御部104DRに送信する。すると表示駆動制御部104DRは、数字列NAの情報を画像信号に変換して、ディスプレイユニット104に送信するので、ディスプレイユニット104は図5に示す数字列NAを表示することができる。 The processing unit 112 transmits information on the numeric string NA to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the numeric string NA into an image signal and transmits it to the display unit 104, so that the display unit 104 can display the numeric string NA shown in FIG.
 ユーザーUSは、幾何的なパターンPTを覚える必要はあるが、暗号などとは異なり、文字列、数字列、画像など表示される個々の要素の配列順序等を予め覚える必要がなく、覚えているパターンPTに沿って表示された要素を順に発話するだけで良く、利便性が高いといえる。秘匿性を高めるためには、個々の要素はランダムに表示することが好ましい。表示する要素としては、数字、文字(アルファベット、ひらがな、カタカナ、漢字・・・)、色、絵柄でもよい。これらを総称して画像という。表示する要素は1種類(例えばひらがなのみ)だけでもよいし、複数種類(カタカナと数字など)を混在させてもよい。但し、要素は音声認識で照合されるため、読み方が登録されていることが好ましい。絵柄の場合には、予め処理部112に絵柄とその読みが対応づけて登録され、ユーザーUSも絵柄とその登録されている読みを覚えておくと好ましい。ひとつの絵柄に対して読みは複数登録されていてもよい。ただし、複数の絵柄で同じ読みを持つもの(橋、箸など)は使用を避けることが望ましい。 Although it is necessary for the user US to remember the geometric pattern PT, unlike the encryption, the user US does not need to remember the arrangement order of individual elements to be displayed such as character strings, numeric strings, and images in advance. It is only necessary to utter the elements displayed along the pattern PT in order, which can be said to be highly convenient. In order to improve confidentiality, it is preferable to display individual elements at random. Elements to be displayed may be numbers, letters (alphabet, hiragana, katakana, kanji ...), colors, and patterns. These are collectively called images. Only one type (for example, only hiragana) may be displayed, or a plurality of types (such as katakana and numbers) may be mixed. However, since the elements are collated by voice recognition, it is preferable that the reading method is registered. In the case of a pattern, it is preferable that the pattern and its reading are registered in advance in the processing unit 112 and the user US also remembers the pattern and its registered reading. A plurality of readings may be registered for one pattern. However, it is desirable to avoid using multiple pictures with the same reading (bridge, chopsticks, etc.).
 認証前の状態では、HMD100は画面がロックされているものとする。まず、ユーザーUSがHMD100の不図示のスイッチをオン操作すると、制御部110は、表示駆動制御部104DRを介してディスプレイユニット104に、認証コードを入力することを求める表示(不図示)を行うと共に、図5に示す数字列NAを表示する。 Suppose that the screen of the HMD 100 is locked before authentication. First, when the user US turns on an unillustrated switch of the HMD 100, the control unit 110 performs a display (not illustrated) for requesting an authentication code to be input to the display unit 104 via the display drive control unit 104DR. The numeric string NA shown in FIG. 5 is displayed.
 ユーザーUSは、認証コードとしてのパターンPTを記憶しているので、図5に示すように表示された数字列NAに、記憶したパターンPTを当てはめたとき、矢印ARに示す順序で数字を「5・3・6・2・9」と読んで行くと、パターンPTを再現できることが分かる。そこで、ユーザーUSが「ごー・さん・ろく・にー・きゅう」とパターンに沿う形で数字を発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「5・3・6・2・9」という数字列に変換し、数字列情報として処理部112に送信する。 Since the user US stores the pattern PT as the authentication code, when the stored pattern PT is applied to the numerical string NA displayed as shown in FIG. 5, the numbers “5” are entered in the order indicated by the arrow AR. If you read “3, 6, 2, 9”, you can see that the pattern PT can be reproduced. Therefore, when the user US utters a number in the form of “Go, San, Roku, Ni, Kyu”, the microphone 105 collects the sound, and the speech recognition unit 113 via the speech processing unit 106 collects this. Is converted into a numeric string “5, 3, 6, 2, 9” and transmitted to the processing unit 112 as numeric string information.
 このとき、処理部112は、表示駆動制御部104DRを介してディスプレイユニット104に、図6に示すような表示を行わせて、入力された数字列がユーザーUSの意図通りのものか、確認するようにしても良い。図6の表示がなされている間、ユーザーUSが「いいえ」と発話すると、マイク105から音声処理部106を介して音声信号が処理部112に入力され、表示している数字列が不適切なものであると処理部112は判断し、再度ユーザーUSの発話を求める。一方、ユーザーUSが「はい」と発話すると、表示している数字列が適切なものであると処理部112は判断し、以降の処理を行う。 At this time, the processing unit 112 causes the display unit 104 to perform a display as shown in FIG. 6 via the display drive control unit 104DR, and confirms whether the input number string is as intended by the user US. You may do it. When the user US speaks “No” while the display of FIG. 6 is being performed, a voice signal is input from the microphone 105 via the voice processing unit 106 to the processing unit 112, and the displayed number string is inappropriate. The processing unit 112 determines that it is a thing and requests the utterance of the user US again. On the other hand, when the user US speaks “Yes”, the processing unit 112 determines that the displayed number string is appropriate, and performs the subsequent processing.
 音声認識部113から数字列情報を受信した処理部112は、認証コード記憶部114に記憶されたパターンPTを読み出して、自身が記憶していた数字列NAに当てはめる。より具体的には、音声認識部113が変換した数字列(「5・3・6・2・9」)の順序で、数字列NAを繋いでいったとき、その軌跡がパターンPTと一致したときは、処理部112は認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。画面ロックを解除後、同様な認証を経て認証コード記憶部114に記憶されたパターンPTを更新しても良い。一方、当該軌跡がパターンPTと一致しなければ、処理部112は認証コードが不一致であると判断し、HMD100の画面ロックを続行する。このとき、新たな認証コードの入力を求めるようにしても良い。 The processing unit 112 that has received the number string information from the voice recognition unit 113 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the number string NA stored in itself. More specifically, when the numeric strings NA are connected in the order of the numeric strings (“5, 3, 6, 2, 9”) converted by the speech recognition unit 113, the locus coincides with the pattern PT. In this case, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. After releasing the screen lock, the pattern PT stored in the authentication code storage unit 114 may be updated through similar authentication. On the other hand, if the locus does not match the pattern PT, the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested.
 図7は、数字列の代わりに処理部112で生成される文字列CAの例を示す図である。但し、図7において、矢印ARを理解しやすいように重ねて描画しているが、実際は表示されることはない。ユーザーUSが、同じ認証コードとしてのパターンPTを記憶しているものとして、図7に示すように表示された文字列CAに、記憶したパターンPTを当てはめたとき、矢印ARに示す順序で英文字を「C・H・G・D・E」と読んで行くと、パターンPTを再現できることが分かる。そこで、ユーザーUSが「しー・えいち・じー・でぃー・いー」とパターンに沿う形でアルファベットを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「C・H・G・D・E」という文字列に変換し、文字列情報として処理部112に送信する。処理部112は、同様に認証コード記憶部114に記憶されたパターンPTを読み出して、自身が記憶していた数字列NAに当てはめて、認証コードが一致するか否か判断できる。
それ以外の構成は、上述した実施形態と同様である。
FIG. 7 is a diagram illustrating an example of a character string CA generated by the processing unit 112 instead of a numeric string. However, in FIG. 7, the arrow AR is drawn so as to be easy to understand, but it is not actually displayed. Assuming that the user US stores the pattern PT as the same authentication code, when the stored pattern PT is applied to the character string CA displayed as shown in FIG. Is read as “C, H, G, D, E”, it can be seen that the pattern PT can be reproduced. Therefore, when the user US utters the alphabet in the form of “Shi-Eichi-Ji-Die-I”, the microphone 105 collects the sound and passes through the voice processing unit 106. A voice signal is input to the voice recognition unit 113, converted into a character string “C · H · G · D · E”, and transmitted to the processing unit 112 as character string information. Similarly, the processing unit 112 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the numeric string NA stored in the processing unit 112 to determine whether or not the authentication codes match.
Other configurations are the same as those in the above-described embodiment.
 図8は、数字列や文字列の代わりに処理部112で生成される野菜又は果物の絵柄の並びGAを示す図である。但し、図8において、矢印ARを理解しやすいように重ねて描画しているが、実際は表示されることはない。ここで、処理部112は、個々の画像に対応づけて画像の名称を予め登録しておくものとする。画像の並びはランダムである。 FIG. 8 is a diagram showing an arrangement GA of vegetable or fruit patterns generated by the processing unit 112 instead of a numeric string or a character string. However, in FIG. 8, the arrow AR is drawn so as to be easily understood, but it is not actually displayed. Here, it is assumed that the processing unit 112 registers image names in advance in association with individual images. The sequence of images is random.
 ユーザーUSが、同じ認証コードとしてのパターンPTを記憶しているものとして、図8に示すように表示された画像の並びGAに、記憶したパターンPTを当てはめたとき、矢印ARに示す順序で「トマト・トウモロコシ・ミカン・柿・ピーマン」の画像を繋いで行くと、パターンPTを再現できることが分かる。そこで、ユーザーUSが「とまと・とうもろこし・みかん・かき・ぴーまん」とパターンに沿う形で絵柄のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「とまと・とうもろこし・みかん・かき・ぴーまん」という文字列に変換し、文字情報として処理部112に送信する。尚、ユーザーの発話において、ワード間に一定の無音声時間を設けると、音声認識を行う際に適切なワードに変換しやすいので好ましい。 Assuming that the user US stores the pattern PT as the same authentication code, when the stored pattern PT is applied to the displayed image array GA as shown in FIG. It can be seen that the pattern PT can be reproduced by connecting images of “tomato, corn, mandarin orange, persimmon, green pepper”. Therefore, when the user US utters a word of a pattern in the form of “Tomato, corn, mandarin orange, kaki, peaman”, the microphone 105 collects the sound, and the voice recognition unit 113 via the voice processing unit 106 collects the words. Is converted to a character string “Tomato, corn, mandarin orange, kaki, piman” and transmitted to the processing unit 112 as character information. In the user's utterance, it is preferable to provide a certain silent time between words because it is easy to convert to an appropriate word when performing speech recognition.
 処理部112は、「とまと」の文字情報が,表示画像としてのトマトの名称に一致する場合にトマトの絵柄TOを選択し、「とうもろこし」の文字情報が,表示画像としてのトウモロコシの名称に一致する場合にトウモロコシの絵柄CRを選択し、「みかん」の文字情報が,表示画像としてのミカンの名称に一致する場合にミカンの絵柄ORを選択し、「かき」の文字情報が,表示画像としての柿の名称に一致する場合に柿の絵柄PRを選択し、「ぴーまん」の文字情報が,表示画像としてのピーマンの名称に一致する場合にピーマンの絵柄PMを選択して、選択された画像をこの順序でつなぎ合わせていったとき、その軌跡がパターンPTと一致すれば、認証コードが一致したものと判断するのである。それ以外の構成は、上述した実施形態と同様である。以上のように、処理部112は、複数の表示画像の中から文字/数字列に対応する画像を特定し選択し、その選択された画像に基づく処理を行う。 The processing unit 112 selects the tomato pattern TO when the character information of “Tomato” matches the name of the tomato as the display image, and the character information of “corn” matches the name of the corn as the display image When selecting the corn pattern CR, when the character information of “mandarin orange” matches the name of the mandarin orange as the display image, select the pattern orange of the mandarin orange, and the character information of “oyster” is displayed as the display image. If the name matches the name of the rice cake, the image PR of the rice cake is selected, and if the character information of “Piman” matches the name of the green pepper as the display image, the image of the green pepper is selected and the selected image is selected. Are connected in this order, if the trajectory matches the pattern PT, it is determined that the authentication codes match. Other configurations are the same as those in the above-described embodiment. As described above, the processing unit 112 specifies and selects an image corresponding to a character / number string from a plurality of display images, and performs processing based on the selected image.
(第2の実施形態)
 次に、第2の実施形態にかかるHMD100の動作について説明する。図9(a)は、認証コード記憶部114に認証コード情報(文字/数字列であるパスワード)として記憶される数字列(ここでは「4・9・1・3・5」)を示す図である。図9(b)は、処理部112で生成される数字と画像の組み合わせ画像NGを、説明画像RGと共に示す図である。図9(b)に示す組み合わせ画像NGは、有色のブロックを3行3列に並べたものであり、その中央に数字を配することで両者の対応付けを行っている。但し、各色は隣の説明画像RGに示すように、ハッチング又は縦横線で表している。数字は、認証コード記憶部114に記憶される数字列を構成するものを必ず含んでいる。但し、数字と色の組み合わせはランダムである。各色と,それに対応する数字との組み合わせが、所定の関係を構成している。尚、説明画像RGは、ブロックの線図に対応した色を示す為に付記されたものであり、実際には表示されない。
(Second Embodiment)
Next, the operation of the HMD 100 according to the second embodiment will be described. FIG. 9A is a diagram showing a number string (here, “4 · 9 · 1 · 3 · 5”) stored as authentication code information (password that is a character / number string) in the authentication code storage unit 114. is there. FIG. 9B is a diagram showing a combination image NG of numbers and images generated by the processing unit 112 together with the explanation image RG. The combined image NG shown in FIG. 9B is an image in which colored blocks are arranged in 3 rows and 3 columns, and the two are associated by arranging a number in the center. However, each color is represented by hatching or vertical and horizontal lines as shown in the adjacent explanatory image RG. The numbers always include those constituting the number string stored in the authentication code storage unit 114. However, the combination of numbers and colors is random. The combination of each color and the corresponding number constitutes a predetermined relationship. The explanation image RG is added to indicate the color corresponding to the block diagram, and is not actually displayed.
 処理部112は、生成した組み合わせ画像NGの情報を表示駆動制御部104DRに送信する。すると表示駆動制御部104DRは、組み合わせ画像NGの情報を画像信号に変換して、ディスプレイユニット104に送信するので、ディスプレイユニット104は図9(b)に示す組み合わせ画像NG(説明画像RGは除く)を表示することができる。 The processing unit 112 transmits information on the generated combination image NG to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the combination image NG into an image signal and transmits it to the display unit 104. Therefore, the display unit 104 displays the combination image NG (excluding the explanation image RG) shown in FIG. Can be displayed.
 認証前の状態では、HMD100は画面がロックされているものとする。まず、ユーザーUSがHMD100の不図示のスイッチをオン操作すると、制御部110は、表示駆動制御部104DRを介してディスプレイユニット104に、認証コードを入力することを求める表示を行うと共に、図9(b)に示す組み合わせ画像NGを表示する。 Suppose that the screen of the HMD 100 is locked before authentication. First, when the user US turns on a switch (not shown) of the HMD 100, the control unit 110 performs a display requesting to input an authentication code on the display unit 104 via the display drive control unit 104DR, and FIG. The combined image NG shown in b) is displayed.
 ユーザーUSは、認証コードとしての数字列(「4・9・1・3・5」)を記憶しているので、図9(b)に示すような組み合わせ画像NGを見たときに、数字「4」に対応するブロックの色が黄色であり、数字「9」に対応するブロックの色が緑色であり、数字「1」に対応するブロックの色が青色であり、数字「3」に対応するブロックの色が黄色であり、数字「5」に対応するブロックの色が青色であることが分かる。そこで、ユーザーUSが「き・みどり・あお・き・あお」と色のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「き・みどり・あお・き・あお」という文字列(第1の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 Since the user US stores a numerical string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG. The color of the block corresponding to “4” is yellow, the color of the block corresponding to the number “9” is green, the color of the block corresponding to the number “1” is blue, and corresponds to the number “3”. It can be seen that the color of the block is yellow and the color of the block corresponding to the number “5” is blue. Therefore, when the user US utters the word “Ki / Midori / Ao / Ki / Ao”, the microphone 105 collects the sound and inputs the audio signal to the audio recognition unit 113 via the audio processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) of “Ki / Midori / Ao / Ki / Ao” and transmitted to the processing unit 112 as character string information.
 これと並行して、変換装置を兼ねる処理部112は、認証コード記憶部114に記憶された数字列(「4・9・1・3・5」)を読み出して、自身が生成した組み合わせ画像NGに当てはめて、対応する画像の色から、「き・みどり・あお・き・あお」という文字を拾い出し、この順序で並べた文字列(第2の文字/数字列という)に変換する。更に処理部112は、第1の文字/数字列と第2の文字/数字列とが一致していた場合、認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。一方、第1の文字/数字列と第2の文字/数字列とが一致しなければ、処理部112は認証コードが不一致であると判断し、HMD100の画面ロックを続行する。このとき、新たな認証コードの入力を求めるようにしても良い。それ以外の構成は、上述した実施形態と同様である。 In parallel with this, the processing unit 112 that also serves as the conversion device reads the numeric string (“4 · 9 · 1 · 3 · 5”) stored in the authentication code storage unit 114 and generates the combined image NG generated by itself. Then, the characters “ki, green, ao, ki, ao” are picked up from the corresponding image color and converted to a character string (called a second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. On the other hand, if the first character / number string does not match the second character / number string, the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested. Other configurations are the same as those in the above-described embodiment.
 次に、認証コード記憶部114に記憶されたパスワードの更新方法について,以下に説明する。パスワードの更新を望むユーザーUSは、HMD100の画面ロックを解除した後、設定画面(不図示)から、パスワードの更新を選択して後述する操作を行うものとする。ここでは、認証コード記憶部114に記憶されたパスワード「1・2・3・4」を更新コードとして用い、新たなパスワード「9・8・5・6」へと更新するものとする。 Next, a method for updating the password stored in the authentication code storage unit 114 will be described below. It is assumed that the user US who wants to update the password releases the screen lock of the HMD 100, selects the password update from the setting screen (not shown), and performs an operation described later. Here, it is assumed that the password “1, 2, 3, 4” stored in the authentication code storage unit 114 is used as an update code and updated to a new password “9, 8, 5, 6”.
 図10(a)~(c)は、処理部112で生成される数字と画像の組み合わせ画像NGを、説明画像RGと共に示す図である。処理部112は、ユーザーUSからパスワード更新を要求された場合、表示駆動制御部104DRを介して図10(a)に示す組み合わせ画像NGをディスプレイユニット104に表示する。但し、数字と色の組み合わせはランダムである。 10 (a) to 10 (c) are diagrams showing a combination image NG of numbers and images generated by the processing unit 112 together with an explanation image RG. When a password update is requested from the user US, the processing unit 112 displays the combined image NG illustrated in FIG. 10A on the display unit 104 via the display drive control unit 104DR. However, the combination of numbers and colors is random.
 ユーザーUSは、更新コードとしての数字列(「1・2・3・4」)を記憶しているので、図10(a)に示すような組み合わせ画像NGを見たときに、数字「1」に対応するブロックの色が青色であり、数字「2」に対応するブロックの色が赤色であり、数字「3」に対応するブロックの色が黄色であり、数字「4」に対応するブロックの色が黄色であることが分かる。そこで、ユーザーUSが「あお・あか・き・き」と色のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「あお・あか・き・き」という文字列(第1の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 Since the user US stores a numeric string (“1, 2, 3, 4”) as an update code, when viewing the combined image NG as shown in FIG. The color of the block corresponding to is blue, the color of the block corresponding to the number “2” is red, the color of the block corresponding to the number “3” is yellow, and the block corresponding to the number “4” It can be seen that the color is yellow. Therefore, when the user US utters a word of the color “Ao / Aka / Ki / Ki”, the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / number string) “Ao, Aka, Ki, Ki” and transmitted to the processing unit 112 as character string information.
 これと並行して、処理部112は、認証コード記憶部114に更新コードとして記憶された数字列(「1・2・3・4」)を読み出して、自身が生成した組み合わせ画像NG(図10(a)に当てはめて、対応する画像の色から、「あお・あか・き・き」という文字を拾い出し、この順序で並べた文字列(第2の文字/数字列という)に変換する。更に処理部112は、第1の文字/数字列と第2の文字/数字列とが一致していた場合、更新コードが一致したものと判断し、パスワードの更新を許可する。より具体的には、パスワードの更新を許可した処理部112は、図10(b)に示すように数字と色の対応付けを変えた組み合わせ画像NGを生成して、表示駆動制御部104DRを介してディスプレイユニット104に表示する。 In parallel with this, the processing unit 112 reads the numeric string (“1, 2, 3, 4”) stored as the update code in the authentication code storage unit 114, and generates the combined image NG (FIG. 10). Applying to (a), the characters “Ao, Aka, Ki, Ki” are picked up from the color of the corresponding image, and converted into a character string (second character / number string) arranged in this order. Further, when the first character / numeric string and the second character / numeric string match, the processing unit 112 determines that the update codes match and permits the password update. The processing unit 112 that has permitted the update of the password generates a combined image NG in which the correspondence between the numbers and the colors is changed as shown in FIG. 10B, and displays the display unit 104 via the display drive control unit 104DR. To display.
 新たなパスワード「9・8・5・6」へと更新を望むユーザーUSは、図10(b)に示す組み合わせ画像NGを見たときに、数字「9」に対応するブロックの色が黄色であり、数字「8」に対応するブロックの色が紫色であり、数字「5」に対応するブロックの色が緑色であり、数字「6」に対応するブロックの色が赤色であることが分かる。 When the user US who wishes to update to the new password “9, 8, 5, 6” sees the combined image NG shown in FIG. 10B, the color of the block corresponding to the number “9” is yellow. In other words, the color of the block corresponding to the number “8” is purple, the color of the block corresponding to the number “5” is green, and the color of the block corresponding to the number “6” is red.
 そこで、ユーザーUSが「き・むらさき・みどり・あか」と色のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「き・むらさき・みどり・あか」という更新文字列に変換し、更新文字列情報として処理部112に送信する。 Therefore, when the user US utters a word of the color “Ki / Murasaki / Midori / Aka”, the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an updated character string “Ki / Murasaki / Midori / Aka” and transmitted to the processing unit 112 as updated character string information.
 処理部112は、図10(b)に示す組み合わせ画像NGに照合することで、更新文字列中の文字「き」は黄色を示し、それに対応する数字は「7,9」であると判断し、文字「むらさき」は紫色を示し、それに対応する数字は「4,8」であると判断し、文字「みどり」は緑色を示し、それに対応する数字は「5」であると判断し、文字「あか」は赤色を示し、それに対応する数字は「1,6」であると判断する。すると、処理部112は、ユーザーUSが更新を所望するパスワードの候補が複数あり、それは「7・4・5・1」、「9・4・5・1」、「7・8・5・1」、「9・8・5・1」、「7・4・5・6」、「9・4・5・6」、「7・8・5・6」、「9・8・5・6」の8通りあると判断する。これら8通りの数字列を、パスワード候補として処理部112は記憶する。このように、処理部112は、文字列の中の文字に対応する数字列を特定し、その特定された数字列に基づく処理を行う。 The processing unit 112 collates with the combined image NG shown in FIG. 10B to determine that the character “ki” in the updated character string is yellow and the corresponding number is “7, 9”. The character “Murasaki” indicates purple, the corresponding number is determined to be “4,8”, the character “Midori” indicates green, the corresponding number is determined to be “5”, and the character “Red” indicates red, and the corresponding numbers are determined to be “1, 6”. Then, the processing unit 112 has a plurality of password candidates that the user US desires to update, which are “7 · 4 · 5 · 1”, “9 · 4 · 5 · 1”, “7 · 8 · 5 · 1”. "," 9,5,5 "," 7,4,5,6 "," 9,4,5,6 "," 7,8,5,6 "," 9,8,5,6 " It is determined that there are eight ways. The processing unit 112 stores these eight numeric strings as password candidates. As described above, the processing unit 112 identifies a numeric string corresponding to a character in the character string, and performs processing based on the identified numeric string.
 かかる場合、処理部112は、図10(c)に示すように数字と色の対応付けを変えた組み合わせ画像NGを新たに生成して、表示駆動制御部104DRを介してディスプレイユニット104に再度表示する。ユーザーUSは、図10(c)に示す組み合わせ画像NGを見たときに、数字「9」に対応するブロックの色が青色であり、数字「8」に対応するブロックの色が緑色であり、数字「5」に対応するブロックの色が黄色であり、数字「6」に対応するブロックの色が緑色であることが分かる。 In such a case, the processing unit 112 newly generates a combination image NG in which the correspondence between numbers and colors is changed as illustrated in FIG. 10C and displays the combined image NG again on the display unit 104 via the display drive control unit 104DR. To do. When the user US views the combination image NG shown in FIG. 10C, the color of the block corresponding to the number “9” is blue, and the color of the block corresponding to the number “8” is green. It can be seen that the color of the block corresponding to the number “5” is yellow and the color of the block corresponding to the number “6” is green.
 そこで、ユーザーUSが「あお・みどり・き・みどり」と色のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「あお・みどり・き・みどり」という更新文字列に変換し、更新文字列情報として処理部112に送信する。 Therefore, when the user US utters a word of the color “Ao / Midori / Ki / Midori”, the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an update character string “Ao / Midori / Ki / Midori” and transmitted to the processing unit 112 as update character string information.
 処理部112は、図10(c)に示す組み合わせ画像NGに照合することで、更新文字列中の文字「あお」は青色を示し、それに対応する数字は「2,9」であると判断し、文字「みどり」は緑色を示し、それに対応する数字は「6,8」であると判断し、文字「き」は黄色を示し、それに対応する数字は「1,5」であると判断する。これを記憶していたパスワード候補と照合すると、「9・8・5・6」の1通りのみが残るので、処理部112は、数字列「9・8・5・6」を新たなパスワードとして、認証コード記憶部114のパスワードを更新するのである。2度目の処理でもパスワードの候補が複数残る場合、処理部112は、新たな組み合わせ画像を更に表示して、ユーザーUSの発話を求めれば良い。尚、以上の実施形態では、組み合わせ画像に用いる色の種類を区別しやすいように5つに限定しているので、パスワード入力のためユーザーUSの複数回の発話を要したが、色の種類をパスワードの数字の数に一対一で合わせれば、ユーザーの発話は1回で済むこととなる。或いは、組み合わせ画像として色の種類を用いることなく、以下の実施形態のように文字や絵柄などを用いれば、画像の数とパスワードの数字の数とを一対一で合わせることができるから、ユーザーの発話は同様に1回で済むこととなる。 The processing unit 112 collates with the combination image NG shown in FIG. 10C to determine that the character “AO” in the updated character string is blue and the corresponding number is “2, 9”. The character “Midori” indicates green, and the corresponding number is determined to be “6, 8”, the character “K” indicates yellow, and the corresponding number is determined to be “1, 5”. . When this is compared with the stored password candidate, only one of “9, 8, 5, 6” remains, so the processing unit 112 uses the numeric string “9, 8, 5, 6” as a new password. The password in the authentication code storage unit 114 is updated. If a plurality of password candidates remain in the second process, the processing unit 112 may further display a new combined image and obtain the user's US utterance. In the above embodiment, the number of colors used for the combined image is limited to five so that it can be easily distinguished. Therefore, the user US has to speak a plurality of times for password input. If the number of passwords is matched one-on-one, the user's utterance can be done only once. Alternatively, the number of images and the number of passwords can be matched on a one-to-one basis by using characters and patterns as in the following embodiment without using color types as combination images. Similarly, one utterance is sufficient.
 図11は、認証コード記憶部114に記憶された数字列を構成する数字に対応して、有色ブロックの代わりに文字を、所定の関係として対応づけて配置した対応表CTを示す図である。認証時には、図11の対応表CTがディスプレイユニット104に表示される。 FIG. 11 is a diagram showing a correspondence table CT in which characters are arranged in association with a predetermined relationship instead of colored blocks in correspondence with the numbers constituting the number string stored in the authentication code storage unit 114. At the time of authentication, the correspondence table CT in FIG. 11 is displayed on the display unit 104.
 ここでユーザーUSは、認証コードとしての数字列(「4・9・1・3・5」)を記憶しているので、図11に示すような対応表CTを見たときに、数字「4」に対応する文字が「こ」であり、数字「9」に対応する文字が「た」であり、数字「1」に対応する文字が「の」であり、数字「3」に対応する文字が「か」であり、数字「5」に対応する文字が「ま」であることが分かる。そこで、ユーザーUSが「こ・た・の・か・ま」とひらがなを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「こ・た・の・か・ま」という文字列(第1の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 Here, since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the correspondence table CT as shown in FIG. Is the character corresponding to the number “9”, the character corresponding to the number “9” is “no”, the character corresponding to the number “1” is “no”, and the character corresponding to the number “3”. Is “ka”, and the character corresponding to the number “5” is “ma”. Therefore, when the user US speaks hiragana “ko, ta, no, ka, or ma”, the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / numerical string) of “ko, ta, no, koma” and transmitted to the processing unit 112 as character string information.
 これと並行して、処理部112は、認証コード記憶部114に記憶された数字列(「4・9・1・3・5」)を読み出して、自身が生成した対応表CT(図11)に照合することで、それぞれ数字に対応する「こ・た・の・か・ま」という文字を拾い出し、この順序で並べた文字列(第2の文字/数字列という)に変換する。更に処理部112は、第1の文字/数字列と第2の文字/数字列とが一致していた場合、認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。それ以外の構成は、上述した実施形態と同様である。 In parallel with this, the processing unit 112 reads the numeric string (“4 · 9 · 1 · 3 · 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 11) generated by itself. Are collated to pick up the characters “ko, ta, no, ka, or ma” corresponding to the numbers, respectively, and convert them into a character string arranged in this order (referred to as a second character / number string). Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
 以上の変形例として、認証コードとしての数字列(「4・9・1・3・5」)を記憶しているユーザーUSが、図11に示すような対応表CTを見たときに、「こ・た・の・か・ま」とひらがなを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「こ・た・の・か・ま」という文字列(第3の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 As a modification of the above, when the user US who stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code looks at the correspondence table CT as shown in FIG. When the hiragana is spoken, the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. Is converted into a character string (referred to as a third character / number string) and transmitted to the processing unit 112 as character string information.
 更に、処理部112は、「こ・た・の・か・ま」という文字列を、自身が生成した対応表CT(図11)に照合することで、それぞれひらがなに対応する数字に変換し、この順序で並べた数字列「4・9・1・3・5」(第4の文字/数字列という)に変換する。更に処理部112は、第4の文字/数字列とパスワードとが一致していた場合、認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。それ以外の構成は、上述した実施形態と同様である。 Further, the processing unit 112 compares the character string “ko, ta, no, ka, or ma” with the correspondence table CT (FIG. 11) generated by the processing unit 112 to convert the character string into numbers corresponding to hiragana, respectively. It is converted to a numeric string “4 · 9 · 1, 3 · 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
 図12は、認証コード記憶部114に記憶された数字列を構成する数字に対応して、有色ブロック又は文字の代わりに野菜又は果物の絵柄を、所定の関係として対応づけて配置した対応表CTを示す図である。尚、対応表CTには、表示した野菜又は果物の絵柄に対応づけて、野菜又は果物の名称を登録しているが、必ずしも必要なものではない。認証時には、図12の対応表CTがディスプレイユニット104に表示される。 FIG. 12 shows a correspondence table CT in which vegetables or fruit patterns are arranged in association with a predetermined relationship in place of colored blocks or characters, corresponding to the numbers constituting the number string stored in the authentication code storage unit 114. FIG. In the correspondence table CT, the names of vegetables or fruits are registered in association with the displayed patterns of vegetables or fruits, but this is not always necessary. At the time of authentication, the correspondence table CT shown in FIG. 12 is displayed on the display unit 104.
 ここでユーザーUSは、認証コードとしての数字列(「4・9・1・3・5」)を記憶しているので、図12に示すような組み合わせ画像NGを見たときに、数字「4」に対応する絵柄が「しいたけ」であり、数字「9」に対応する絵柄が「ピーマン」であり、数字「1」に対応する絵柄が「トマト」であり、数字「3」に対応する絵柄が「サクランボ」であり、数字「5」に対応する絵柄が「ミカン」であることが分かる。そこで、ユーザーUSが「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」と絵柄のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」という文字列(第1の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 Here, since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG. The pattern corresponding to "" is shiitake, the pattern corresponding to the number "9" is "green pepper", the pattern corresponding to the number "1" is "tomato", and the pattern corresponding to the number "3" Is “cherry” and the pattern corresponding to the number “5” is “mandarin orange”. Therefore, when the user US utters the word “Shitake / Piman / Tomato / Sakurabo / Mikan”, the microphone 105 collects the sound and inputs the speech signal to the speech recognition unit 113 via the speech processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) “Shitake / Piman / Tomato / Cherry Bamboo / Mikan” and transmitted to the processing unit 112 as character string information.
 これと並行して、処理部112は、認証コード記憶部114に記憶された数字列(「4・9・1・3・5」)を読み出して、自身が生成した対応表CT(図12)に照合することで、それぞれ数字に対応する「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」という文字を拾い出し、この順序で並べた文字列(第2の文字/数字列という)に変換する。更に処理部112は、第1の文字/数字列と第2の文字/数字列とが一致していた場合、認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。それ以外の構成は、上述した実施形態と同様である。 In parallel with this, the processing unit 112 reads the numeric string (“4 · 9 · 1 · 3 · 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 12) generated by itself. By collating with each other, the characters “Shitake / Piman / Tomato / Sakurambo / Mikan” corresponding to the numbers are picked up and converted into a character string (second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
 以上の変形例として、認証コードとしての数字列(「4・9・1・3・5」)を記憶しているユーザーUSが、図12に示すような対応表CTを見たときに、「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」と絵柄のワードを発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」という文字列(第3の文字/数字列という)に変換し、文字列情報として処理部112に送信する。 As a modification example described above, when the user US storing a numeric string (“4, 9, 1, 3, 5”) as an authentication code looks at the correspondence table CT as shown in FIG. When a word with a pattern such as “Shitake, Piman, Tomato, Sakura, Mikan” is uttered, the microphone 105 collects the sound, and a speech signal is input to the speech recognition unit 113 via the speech processing unit 106. It is converted into a character string (referred to as a third character / number string) “Tomato, cherry, orange” and transmitted to the processing unit 112 as character string information.
 更に、処理部112は、「しいたけ・ぴーまん・とまと・さくらんぼ・みかん」という文字列を、自身が生成した対応表CT(図11)に照合することで、それぞれ絵柄に対応する数字に変換し、この順序で並べた数字列「4・9・1・3・5」(第4の文字/数字列という)に変換する。更に処理部112は、第4の文字/数字列とパスワードとが一致していた場合、認証コードが一致したものと判断し、HMD100の画面ロックを解除するのである。それ以外の構成は、上述した実施形態と同様である。 Further, the processing unit 112 converts the character string “Shitake / Pi-Man / Tomato / Cherry Bamboo / Tangerine” into a correspondence table CT (FIG. 11) generated by itself, thereby converting the character string into a number corresponding to the design. It is converted to a numeric string “4 · 9 · 1, 3 · 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
 図13は、上述した実施形態にかかる処理部112の制御動作(ただし、ステップS102を除く)を示すフローチャートである。ステップS101で、処理部112は、画像として数字列NA(図5)、文字列CA(図7)、画像の並びGA(図8)、組み合わせ画像NG(図9(b))、対応表CT(図11,12)を表示する。 FIG. 13 is a flowchart showing a control operation (except for step S102) of the processing unit 112 according to the above-described embodiment. In step S101, the processing unit 112, as an image, includes a numeric string NA (FIG. 5), a character string CA (FIG. 7), an image arrangement GA (FIG. 8), a combined image NG (FIG. 9B), and a correspondence table CT. (FIGS. 11 and 12) are displayed.
 表示画像を見たユーザーUSが、これに対応して発話すると(ステップS102)、音声認識部113がその音声が表す文字/数字列を認識することで音声認識を行い(ステップS103)、処理部112は,ステップS104に示すように音声処理結果を表示する(図6参照) When the user US who has seen the display image speaks in response to this (step S102), the voice recognition unit 113 recognizes the character / numeric string represented by the voice and performs voice recognition (step S103). 112 displays the audio processing result as shown in step S104 (see FIG. 6).
 ここで、音声処理結果が不適切であれば(ステップS105でNO)、フローはステップS102へと戻り、同様の処理を繰り返す。一方、音声処理結果が適切であれば(ステップS105でYES)、処理部112は、ステップS106で認証コードを読み出して、ステップS107で音声認識結果と認証コードとを照合して、両者が一致しているか否かを処理する。 Here, if the voice processing result is inappropriate (NO in step S105), the flow returns to step S102, and the same processing is repeated. On the other hand, if the voice processing result is appropriate (YES in step S105), the processing unit 112 reads the authentication code in step S106, collates the voice recognition result and the authentication code in step S107, and the two match. Process whether or not.
 照合の結果、音声認識結果と認証コードとが不一致であると判断した場合、ステップS109にて処理部112は、例えば「認証に失敗しました」などのメッセージを、ディスプレイユニット104に表示して、画面ロックを続行する。 As a result of the collation, when it is determined that the voice recognition result and the authentication code do not match, the processing unit 112 displays a message such as “authentication failed” on the display unit 104 in step S109, Continue screen lock.
 これに対し、音声認識結果と認証コードとが一致していると判断した場合、ステップS108にて処理部112は、認証に成功したとして画面ロックを解除する。 On the other hand, if it is determined that the voice recognition result matches the authentication code, in step S108, the processing unit 112 releases the screen lock because the authentication is successful.
 例えば、一般的な生体認証ではユーザー個人を特定することはできるが、解除コードとしての生体情報は、複数人で共有することが困難というというデメリットがある。すなわち、複数人分の生体情報を登録すれば共有できるが、登録に手間がかかる,特に、工場で電子機器を使用する場合など、ひとつの電子機器を複数人で共有することが予想されるが、そのような場合には共通するパスワードを使用して、共用の利便性を高めたいという要請がある。本実施形態によれば、従来のパスワードの認証とほぼ同様なユーザー負荷で使用可能というメリットがある。又、音声認識はハンズフリーで入力できるユーザーインターフェースであり、特にHMDと親和性が高いという利点もある。更に、ユーザーの頭部に装着するHMDであれば、ディスプレイユニットに表示される画像が他人からは見えくいため、秘匿性を高く保てるという利点がある。ユーザーが発話する文字列を、万が一他人が覚えたとしても、表示毎に異なる画像(或いは文字/数字列と画像の組み合わせ)が表示されるようにすれば、覚えた文字列を別のタイミングで他人が発話したとしても、電子機器のロックが解除されず、強固なセキュリティを確保できる。更に、従来の電子機器にパスワードを入力する際に行われていた、入力した文字等を伏せ字「******」にして表示する処理などの必要がない。ディスプレイユニットに表示される画像が他人から見えなければ、音声認識した結果を表示する際に逆変換したコードを表示してもよい。 For example, although general biometric authentication can identify an individual user, there is a demerit that biometric information as a release code is difficult to share with multiple people. In other words, it can be shared by registering biometric information for a plurality of people, but it takes time to register, and it is expected that a single electronic device will be shared by a plurality of people, especially when an electronic device is used in a factory. In such a case, there is a request to increase the convenience of sharing by using a common password. According to the present embodiment, there is an advantage that it can be used with almost the same user load as conventional password authentication. Speech recognition is a user interface that can be input hands-free, and has an advantage of high affinity with HMD. Furthermore, if the HMD is mounted on the user's head, the image displayed on the display unit is difficult for others to see, so there is an advantage that confidentiality can be kept high. Even if someone remembers the character string that the user utters, if a different image (or a combination of a character / number string and an image) is displayed for each display, the learned character string will be displayed at a different timing. Even if another person speaks, the electronic device is not unlocked and strong security can be secured. In addition, there is no need for the process of displaying the input characters etc. as the hidden characters “******”, which is performed when inputting the password to the conventional electronic device. If the image displayed on the display unit is not visible to others, a code reversely converted when displaying the result of voice recognition may be displayed.
(第3の実施形態)
 本実施形態のHMD100を、情報の入力手段として用いることもできる。例えば、「030-1234-5678」という電話番号に、ユーザーUSが架電しようとする場合、ユーザーUSの要求に応じて、処理部112は、図11に示すような対応表CTを生成してディスプレイユニット104に表示する。ここで、ユーザーUSは、表示された対応表CTに従って、電話番号に対応する文字を「け・か・け・の・る・か・こ・ま・み・に・い」と発話すると、これをマイク105が集音して、音声処理部106を介して音声認識部113に音声信号が入力され、「け・か・け・の・る・か・こ・ま・み・に・い」という文字列(第1の文字/数字列)に変換し、文字列情報として処理部112に送信する。
(Third embodiment)
The HMD 100 of the present embodiment can also be used as information input means. For example, when the user US tries to make a call to a telephone number “030-1234-5678”, the processing unit 112 generates a correspondence table CT as shown in FIG. 11 in response to a request from the user US. Displayed on the display unit 104. Here, according to the displayed correspondence table CT, the user US utters the characters corresponding to the telephone number as “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i”. Is picked up by the microphone 105, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106, and "Ke-ka-Ke-no-ru-ka-ko-ma-mi-ni-i" Is converted to a character string (first character / number string) and transmitted to the processing unit 112 as character string information.
 処理部112は、自身が生成した対応表CT(図11)を用いて、入力された文字列「け・か・け・の・る・か・こ・ま・み・に・い」を、「0・3・0・1・2・3・4・5・6・7・8」という数字列(第2の文字/数字列)に変換する。更に処理部112は、電話機能を有している場合、得られた数字列を電話番号として入力することで架電することができる。かかる場合、処理部112は入力装置を兼ねることとなる。これにより、ユーザーUSは相手先の電話番号を周囲の者に知られることなく、ハンズフリーで架電することが可能になる。変換された電話番号をディスプレイユニット104に表示して、ユーザーUSの確認後に架電するようにしても良い。HMD100は、電話番号に限らずマイナンバーやクレジットカードナンバー等の入力に用いても良い。 The processing unit 112 uses the correspondence table CT (FIG. 11) generated by itself to process the input character string “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i” It is converted into a number string (second character / number string) of “0, 3, 0, 1, 2, 3, 4, 5, 6, 7, 8”. Further, when the processing unit 112 has a telephone function, the processing unit 112 can make a call by inputting the obtained numeric string as a telephone number. In such a case, the processing unit 112 also serves as an input device. As a result, the user US can make a hands-free call without knowing the other party's telephone number. The converted telephone number may be displayed on the display unit 104 to make a call after confirming the user US. The HMD 100 may be used for inputting not only a telephone number but also a My Number, a credit card number, and the like.
 本発明は、明細書に記載の実施形態に限定されるものではなく、他の実施形態・変形例を含むことは、本明細書に記載された実施形態や技術思想から本分野の当業者にとって明らかである。明細書の記載及び実施形態は、あくまでも例証を目的としており、本発明の範囲は後述するクレームによって示されている。例えば、以上の実施形態では、HMDを例にとり本発明を説明してきたが、本発明はHMDに限らず、携帯端末などの電子機器全般に適用可能である。又、上述した認証により許可される機能は、一部であっても全てであってもよい。 The present invention is not limited to the embodiments described in the specification, and other embodiments and modifications are included for those skilled in the art from the embodiments and technical ideas described in the present specification. it is obvious. The description and the embodiments are for illustrative purposes only, and the scope of the present invention is indicated by the following claims. For example, in the above embodiment, the present invention has been described by taking the HMD as an example. However, the present invention is not limited to the HMD and can be applied to all electronic devices such as portable terminals. Further, some or all of the functions permitted by the authentication described above may be used.
 また、上述の実施形態では、パスワードの入力により所定機能を許可する例として画面ロックの解除を挙げたが、パスワードの入力により特定のアプリケーションの起動許可を付与することもできる。より具体的には、ディスプレイユニット104にアプリケーション起動時のログイン画面などが表示された状態で、上述したようにしてパスワードを適切に入力することで、かかるアプリケーションを起動することができる。尚、特定のアプリケーションを起動した後、その中で認証を行うこともできる。かかる場合、スイッチをオンするような操作を用いず、ユーザーの発話を用いたハンズフリーの操作の中で認証画面に遷移するようにすることが望ましい。 In the above-described embodiment, the screen lock is released as an example of permitting a predetermined function by inputting a password. However, it is also possible to grant start permission of a specific application by inputting a password. More specifically, the application can be started by appropriately inputting a password as described above in a state where a login screen at the time of starting the application is displayed on the display unit 104. In addition, after starting a specific application, it can also authenticate in it. In such a case, it is desirable to change to the authentication screen during a hands-free operation using the user's utterance without using an operation to turn on the switch.
101      フレーム
102      眼鏡レンズ
103      主本体部
104      ディスプレイユニット
104A     画像形成部
104B     画像表示部
104DR    表示駆動制御部
104a     光源
104b     一方向拡散板
104c     集光レンズ
104d     表示素子
104f     接眼プリズム
104g     偏向プリズム
104h     ホログラム光学素子
105      マイク
106      音声処理部
110      制御部
112      処理部
113      音声認識部
114      認証コード記憶部
CA       文字列
CT       対応表
NA       数字列
NG       画像
PT       パターン
US       ユーザー
DESCRIPTION OF SYMBOLS 101 Frame 102 Eyeglass lens 103 Main body part 104 Display unit 104A Image formation part 104B Image display part 104DR Display drive control part 104a Light source 104b Unidirectional diffuser 104c Condensing lens 104d Display element 104f Eyepiece prism 104g Deflection prism 104h Hologram optical element 105 Microphone 106 Voice processing unit 110 Control unit 112 Processing unit 113 Voice recognition unit 114 Authentication code storage unit CA Character string CT Correspondence table NA Number string NG Image PT Pattern US User

Claims (16)

  1.  パターンを記憶する記憶装置と、
     複数の画像を表示する表示装置と、
     前記画像に応じてユーザーが発話した音声を取得して、対応する文字/数字列に変換する音声認識装置と、
     前記音声認識装置が変換した文字/数字列の順序で、前記表示装置に表示された前記複数の画像を繋いでいったとき、その軌跡が前記記憶装置に記憶された前記パターンと一致したときは、所定の機能を許可する処理装置とを有する電子機器。
    A storage device for storing patterns;
    A display device for displaying a plurality of images;
    A voice recognition device that acquires voice spoken by a user according to the image and converts the voice into a corresponding character / number string;
    When the plurality of images displayed on the display device are connected in the order of the character / number string converted by the voice recognition device, and the locus matches the pattern stored in the storage device And an electronic device having a processing device that permits a predetermined function.
  2.  前記表示装置は、表示の度に、異なる画像を表示する請求項1に記載の電子機器。 The electronic device according to claim 1, wherein the display device displays a different image each time it is displayed.
  3.  前記所定の機能とは前記パターンの更新であり、前記処理装置が前記パターンの更新を許可したときは、新たなパターンが入力されたことに応じて、前記記憶装置に記憶された前記パターンを更新する請求項1又は2に記載の電子機器。 The predetermined function is the update of the pattern, and when the processing device permits the update of the pattern, the pattern stored in the storage device is updated in response to the input of a new pattern. The electronic device according to claim 1 or 2.
  4.  文字/数字列からなるパスワードを記憶する記憶装置と、
     少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換する音声認識装置と、
     前記パスワードを、前記所定の関係に従って第2の文字/数字列に変換する変換装置と、
     前記音声認識装置が変換した前記第1の文字/数字列と、前記変換装置が変換した前記第2の文字/数字列が一致したときは、所定の機能を許可する処理装置とを有する電子機器。
    A storage device for storing a password consisting of a character / number string;
    A display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship;
    A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string;
    A conversion device for converting the password into a second character / number string in accordance with the predetermined relationship;
    An electronic apparatus having a processing device that permits a predetermined function when the first character / numeric string converted by the voice recognition device matches the second character / numeric string converted by the conversion device .
  5.  文字/数字列からなるパスワードを記憶する記憶装置と、
     少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第3の文字/数字列に変換する音声認識装置と、
     前記第3の文字/数字列を、前記所定の関係に従って第4の文字/数字列に変換し、前記第4の文字/数字列と前記パスワードが一致したときは、所定の機能を許可する処理装置とを有する電子機器。
    A storage device for storing a password consisting of a character / number string;
    A display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship;
    A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a third letter / number string;
    Processing for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password An electronic device having the apparatus.
  6.  前記表示装置は、表示の度に、画像と文字/数字とを対応付けする前記所定の関係を変更する請求項4又は5に記載の電子機器。 The electronic device according to claim 4 or 5, wherein the display device changes the predetermined relationship for associating an image with a character / number for each display.
  7.  前記所定の機能とは前記パスワードの更新であり、前記処理装置が前記パスワードの更新を許可したときは、新たなパスワードが入力されたことに応じて、前記記憶装置に記憶された前記パスワードを更新する請求項4~6のいずれかに記載の電子機器。 The predetermined function is update of the password, and when the processing device permits the update of the password, the password stored in the storage device is updated in response to the input of a new password. The electronic device according to any one of claims 4 to 6.
  8.  規定の文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示する表示装置と、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換する音声認識装置と、
     前記音声認識装置が変換した第1の文字/数字列を、前記所定の関係に従って第2の文字/数字列に変換する変換装置と、
     前記規定の文字/数字列として、前記第2の文字/数字列を入力する入力装置と、を有する電子機器。
    A display device that displays letters / numbers constituting a prescribed letter / number string in association with a plurality of images according to a predetermined relationship;
    A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string;
    A conversion device that converts the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
    An electronic apparatus comprising: an input device that inputs the second character / number string as the prescribed character / number string.
  9.  前記表示装置は、表示の度に、画像と文字/数字とを対応付けする前記所定の関係を変更する請求項8に記載の電子機器。 9. The electronic device according to claim 8, wherein the display device changes the predetermined relationship for associating an image with a character / number for each display.
  10.  複数の画像を表示する表示装置と、
     マイクと、
     前記マイクを通して取得した音声を解析し、当該音声が表す文字/数字列を認識する音声認識装置と、
     前記複数の画像の中で、前記文字/数字列に対応する画像を特定し、前記特定された画像に基づく処理を行う処理部と、を有する電子機器。
    A display device for displaying a plurality of images;
    With a microphone,
    A voice recognition device that analyzes voice acquired through the microphone and recognizes a character / number string represented by the voice;
    An electronic device comprising: a processing unit that specifies an image corresponding to the character / number string among the plurality of images and performs processing based on the specified image.
  11.  請求項1~10のいずれかに記載の電子機器を有するヘッドマウントディスプレイ。 A head mounted display having the electronic device according to any one of claims 1 to 10.
  12.  パターンを記憶し、
     複数の画像を表示し、
     前記画像に応じてユーザーが発話した音声を取得して、対応する文字/数字列に変換し、
     変換された文字/数字列の順序で、表示された前記複数の画像を繋いでいったとき、その軌跡が前記記憶装置に記憶された前記パターンと一致したときは、所定の認証を行う電子機器の処理方法。
    Remember the pattern,
    Display multiple images,
    Acquire the voice spoken by the user according to the image, convert it to the corresponding character / number string,
    An electronic device that performs predetermined authentication when the plurality of displayed images are connected in the order of the converted character / numerical string, and the trajectory matches the pattern stored in the storage device Processing method.
  13.  文字/数字列からなるパスワードを記憶し、
     少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換し、
     前記パスワードを、前記所定の関係に従って第2の文字/数字列に変換し、
     前記第1の文字/数字列と前記第2の文字/数字列が一致したときは、所定の認証を行う電子機器の処理方法。
    Memorize passwords consisting of letters and numbers,
    Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
    Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string;
    Converting the password into a second character / number string according to the predetermined relationship;
    A processing method of an electronic device that performs predetermined authentication when the first character / numeric string matches the second character / numeric string.
  14.  文字/数字列からなるパスワードを記憶し、
     少なくとも前記パスワードの文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第3の文字/数字列に変換し、
     前記第3の文字/数字列を、前記所定の関係に従って第4の文字/数字列に変換し、前記第4の文字/数字列と前記パスワードが一致したときは、所定の認証を行う電子機器の処理方法。
    Memorize passwords consisting of letters and numbers,
    Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
    Obtaining voice spoken by the user according to the image associated with the letters / numbers, and converting the voice into a third letter / number string;
    An electronic device that converts the third character / numeric string into a fourth character / numeric string according to the predetermined relationship, and performs predetermined authentication when the fourth character / numeric string matches the password Processing method.
  15.  規定の文字/数字列を構成する文字/数字を、所定の関係に従って複数の画像に対応づけて表示し、
     前記文字/数字に対応づけられた前記画像に応じてユーザーが発話した音声を取得して、第1の文字/数字列に変換し、
     前記音声認識装置が変換した第1の文字/数字列を、前記所定の関係に従って第2の文字/数字列に変換し、
     前記規定の文字/数字列として、前記第2の文字/数字列を入力する電子機器の処理方法。
    The letters / numbers that make up the prescribed letter / number string are displayed in association with a plurality of images according to a predetermined relationship,
    Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string;
    Converting the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
    A processing method of an electronic device for inputting the second character / number string as the prescribed character / number string.
  16.  請求項12~15のいずれかに記載の電子機器の処理方法をコンピュータに実行させるためのプログラム。 A program for causing a computer to execute the processing method for an electronic device according to any one of claims 12 to 15.
PCT/JP2017/021067 2016-06-14 2017-06-07 Electronic apparatus, head-mounted display, processing method of electronic apparatus, and program therefor WO2017217288A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2016118046 2016-06-14
JP2016-118046 2016-06-14

Publications (1)

Publication Number Publication Date
WO2017217288A1 true WO2017217288A1 (en) 2017-12-21

Family

ID=60663453

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/021067 WO2017217288A1 (en) 2016-06-14 2017-06-07 Electronic apparatus, head-mounted display, processing method of electronic apparatus, and program therefor

Country Status (1)

Country Link
WO (1) WO2017217288A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008257701A (en) * 2007-03-12 2008-10-23 Yahoo Japan Corp Authentication system
JP2010009544A (en) * 2008-06-30 2010-01-14 Toppan Forms Co Ltd Personal identification system and personal identification method
JP2014092941A (en) * 2012-11-02 2014-05-19 Sony Corp Information processor and information processing method and computer program
CN104468522A (en) * 2014-11-07 2015-03-25 百度在线网络技术(北京)有限公司 Voiceprint authentication method and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008257701A (en) * 2007-03-12 2008-10-23 Yahoo Japan Corp Authentication system
JP2010009544A (en) * 2008-06-30 2010-01-14 Toppan Forms Co Ltd Personal identification system and personal identification method
JP2014092941A (en) * 2012-11-02 2014-05-19 Sony Corp Information processor and information processing method and computer program
CN104468522A (en) * 2014-11-07 2015-03-25 百度在线网络技术(北京)有限公司 Voiceprint authentication method and device

Similar Documents

Publication Publication Date Title
US10360412B2 (en) Contextual contemporaneous gesture and keyboard entry authentication
US9503800B2 (en) Glass-type terminal and method of controlling the same
US8090201B2 (en) Image-based code
US9275213B2 (en) Method and system for securing the entry of data to a device
EP2851831B1 (en) Mobile Information Gateway for Home Healthcare
EP2851832B1 (en) Mobile information gateway for use by medical personnel
US8873147B1 (en) Chord authentication via a multi-touch interface
KR102393892B1 (en) Terminal device and method for performing user authentication using biometric information
US9336779B1 (en) Dynamic image-based voice entry of unlock sequence
JP2014092940A (en) Image display device and image display method and computer program
US9552471B1 (en) Personal familiarity authentication
JP2008241822A (en) Image display device
CN105900103A (en) Touch terminal and password generation method thereof
WO2017217288A1 (en) Electronic apparatus, head-mounted display, processing method of electronic apparatus, and program therefor
Saulynas et al. Towards the use of brain–computer interface and gestural technologies as a potential alternative to PIN authentication
US20160224808A1 (en) Information input method having confidentiality
WO2016200084A1 (en) Iris recognition usb device using otp function and method of controlling same
RU2751095C2 (en) Providing access to structured stored data
Tkauc et al. Cloud-Based Face and Speech Recognition for Access Control Applications
KR20050090102A (en) Number fade password input system
TW200841204A (en) Hand cryptographic device
TW201617951A (en) Device and method for password input
TWI644232B (en) Method and apparatus for password entering
KR20020076487A (en) A method for authentication of a person using motion picture information
WO2015093221A1 (en) Electronic device and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17813186

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17813186

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP