WO2017217288A1 - Appareil électronique, visiocasque, procédé de traitement d'appareil électronique et programme associé - Google Patents

Appareil électronique, visiocasque, procédé de traitement d'appareil électronique et programme associé Download PDF

Info

Publication number
WO2017217288A1
WO2017217288A1 PCT/JP2017/021067 JP2017021067W WO2017217288A1 WO 2017217288 A1 WO2017217288 A1 WO 2017217288A1 JP 2017021067 W JP2017021067 W JP 2017021067W WO 2017217288 A1 WO2017217288 A1 WO 2017217288A1
Authority
WO
WIPO (PCT)
Prior art keywords
character
string
voice
number string
password
Prior art date
Application number
PCT/JP2017/021067
Other languages
English (en)
Japanese (ja)
Inventor
軌行 石井
Original Assignee
コニカミノルタ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by コニカミノルタ株式会社 filed Critical コニカミノルタ株式会社
Publication of WO2017217288A1 publication Critical patent/WO2017217288A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/36User authentication by graphic or iconic representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/22Interactive procedures; Man-machine interfaces
    • G10L17/24Interactive procedures; Man-machine interfaces the user being prompted to utter a password or a predefined phrase
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/64Constructional details of receivers, e.g. cabinets or dust covers

Definitions

  • the present invention relates to an electronic device having a voice recognition function, a head mounted display, a processing method of the electronic device, and a program thereof.
  • a typical mobile terminal has a touch panel screen that serves both as an image display and a user interface. By touching this screen, the user can make necessary inputs to display a desired image or input information. Can be performed.
  • biometric authentication is performed in which a biometric pattern such as a user's fingerprint, voiceprint, vein, retina, etc. is read for authentication.
  • biometric authentication requires a sensor for reading a biometric pattern and dedicated software for performing pattern matching processing, resulting in a problem that the system becomes complicated and costs increase.
  • the biometric patterns of the plurality of people must be stored, which is inconvenient.
  • some portable terminals have a voice recognition function to realize hands-free. Therefore, there is an idea that the mobile terminal can be unlocked without performing a button operation by using the voice recognition function.
  • voice recognition the words spoken by the user must first be collected with a microphone in order to convert them into character strings, etc. There is a risk that the password will be known.
  • the same problem also occurs when it is desired to input information such as a telephone number to be concealed into a portable terminal or the like using a voice recognition function.
  • Patent Document 1 a pair of keywords randomly selected from a plurality of pairs of keywords and passwords registered in advance is displayed on the display unit.
  • Patent Document 1 mentions the “correct answer rate”, which is a proof that the user is assumed to make a mistake, and there is a concern that there is difficulty in terms of usability.
  • the present invention has been made in view of the above circumstances, and is capable of suppressing leakage of information to a third party who has heard a user's utterance while using a voice recognition function, a head mounted display, and an electronic device. It is an object of the present invention to provide a device processing method and a program thereof.
  • an electronic device reflecting one aspect of the present invention.
  • a storage device for storing patterns;
  • a display device for displaying a plurality of images;
  • a voice recognition device that acquires voice spoken by a user according to the image and converts the voice into a corresponding character / number string;
  • the plurality of images displayed on the display device are connected in the order of the character / number string converted by the voice recognition device, and the locus matches the pattern stored in the storage device
  • a processing device that permits a predetermined function.
  • a storage device for storing a password consisting of a character / number string
  • a display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship
  • a voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string
  • a conversion device for converting the password into a second character / number string in accordance with the predetermined relationship
  • a processing device that permits a predetermined function when the first character / numeric string converted by the voice recognition device matches the second character / numeric string converted by the conversion device; is there.
  • a storage device for storing a password consisting of a character / number string
  • a display device that displays at least letters / numbers constituting the letter / number string of the password in association with a plurality of images according to a predetermined relationship
  • a voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a third letter / number string
  • Processing for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password
  • a device for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric string matches the password
  • a device for converting the third character / numeric string to a fourth character / numeric string in accordance with the predetermined relationship and permitting a predetermined function when the fourth character / numeric
  • still another electronic device reflecting one aspect of the present invention is: A display device that displays letters / numbers constituting a prescribed letter / number string in association with a plurality of images according to a predetermined relationship; A voice recognition device that acquires a voice spoken by a user according to the image associated with the letters / numbers and converts the voice into a first letter / number string; A conversion device that converts the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship; And an input device for inputting the second character / number string as the prescribed character / number string.
  • still another electronic device reflecting one aspect of the present invention is: A display device for displaying a plurality of images; With a microphone, A voice recognition device that analyzes voice acquired through the microphone and recognizes a character / number string represented by the voice; A processing unit that specifies an image corresponding to the character / number string among the plurality of images and performs processing based on the specified image.
  • a processing method of an electronic device reflecting one aspect of the present invention is as follows.
  • the pattern Display multiple images, Acquire the voice spoken by the user according to the image, convert it to the corresponding character / number string,
  • the plurality of displayed images are connected in the order of the converted character / numerical string, when the locus coincides with the pattern stored in the storage device, predetermined authentication is performed. is there.
  • another electronic device processing method includes: Memorize passwords consisting of letters and numbers, Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship; Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string; Converting the password into a second character / number string according to the predetermined relationship; When the first character / numeric string matches the second character / numeric string, predetermined authentication is performed.
  • still another electronic device processing method reflecting one aspect of the present invention is as follows.
  • Memorize passwords consisting of letters and numbers, Displaying at least letters / numbers constituting the password letter / number string in association with a plurality of images according to a predetermined relationship;
  • the third character / number string is converted into a fourth character / number string according to the predetermined relationship, and when the fourth character / number string matches the password, predetermined authentication is performed. is there.
  • still another electronic device processing method reflecting one aspect of the present invention is as follows.
  • the letters / numbers that make up the prescribed letter / number string are displayed in association with a plurality of images according to a predetermined relationship, Obtaining a voice spoken by the user according to the image associated with the letter / number, and converting the voice into a first letter / number string; Converting the first character / number string converted by the voice recognition device into a second character / number string in accordance with the predetermined relationship;
  • the second character / number string is input as the prescribed character / number string.
  • an electronic device a head-mounted display, a method for processing an electronic device, and a program thereof that can prevent information from leaking to a third party who has heard a user's speech while using a voice recognition function. Can do.
  • FIG. 1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state in which a user is wearing it.
  • 3 is a schematic cross-sectional view showing a configuration of a display unit 104.
  • FIG. It is a block diagram of HMD100 concerning this embodiment. It is a figure which shows the pattern PT memorize
  • FIG. It is a figure which shows the message displayed in order to confirm whether speech recognition was performed appropriately. It is a figure which shows the example of the character string CA produced
  • FIG. 1 is a figure which shows the arrangement
  • A is a figure which shows the numerical sequence (here "4 * 9 * 1 * 3 * 5") memorize
  • FIG. (B) is a figure which shows the combination image NG of the number and image produced
  • FIG. 1 is a front view showing a head mounted display (hereinafter referred to as HMD) 100, which is an electronic device, according to the present embodiment in a state where a user wears it.
  • HMD head mounted display
  • the right side and the left side of the HMD 100 refer to the right side and the left side for the user wearing the HMD 100.
  • the frame 101 to be mounted on the head of the user US has two spectacle lenses 102 arranged in front of the user US.
  • a cylindrical main body 103 is fixed on the upper part of the spectacle lens 102 on the right side (which may be on the left side according to the user's dominant eye).
  • the main body 103 is provided with a display unit 104.
  • a display drive control unit 104DR (see FIG. 3 described later) that controls display control of the display unit 104 is disposed in the main body unit 103. If necessary, a display unit may be arranged in front of both eyes.
  • FIG. 2 is a schematic cross-sectional view showing the configuration of the display unit 104.
  • the display unit 104 as a display device includes an image forming unit 104A and an image display unit 104B.
  • the image forming unit 104A is incorporated in the main body unit 103, and includes a light source 104a, a unidirectional diffuser 104b, a condenser lens 104c, and a display element 104d.
  • the image display unit 104B which is a so-called see-through type display member, is disposed on the entire plate so as to extend downward from the main body unit 103 and extend in parallel to one eyeglass lens 102 (see FIG. 1).
  • the eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h is disposed on the entire plate so as to extend downward from the main body unit 103 and extend in parallel to one eyeglass lens 102 (see FIG. 1).
  • the light emitted from the light source 104a is diffused by the unidirectional diffusion plate 104b, condensed by the condenser lens 104c, and enters the display element 104d.
  • the light incident on the display element 104d is modulated for each pixel based on the image data input from the display drive control unit 104DR, and is emitted as image light. As a result, a color image is displayed on the display element 104d.
  • Image light from the display element 104d enters the eyepiece prism 104f from its base end face PL1, is totally reflected a plurality of times by the inner side face PL2 and the outer side face PL3, and enters the hologram optical element 104h.
  • the light incident on the hologram optical element 104h is reflected there, passes through the inner side surface PL2, and reaches the pupil B.
  • the user can observe an enlarged virtual image of the image displayed on the display element 104d, and can visually recognize it as a screen formed on the image display unit 104B.
  • the hologram optical element 104h constitutes a screen, or it can be considered that a screen is formed on the inner surface PL2.
  • “screen” may refer to an image to be displayed.
  • the eyepiece prism 104f, the deflecting prism 104g, and the hologram optical element 104h transmit almost all of the external light, the user can observe an external field image (real image) through them. Therefore, the virtual image of the image displayed on the display element 104d is observed so as to overlap with a part of the external image. In this manner, the user of the HMD 100 can simultaneously observe the image provided from the display element 104d and the external image via the hologram optical element 104h. Note that when the display unit 104 is in the non-display state, the image display unit 104B is transparent, and only the external image can be observed.
  • a display unit is configured by combining a light source, a liquid crystal display element, and an optical system.
  • a self-luminous display element for example, an organic EL display
  • Element for example, an organic EL display
  • a transmissive organic EL display panel having transparency in a non-light emitting state may be used.
  • FIG. 3 is a block diagram of the HMD 100 according to the present embodiment, which is shown together with the user US.
  • the HMD 100 collects the above-described display unit (display unit) 104, the microphone 105 that collects the voice spoken by the user US and converts it into a signal, and the voice that processes the signal output from the microphone 105 and outputs it as a voice signal.
  • the processing unit 106 and the control unit 110 that receives the audio signal output from the audio processing unit 106 are included.
  • the control unit 110 receives the voice signal output from the voice processing unit 106, recognizes a character / number string represented by the voice, and converts the character / number string to the corresponding character / number, and the voice recognition unit 113.
  • a processing unit (processing device) 112 that processes characters / numbers output from the display unit, a display drive control unit 104DR that receives signals from the processing unit 112 and drives and controls the display unit 104, and an authentication code (here, a pattern) Or an authentication code storage unit (storage device) 114 that stores a password.
  • the microphone 105, the voice processing unit 106, and the voice recognition unit 113 constitute a voice recognition device.
  • letters / numbers when referring to letters / numbers, it means at least one of letters and numbers, and when referring to letters / number strings, a plurality of letters or numbers are arranged (including only letters or numbers only).
  • An image includes letters / numbers.
  • a word includes a single character.
  • FIG. 4 is a diagram showing a pattern PT stored as authentication code information in the authentication code storage unit 114.
  • the pattern PT is formed by connecting a horizontal line from left to right, a vertical line from top to bottom, and a horizontal line from left to right in this order.
  • the pattern PT is assumed to be stored in advance by the user US.
  • the “pattern” refers to a traced geometric shape, and corresponds to, for example, a one-stroke stroke.
  • FIG. 5 is a diagram schematically showing a numeric string NA generated by the processing unit 112.
  • the arrow AR is drawn so as to be easily understood, but it is not actually displayed.
  • the numerical string NA is, for example, a random number generated by the processing unit 112 assigned to 3 rows and 3 columns.
  • the processing unit 112 transmits information on the numeric string NA to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the numeric string NA into an image signal and transmits it to the display unit 104, so that the display unit 104 can display the numeric string NA shown in FIG.
  • the user US does not need to remember the arrangement order of individual elements to be displayed such as character strings, numeric strings, and images in advance. It is only necessary to utter the elements displayed along the pattern PT in order, which can be said to be highly convenient. In order to improve confidentiality, it is preferable to display individual elements at random. Elements to be displayed may be numbers, letters (alphabet, hiragana, katakana, kanji ...), colors, and patterns. These are collectively called images. Only one type (for example, only hiragana) may be displayed, or a plurality of types (such as katakana and numbers) may be mixed.
  • the reading method is registered.
  • the pattern and its reading are registered in advance in the processing unit 112 and the user US also remembers the pattern and its registered reading.
  • a plurality of readings may be registered for one pattern. However, it is desirable to avoid using multiple pictures with the same reading (bridge, chopsticks, etc.).
  • the control unit 110 performs a display (not illustrated) for requesting an authentication code to be input to the display unit 104 via the display drive control unit 104DR.
  • the numeric string NA shown in FIG. 5 is displayed.
  • the user US Since the user US stores the pattern PT as the authentication code, when the stored pattern PT is applied to the numerical string NA displayed as shown in FIG. 5, the numbers “5” are entered in the order indicated by the arrow AR. If you read “3, 6, 2, 9”, you can see that the pattern PT can be reproduced. Therefore, when the user US utters a number in the form of “Go, San, Roku, Ni, Kyu”, the microphone 105 collects the sound, and the speech recognition unit 113 via the speech processing unit 106 collects this. Is converted into a numeric string “5, 3, 6, 2, 9” and transmitted to the processing unit 112 as numeric string information.
  • the processing unit 112 causes the display unit 104 to perform a display as shown in FIG. 6 via the display drive control unit 104DR, and confirms whether the input number string is as intended by the user US. You may do it.
  • the user US speaks “No” while the display of FIG. 6 is being performed, a voice signal is input from the microphone 105 via the voice processing unit 106 to the processing unit 112, and the displayed number string is inappropriate.
  • the processing unit 112 determines that it is a thing and requests the utterance of the user US again.
  • the processing unit 112 determines that the displayed number string is appropriate, and performs the subsequent processing.
  • the processing unit 112 that has received the number string information from the voice recognition unit 113 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the number string NA stored in itself. More specifically, when the numeric strings NA are connected in the order of the numeric strings (“5, 3, 6, 2, 9”) converted by the speech recognition unit 113, the locus coincides with the pattern PT. In this case, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. After releasing the screen lock, the pattern PT stored in the authentication code storage unit 114 may be updated through similar authentication. On the other hand, if the locus does not match the pattern PT, the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested.
  • FIG. 7 is a diagram illustrating an example of a character string CA generated by the processing unit 112 instead of a numeric string.
  • the arrow AR is drawn so as to be easy to understand, but it is not actually displayed.
  • the user US stores the pattern PT as the same authentication code
  • the stored pattern PT is applied to the character string CA displayed as shown in FIG. Is read as “C, H, G, D, E”
  • the pattern PT can be reproduced. Therefore, when the user US utters the alphabet in the form of “Shi-Eichi-Ji-Die-I”, the microphone 105 collects the sound and passes through the voice processing unit 106.
  • a voice signal is input to the voice recognition unit 113, converted into a character string “C ⁇ H ⁇ G ⁇ D ⁇ E”, and transmitted to the processing unit 112 as character string information.
  • the processing unit 112 reads the pattern PT stored in the authentication code storage unit 114 and applies it to the numeric string NA stored in the processing unit 112 to determine whether or not the authentication codes match.
  • Other configurations are the same as those in the above-described embodiment.
  • FIG. 8 is a diagram showing an arrangement GA of vegetable or fruit patterns generated by the processing unit 112 instead of a numeric string or a character string.
  • the arrow AR is drawn so as to be easily understood, but it is not actually displayed.
  • the processing unit 112 registers image names in advance in association with individual images. The sequence of images is random.
  • the user US stores the pattern PT as the same authentication code
  • the stored pattern PT is applied to the displayed image array GA as shown in FIG.
  • the pattern PT can be reproduced by connecting images of “tomato, corn, mandarin orange, persimmon, green pepper”. Therefore, when the user US utters a word of a pattern in the form of “Tomato, corn, mandarin orange, kaki, peaman”, the microphone 105 collects the sound, and the voice recognition unit 113 via the voice processing unit 106 collects the words. Is converted to a character string “Tomato, corn, mandarin orange, kaki, piman” and transmitted to the processing unit 112 as character information. In the user's utterance, it is preferable to provide a certain silent time between words because it is easy to convert to an appropriate word when performing speech recognition.
  • the processing unit 112 selects the tomato pattern TO when the character information of “Tomato” matches the name of the tomato as the display image, and the character information of “corn” matches the name of the corn as the display image
  • the corn pattern CR when the character information of “mandarin orange” matches the name of the mandarin orange as the display image, select the pattern orange of the mandarin orange, and the character information of “oyster” is displayed as the display image. If the name matches the name of the rice cake, the image PR of the rice cake is selected, and if the character information of “Piman” matches the name of the green pepper as the display image, the image of the green pepper is selected and the selected image is selected.
  • the processing unit 112 specifies and selects an image corresponding to a character / number string from a plurality of display images, and performs processing based on the selected image.
  • FIG. 9A is a diagram showing a number string (here, “4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored as authentication code information (password that is a character / number string) in the authentication code storage unit 114. is there.
  • FIG. 9B is a diagram showing a combination image NG of numbers and images generated by the processing unit 112 together with the explanation image RG.
  • the combined image NG shown in FIG. 9B is an image in which colored blocks are arranged in 3 rows and 3 columns, and the two are associated by arranging a number in the center.
  • each color is represented by hatching or vertical and horizontal lines as shown in the adjacent explanatory image RG.
  • the numbers always include those constituting the number string stored in the authentication code storage unit 114.
  • the combination of numbers and colors is random.
  • the combination of each color and the corresponding number constitutes a predetermined relationship.
  • the explanation image RG is added to indicate the color corresponding to the block diagram, and is not actually displayed.
  • the processing unit 112 transmits information on the generated combination image NG to the display drive control unit 104DR. Then, the display drive control unit 104DR converts the information of the combination image NG into an image signal and transmits it to the display unit 104. Therefore, the display unit 104 displays the combination image NG (excluding the explanation image RG) shown in FIG. Can be displayed.
  • the control unit 110 performs a display requesting to input an authentication code on the display unit 104 via the display drive control unit 104DR, and FIG.
  • the combined image NG shown in b) is displayed.
  • the user US Since the user US stores a numerical string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG.
  • the color of the block corresponding to “4” is yellow
  • the color of the block corresponding to the number “9” is green
  • the color of the block corresponding to the number “1” is blue
  • the color of the block is yellow
  • the color of the block corresponding to the number “5” is blue. Therefore, when the user US utters the word “Ki / Midori / Ao / Ki / Ao”, the microphone 105 collects the sound and inputs the audio signal to the audio recognition unit 113 via the audio processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) of “Ki / Midori / Ao / Ki / Ao” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number string
  • the processing unit 112 that also serves as the conversion device reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the combined image NG generated by itself. Then, the characters “ki, green, ao, ki, ao” are picked up from the corresponding image color and converted to a character string (called a second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100.
  • the processing unit 112 determines that the authentication codes do not match, and continues the screen lock of the HMD 100. At this time, input of a new authentication code may be requested.
  • Other configurations are the same as those in the above-described embodiment.
  • FIG. 10A to 10 (c) are diagrams showing a combination image NG of numbers and images generated by the processing unit 112 together with an explanation image RG.
  • the processing unit 112 displays the combined image NG illustrated in FIG. 10A on the display unit 104 via the display drive control unit 104DR.
  • the combination of numbers and colors is random.
  • the user US Since the user US stores a numeric string (“1, 2, 3, 4”) as an update code, when viewing the combined image NG as shown in FIG.
  • the color of the block corresponding to is blue, the color of the block corresponding to the number “2” is red, the color of the block corresponding to the number “3” is yellow, and the block corresponding to the number “4” It can be seen that the color is yellow. Therefore, when the user US utters a word of the color “Ao / Aka / Ki / Ki”, the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / number string) “Ao, Aka, Ki, Ki” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number string
  • the processing unit 112 reads the numeric string (“1, 2, 3, 4”) stored as the update code in the authentication code storage unit 114, and generates the combined image NG (FIG. 10). Applying to (a), the characters “Ao, Aka, Ki, Ki” are picked up from the color of the corresponding image, and converted into a character string (second character / number string) arranged in this order. Further, when the first character / numeric string and the second character / numeric string match, the processing unit 112 determines that the update codes match and permits the password update. The processing unit 112 that has permitted the update of the password generates a combined image NG in which the correspondence between the numbers and the colors is changed as shown in FIG. 10B, and displays the display unit 104 via the display drive control unit 104DR. To display.
  • the color of the block corresponding to the number “9” is yellow.
  • the color of the block corresponding to the number “8” is purple
  • the color of the block corresponding to the number “5” is green
  • the color of the block corresponding to the number “6” is red.
  • the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an updated character string “Ki / Murasaki / Midori / Aka” and transmitted to the processing unit 112 as updated character string information.
  • the processing unit 112 collates with the combined image NG shown in FIG. 10B to determine that the character “ki” in the updated character string is yellow and the corresponding number is “7, 9”.
  • the character “Murasaki” indicates purple, the corresponding number is determined to be “4,8”, the character “Midori” indicates green, the corresponding number is determined to be “5”, and the character “Red” indicates red, and the corresponding numbers are determined to be “1, 6”.
  • the processing unit 112 has a plurality of password candidates that the user US desires to update, which are “7 ⁇ 4 ⁇ 5 ⁇ 1”, “9 ⁇ 4 ⁇ 5 ⁇ 1”, “7 ⁇ 8 ⁇ 5 ⁇ 1”.
  • the processing unit 112 stores these eight numeric strings as password candidates. As described above, the processing unit 112 identifies a numeric string corresponding to a character in the character string, and performs processing based on the identified numeric string.
  • the processing unit 112 newly generates a combination image NG in which the correspondence between numbers and colors is changed as illustrated in FIG. 10C and displays the combined image NG again on the display unit 104 via the display drive control unit 104DR.
  • the color of the block corresponding to the number “9” is blue, and the color of the block corresponding to the number “8” is green. It can be seen that the color of the block corresponding to the number “5” is yellow and the color of the block corresponding to the number “6” is green.
  • the microphone 105 collects the sound, and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into an update character string “Ao / Midori / Ki / Midori” and transmitted to the processing unit 112 as update character string information.
  • the processing unit 112 collates with the combination image NG shown in FIG. 10C to determine that the character “AO” in the updated character string is blue and the corresponding number is “2, 9”.
  • the character “Midori” indicates green, and the corresponding number is determined to be “6, 8”, the character “K” indicates yellow, and the corresponding number is determined to be “1, 5”. .
  • the processing unit 112 uses the numeric string “9, 8, 5, 6” as a new password.
  • the password in the authentication code storage unit 114 is updated. If a plurality of password candidates remain in the second process, the processing unit 112 may further display a new combined image and obtain the user's US utterance.
  • the number of colors used for the combined image is limited to five so that it can be easily distinguished. Therefore, the user US has to speak a plurality of times for password input. If the number of passwords is matched one-on-one, the user's utterance can be done only once. Alternatively, the number of images and the number of passwords can be matched on a one-to-one basis by using characters and patterns as in the following embodiment without using color types as combination images. Similarly, one utterance is sufficient.
  • FIG. 11 is a diagram showing a correspondence table CT in which characters are arranged in association with a predetermined relationship instead of colored blocks in correspondence with the numbers constituting the number string stored in the authentication code storage unit 114.
  • the correspondence table CT in FIG. 11 is displayed on the display unit 104.
  • the user US since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the correspondence table CT as shown in FIG. Is the character corresponding to the number “9”, the character corresponding to the number “9” is “no”, the character corresponding to the number “1” is “no”, and the character corresponding to the number “3”. Is “ka”, and the character corresponding to the number “5” is “ma”. Therefore, when the user US speaks hiragana “ko, ta, no, ka, or ma”, the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. It is converted into a character string (referred to as a first character / numerical string) of “ko, ta, no, koma” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / numerical string
  • the processing unit 112 reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 11) generated by itself.
  • the correspondence table CT (FIG. 11) generated by itself.
  • a character string arranged in this order referred to as a second character / number string.
  • the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100.
  • Other configurations are the same as those in the above-described embodiment.
  • the microphone 105 collects the sound and a voice signal is input to the voice recognition unit 113 via the voice processing unit 106. Is converted into a character string (referred to as a third character / number string) and transmitted to the processing unit 112 as character string information.
  • the processing unit 112 compares the character string “ko, ta, no, ka, or ma” with the correspondence table CT (FIG. 11) generated by the processing unit 112 to convert the character string into numbers corresponding to hiragana, respectively. It is converted to a numeric string “4 ⁇ 9 ⁇ 1, 3 ⁇ 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • FIG. 12 shows a correspondence table CT in which vegetables or fruit patterns are arranged in association with a predetermined relationship in place of colored blocks or characters, corresponding to the numbers constituting the number string stored in the authentication code storage unit 114.
  • FIG. 12 the names of vegetables or fruits are registered in association with the displayed patterns of vegetables or fruits, but this is not always necessary.
  • the correspondence table CT shown in FIG. 12 is displayed on the display unit 104.
  • the user US since the user US stores a numeric string (“4, 9, 1, 3, 5”) as an authentication code, when viewing the combined image NG as shown in FIG.
  • the pattern corresponding to "" is shiitake, the pattern corresponding to the number “9” is “green pepper”, the pattern corresponding to the number “1” is “tomato”, and the pattern corresponding to the number “3” Is “cherry” and the pattern corresponding to the number “5” is “mandarin orange”. Therefore, when the user US utters the word “Shitake / Piman / Tomato / Sakurabo / Mikan”, the microphone 105 collects the sound and inputs the speech signal to the speech recognition unit 113 via the speech processing unit 106. Then, it is converted into a character string (referred to as a first character / number string) “Shitake / Piman / Tomato / Cherry Bamboo / Mikan” and transmitted to the processing unit 112 as character string information.
  • a character string referred to as a first character / number
  • the processing unit 112 reads the numeric string (“4 ⁇ 9 ⁇ 1 ⁇ 3 ⁇ 5”) stored in the authentication code storage unit 114 and generates the correspondence table CT (FIG. 12) generated by itself. By collating with each other, the characters “Shitake / Piman / Tomato / Sakurambo / Mikan” corresponding to the numbers are picked up and converted into a character string (second character / number string) arranged in this order. Further, when the first character / number string and the second character / number string match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • the processing unit 112 converts the character string “Shitake / Pi-Man / Tomato / Cherry Bamboo / Tangerine” into a correspondence table CT (FIG. 11) generated by itself, thereby converting the character string into a number corresponding to the design. It is converted to a numeric string “4 ⁇ 9 ⁇ 1, 3 ⁇ 5” (referred to as a fourth character / numeric sequence) arranged in this order. Further, when the fourth character / number string and the password match, the processing unit 112 determines that the authentication codes match and releases the screen lock of the HMD 100. Other configurations are the same as those in the above-described embodiment.
  • FIG. 13 is a flowchart showing a control operation (except for step S102) of the processing unit 112 according to the above-described embodiment.
  • the processing unit 112 as an image, includes a numeric string NA (FIG. 5), a character string CA (FIG. 7), an image arrangement GA (FIG. 8), a combined image NG (FIG. 9B), and a correspondence table CT. (FIGS. 11 and 12) are displayed.
  • step S102 When the user US who has seen the display image speaks in response to this (step S102), the voice recognition unit 113 recognizes the character / numeric string represented by the voice and performs voice recognition (step S103). 112 displays the audio processing result as shown in step S104 (see FIG. 6).
  • step S105 if the voice processing result is inappropriate (NO in step S105), the flow returns to step S102, and the same processing is repeated.
  • the processing unit 112 reads the authentication code in step S106, collates the voice recognition result and the authentication code in step S107, and the two match. Process whether or not.
  • the processing unit 112 displays a message such as “authentication failed” on the display unit 104 in step S109, Continue screen lock.
  • step S108 the processing unit 112 releases the screen lock because the authentication is successful.
  • biometric authentication can identify an individual user, there is a demerit that biometric information as a release code is difficult to share with multiple people. In other words, it can be shared by registering biometric information for a plurality of people, but it takes time to register, and it is expected that a single electronic device will be shared by a plurality of people, especially when an electronic device is used in a factory. In such a case, there is a request to increase the convenience of sharing by using a common password. According to the present embodiment, there is an advantage that it can be used with almost the same user load as conventional password authentication. Speech recognition is a user interface that can be input hands-free, and has an advantage of high affinity with HMD.
  • the HMD is mounted on the user's head, the image displayed on the display unit is difficult for others to see, so there is an advantage that confidentiality can be kept high. Even if someone remembers the character string that the user utters, if a different image (or a combination of a character / number string and an image) is displayed for each display, the learned character string will be displayed at a different timing. Even if another person speaks, the electronic device is not unlocked and strong security can be secured. In addition, there is no need for the process of displaying the input characters etc. as the hidden characters “******”, which is performed when inputting the password to the conventional electronic device. If the image displayed on the display unit is not visible to others, a code reversely converted when displaying the result of voice recognition may be displayed.
  • the HMD 100 of the present embodiment can also be used as information input means.
  • the processing unit 112 when the user US tries to make a call to a telephone number “030-1234-5678”, the processing unit 112 generates a correspondence table CT as shown in FIG. 11 in response to a request from the user US. Displayed on the display unit 104.
  • the user US utters the characters corresponding to the telephone number as “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i”.
  • the processing unit 112 uses the correspondence table CT (FIG. 11) generated by itself to process the input character string “ke, ka, ke, no, ru, ka, ko, ma, mi, ni, i” It is converted into a number string (second character / number string) of “0, 3, 0, 1, 2, 3, 4, 5, 6, 7, 8”. Further, when the processing unit 112 has a telephone function, the processing unit 112 can make a call by inputting the obtained numeric string as a telephone number. In such a case, the processing unit 112 also serves as an input device. As a result, the user US can make a hands-free call without knowing the other party's telephone number. The converted telephone number may be displayed on the display unit 104 to make a call after confirming the user US.
  • the HMD 100 may be used for inputting not only a telephone number but also a My Number, a credit card number, and the like.
  • the present invention is not limited to the embodiments described in the specification, and other embodiments and modifications are included for those skilled in the art from the embodiments and technical ideas described in the present specification. it is obvious.
  • the description and the embodiments are for illustrative purposes only, and the scope of the present invention is indicated by the following claims.
  • the present invention has been described by taking the HMD as an example.
  • the present invention is not limited to the HMD and can be applied to all electronic devices such as portable terminals. Further, some or all of the functions permitted by the authentication described above may be used.
  • the screen lock is released as an example of permitting a predetermined function by inputting a password.
  • the application can be started by appropriately inputting a password as described above in a state where a login screen at the time of starting the application is displayed on the display unit 104.
  • it can also authenticate in it. In such a case, it is desirable to change to the authentication screen during a hands-free operation using the user's utterance without using an operation to turn on the switch.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Security & Cryptography (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

La présente invention concerne un appareil électronique, un visiocasque, un procédé de traitement de l'appareil électronique, et un programme associé avec lequel il est possible de supprimer les informations de fuite vers une tierce personne qui a entendu l'énoncé d'un utilisateur. L'appareil électronique possède un dispositif de stockage destiné à stocker un modèle, un dispositif d'affichage destiné à afficher une pluralité d'images, un dispositif de reconnaissance de la parole destiné à acquérir la parole énoncée par un utilisateur en réponse aux images et à convertir ces dernières en une chaîne numérique/de caractère correspondante, et un dispositif de traitement destiné à autoriser une fonction prescrite lorsque la pluralité d'images affichées sur le dispositif d'affichage sont liées dans l'ordre de la chaîne numérique/de caractères convertie par le dispositif de reconnaissance de la parole et un locus associé correspondant au modèle stocké dans le dispositif de stockage.
PCT/JP2017/021067 2016-06-14 2017-06-07 Appareil électronique, visiocasque, procédé de traitement d'appareil électronique et programme associé WO2017217288A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2016118046 2016-06-14
JP2016-118046 2016-06-14

Publications (1)

Publication Number Publication Date
WO2017217288A1 true WO2017217288A1 (fr) 2017-12-21

Family

ID=60663453

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/021067 WO2017217288A1 (fr) 2016-06-14 2017-06-07 Appareil électronique, visiocasque, procédé de traitement d'appareil électronique et programme associé

Country Status (1)

Country Link
WO (1) WO2017217288A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008257701A (ja) * 2007-03-12 2008-10-23 Yahoo Japan Corp 認証システム
JP2010009544A (ja) * 2008-06-30 2010-01-14 Toppan Forms Co Ltd 本人認証システムおよび本人認証方法
JP2014092941A (ja) * 2012-11-02 2014-05-19 Sony Corp 情報処理装置及び情報処理方法、並びにコンピューター・プログラム
CN104468522A (zh) * 2014-11-07 2015-03-25 百度在线网络技术(北京)有限公司 一种声纹验证方法和装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008257701A (ja) * 2007-03-12 2008-10-23 Yahoo Japan Corp 認証システム
JP2010009544A (ja) * 2008-06-30 2010-01-14 Toppan Forms Co Ltd 本人認証システムおよび本人認証方法
JP2014092941A (ja) * 2012-11-02 2014-05-19 Sony Corp 情報処理装置及び情報処理方法、並びにコンピューター・プログラム
CN104468522A (zh) * 2014-11-07 2015-03-25 百度在线网络技术(北京)有限公司 一种声纹验证方法和装置

Similar Documents

Publication Publication Date Title
US10360412B2 (en) Contextual contemporaneous gesture and keyboard entry authentication
US9503800B2 (en) Glass-type terminal and method of controlling the same
US8090201B2 (en) Image-based code
US9275213B2 (en) Method and system for securing the entry of data to a device
EP2851831B1 (fr) Passerelle mobile d'information de soins de santé à domicile
US8873147B1 (en) Chord authentication via a multi-touch interface
US20060039686A1 (en) Line-of-sight-based authentication apparatus and method
US20150088546A1 (en) Mobile Information Gateway for Use by Medical Personnel
KR102393892B1 (ko) 생체정보를 이용한 사용자인증을 수행하는 단말 장치 및 방법
US9336779B1 (en) Dynamic image-based voice entry of unlock sequence
JP2014092940A (ja) 画像表示装置及び画像表示方法、並びにコンピューター・プログラム
US9552471B1 (en) Personal familiarity authentication
JP2008241822A (ja) 画像表示装置
CN105900103A (zh) 触摸终端机以及触摸终端机的密码生成方法
WO2017217288A1 (fr) Appareil électronique, visiocasque, procédé de traitement d'appareil électronique et programme associé
Saulynas et al. Towards the use of brain–computer interface and gestural technologies as a potential alternative to PIN authentication
US20160224808A1 (en) Information input method having confidentiality
WO2016200084A1 (fr) Dispositif usb de reconnaissance d'iris au moyen d'une fonction otp et son procédé de commande
US11075920B2 (en) Providing access to structured stored data
KR20190033377A (ko) 이미지 터치 패스워드를 이용한 사용자 인증 방법 및 이를 위한 컴퓨터 프로그램
Tkauc et al. Cloud-Based Face and Speech Recognition for Access Control Applications
KR20050090102A (ko) 번호 숨김 비밀번호 입력 시스템
TW200841204A (en) Hand cryptographic device
TW201617951A (zh) 密碼輸入的裝置及方法
TWI644232B (zh) 密碼輸入的裝置及方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17813186

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17813186

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP