WO2023210158A1 - Eyeglass-type display device and display system - Google Patents

Eyeglass-type display device and display system Download PDF

Info

Publication number
WO2023210158A1
WO2023210158A1 PCT/JP2023/007812 JP2023007812W WO2023210158A1 WO 2023210158 A1 WO2023210158 A1 WO 2023210158A1 JP 2023007812 W JP2023007812 W JP 2023007812W WO 2023210158 A1 WO2023210158 A1 WO 2023210158A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
display
glasses
unit
display device
Prior art date
Application number
PCT/JP2023/007812
Other languages
French (fr)
Japanese (ja)
Inventor
信貴 松嶌
勇一 水越
Original Assignee
株式会社Nttドコモ
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Nttドコモ filed Critical 株式会社Nttドコモ
Publication of WO2023210158A1 publication Critical patent/WO2023210158A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer

Definitions

  • the present invention relates to a glasses-type display device and a display system.
  • An HMD Head Mounted Display
  • a virtual object representing additional information such as explanatory text regarding the real object, superimposed on an object (hereinafter referred to as a real object) in a real-world view.
  • Some HDM devices such as AR (Augmented Reality) glasses or MR (Mixed Reality) glasses, display virtual objects superimposed on real space without blocking the user's field of view (for example, Patent Document 1 and Patent Document 2).
  • JP2022-029429A Japanese Patent Application Publication No. 2014-093050
  • the immersion or convenience may be impaired if there is unnecessary light in the area that includes the real object that the user is gazing at (hereinafter referred to as the attention area).
  • the attention area There is a problem.
  • the entire field of view is always blocked, there is a problem in that the user cannot visually grasp the situation in the surrounding real space.
  • display control such as blocking light from the real space and displaying a virtual object only in a region of interest specified by the user using a mouse or keyboard.
  • a mouse or keyboard to specify the attention area reduces user convenience.
  • a glasses-type display device including a transmissive display section on which a virtual object is displayed includes a specifying section and a display control section.
  • the specifying unit specifies an attention area including a real object that the user is gazing at in the user's field of view based on the voice uttered by the user.
  • the display control unit displays a virtual object corresponding to the real object in a display area of the display unit that overlaps the attention area of the field of view.
  • a display system includes a glasses-type display device that is attached to a user's head and includes a transmissive display section on which a virtual object is displayed, a specific section, and a display control section. Be prepared.
  • the specifying unit specifies an attention area including a real object that the user is gazing at in the user's field of view based on the voice uttered by the user.
  • the display control unit displays a virtual object corresponding to the real object in a display area of the display unit that overlaps the attention area of the field of view.
  • the virtual object can be displayed over the attention area without impairing the user's convenience.
  • FIG. 1 is a block diagram illustrating a configuration example of a display system 1 according to an embodiment of the present disclosure. It is a diagram showing an example of a poster P on which real objects are arranged. It is a figure showing an example of management table TBL in this indication.
  • 1 is a block diagram showing a configuration example of a mobile device 10.
  • FIG. 2 is a flowchart showing the flow of a display method executed by the processing device 18 of the mobile device 10 according to the program PR1.
  • 2 is a block diagram showing a configuration example of a glasses-type display device 20.
  • FIG. FIG. 3 is a diagram for explaining the operation of this embodiment.
  • FIG. 3 is a diagram for explaining the operation of this embodiment.
  • 7 is a diagram illustrating an example of an image viewed by user U in modification example 1.
  • FIG. 7 is a diagram illustrating an example of an image viewed by user U in modification 2.
  • FIG. 1 is a block diagram illustrating a configuration example of a display system 1 according to an embodiment
  • FIG. 1 is a block diagram showing a configuration example of a display system 1 according to an embodiment of the present disclosure.
  • the display system 1 includes a mobile device 10 and a glasses-type display device 20.
  • the glasses-type display device 20 is attached to the head of the user U.
  • the glasses-type display device 20 is an HMD device that displays virtual objects that do not exist in real space without blocking the field of view of the user U wearing the glasses-type display device 20.
  • the glasses-type display device 20 has an imaging function.
  • the glasses-type display device 20 mounted on the head of the user U uses an imaging function to capture an image of real space corresponding to the field of view of the user U.
  • An object existing in the field of view of the user U that is, a real object, appears in the image captured by the glasses-type display device 20 worn on the head of the user U.
  • the mobile device 10 is, for example, a smartphone.
  • the mobile device 10 is worn on the user U's body.
  • the mobile device 10 is attached to the body of the user U by hanging from the neck using a strap or the like.
  • the mobile device 10 has a sound collection function.
  • the mobile device 10 worn on the user's body collects the voice emitted by the user U using a sound collection function.
  • the mobile device 10 is connected by wire to a glasses-type display device 20 that is worn on the head.
  • the mobile device 10 may be connected to the eyeglass-type display device 20 wirelessly.
  • the mobile device 10 acquires image data representing an image captured by the glasses-type display device 20 from the glasses-type display device 20 .
  • the mobile device 10 is not limited to a smartphone, and may be, for example, a tablet or a notebook personal computer.
  • the mobile device 10 communicates with the management device 30 via the communication network NW.
  • the mobile device 10 transmits the image data acquired from the glasses-type display device 20 to the management device 30.
  • the management device 30 is a server device that provides a location recognition service and a content management service in AR.
  • the position recognition service is a service that specifies the position of the glasses-type display device 20 in the global coordinate system based on an image captured by the imaging function of the glasses-type display device 20.
  • Specific implementation modes of the location recognition service include a mode using an AR tag or a mode using a distribution of feature points extracted from an image, such as SLAM (Simultaneous Localization and Mapping).
  • the content management service is a service that distributes information regarding virtual objects corresponding to one or more real objects visible from the position of the glasses-type display device 20 in the global coordinate system to the glasses-type display device 20.
  • the management device 30 stores virtual object information representing an image of the virtual object, and information about the virtual object corresponding to each of one or more real objects visible from the position in association with the position in the global coordinate system. Area information indicating the position and size of the display area to be displayed is stored in advance.
  • the management device 30 specifies one or more real objects from the position of the glasses-type display device 20 specified based on the image data received from the mobile device 10 via the communication network NW. Then, virtual object information and area information corresponding to each of the identified one or more real objects are sent back to the mobile device 10.
  • the mobile device 10 causes the glasses-type display device 20 to display an image of the virtual object according to the virtual object information and area information received from the management device 30. As a result, the virtual object appears superimposed on the real space in the eyes of the user U.
  • the real space in this embodiment is, for example, a venue for a poster session at a research presentation such as an academic conference.
  • the real objects in this embodiment are, for example, each English word on a poster displayed in a poster session venue and in which research content is written in English.
  • FIG. 2 is a diagram showing an example of posters P displayed in the poster session venue.
  • the virtual object in this embodiment is a character string representing a Japanese translation of an English word written on the poster P.
  • the management device 30 transmits virtual object information and area information to the mobile device 10 by transmitting the management table TBL shown in FIG. 3 to the mobile device 10.
  • the management table TBL stores virtual object information and area information about a virtual object corresponding to the real object in association with identification information for identifying the real object.
  • the identification information in this embodiment is character string data representing the pronunciation of the real object identified by the identification information. For example, assume that the English word "Patients" is a real object. The identification information in this case is "patients" in katakana. Katakana are phonetic characters used in Japanese to represent the pronunciation of foreign words.
  • the virtual object information corresponding to this identification information represents an image of the Japanese character string "Kikin" which represents the Japanese translation of the English word "Patients”.
  • the background of the Japanese character string is painted over with a predetermined color such as white to block light from real space.
  • the area information stored in the management table TBL in association with the identification information "Patients" represents an area that overlaps with the real object "Patients”.
  • FIG. 4 is a block diagram showing a configuration example of the mobile device 10.
  • the mobile device 10 includes an input device 11, an output device 12, a microphone 13, a communication device 14, a communication device 15, a storage device 17, a processing device 18, and a bus 19. ,including.
  • the input device 11, the output device 12, the microphone 13, the communication device 14, the communication device 15, the storage device 17, and the processing device 18 are interconnected by a bus 19 for communicating information.
  • the bus 19 may be configured using a single bus, or may be configured using different buses for each device.
  • the input device 11 includes a touch panel.
  • the input device 11 may include a plurality of operation keys in addition to a touch panel.
  • the input device 11 may include a plurality of operation keys without including a touch panel.
  • the input device 11 receives operations performed by the user U.
  • Output device 12 includes a display. A touch panel of the input device 11 is stacked on the display of the output device 12 .
  • the output device 12 displays various information.
  • the microphone 13 picks up user U's voice.
  • the microphone 13 generates sound data indicating the waveform of the collected sound and outputs it to the processing device 18 .
  • the user U's attention area is specified based on the user's U voice picked up by the microphone 13.
  • the communication device 14 is hardware (transmission/reception device) for communicating with the management device 30 via the communication network NW.
  • the communication device 14 is also called, for example, a network device, a network controller, a network card, a communication module, or the like.
  • the communication device 14 transmits the image data given from the processing device 18 to the management device 30. Further, the communication device 14 supplies the management table TBL received from the management device 30 to the processing device 18. Note that the communication device 14 may communicate with the management device 30 without going through the communication network NW.
  • the communication device 15 is hardware (transmission/reception device) for communicating with the eyeglass-type display device 20 by wire.
  • the communication device 15 supplies the image data received from the glasses-type display device 20 to the processing device 18 .
  • the communication device 15 transmits image data provided from the processing device 18 to the glasses-type display device 20.
  • the communication device 15 may communicate with the glasses-type display device 20 wirelessly.
  • the storage device 17 is a recording medium that can be read by the processing device 18.
  • the storage device 17 includes, for example, nonvolatile memory and volatile memory.
  • Nonvolatile memories include, for example, ROM (Read Only Memory), EPROM (Erasable Programmable Read Only Memory), and EEPROM (Electrically Erasable Programmable Read Only Memory).
  • the volatile memory is, for example, RAM (Random Access Memory).
  • the storage device 17 stores in advance a program PR1 that causes the processing device 18 to execute the method of specifying a region of interest according to the present disclosure. Furthermore, the management table TBL received from the management device 30 is written into the storage device 17 by the processing device 18 .
  • the processing device 18 includes one or more CPUs (Central Processing Units). One or more CPUs are an example of one or more processors. Each of the processor and CPU is an example of a computer.
  • the processing device 18 reads the program PR1 from the storage device 17.
  • the processing device 18 operating according to the program PR1 transmits the image data received from the glasses-type display device 20 using the communication device 15 to the management device 30 using the communication device 14. Further, the processing device 18 operating according to the program PR1 writes the management table TBL received from the management device 30 into the storage device 17 using the communication device 14.
  • the processing device 18 operating according to the program PR1 functions as the speech recognition section 181, the identification section 182, and the display control section 183 shown in FIG. That is, the speech recognition section 181, the identification section 182, and the display control section 183 in FIG. 4 are software modules realized by operating the processing device 18 according to software.
  • the speech recognition unit 181 converts the speech represented by the sound data generated by the microphone 13 into a character string. That is, the speech recognition unit 181 performs speech recognition on the user U's speech according to a predetermined speech recognition algorithm. Existing technology may be adopted as appropriate for the speech recognition algorithm.
  • the voice recognition unit 181 generates recognized character string data representing the result of voice recognition of the user U's voice, that is, a character string of one or more words uttered by the user.
  • the identification unit 182 identifies the region of interest based on the recognition result by the voice recognition unit 181. More specifically, the identification unit 182 manages whether or not the character string represented by the recognized character string data generated by the speech recognition unit 181 matches any identification information stored in the management table TBL. The determination is made with reference to the table TBL. When the character string represented by the recognized character string data matches any identification information, the identification unit 182 identifies the area represented by the area information corresponding to the identification information that matches the character string represented by the recognized character string data as a region of interest. do.
  • the display control unit 183 causes a display area of the display unit of the glasses-type display device 20 that overlaps with the area of interest specified by the specifying unit 182 to display a virtual object corresponding to the area of interest. More specifically, the display control unit 183 displays an image in which an image representing virtual object information corresponding to identification information that matches the character string represented by the recognized character string data is arranged in an area indicated by area information corresponding to the identification information. Generate image data representing. Then, the display control unit 183 transmits the image data to the eyeglass-type display device 20 using the communication device 15, thereby causing the eyeglass-type display device 20 to display an image represented by the image data.
  • the processing device 18 operating according to the program PR1 executes the display method shown in FIG. 5 every time sound data is output from the microphone 13. As shown in FIG. 5, this display method includes each process of step SA110 to step SA140.
  • step SA110 the processing device 18 functions as the speech recognition unit 181.
  • step SA110 the processing device 18 generates recognized character string data by performing voice recognition on the voice represented by the sound data output from the microphone 13.
  • step SA120 and step SA130 the processing device 18 functions as the specifying unit 182.
  • step SA120 the processing device 18 checks whether the character string represented by the recognized character string data generated in step SA110 matches any identification information stored in the management table TBL. Determine by referring to.
  • step SA130 the processing device 18 specifies, as a region of interest, the region indicated by the region information stored in the management table TBL in association with the identification information that matches the character string represented by the recognized character string data generated in step SA110. do. If the determination result in step SA120 is "No", that is, if the character string represented by the recognized character string data does not match any identification information, the processing device 18 does not execute the processing from step SA130 onwards. , this display method ends.
  • step SA140 the processing device 18 functions as the display control unit 183.
  • the processing device 18 acquires the virtual object information stored in the management table TBL in association with the identification information that matches the character string represented by the recognized character string data generated in step SA110, and Image data representing an image in which a virtual object representing information is placed in the region of interest specified in step SA130 is generated. Then, the processing device 18 supplies the generated image data to the eyeglass-type display device 20, thereby causing the eyeglass-type display device 20 to display an image represented by the image data.
  • FIG. 6 is a block diagram showing a configuration example of the eyeglass-type display device 20.
  • the eyeglass-type display device 20 includes a display section 2a, a communication device 2b, an imaging device 2c, a storage device 2d, a processing device 2e, and a bus 2f.
  • the display unit 2a, the communication device 2b, the imaging device 2c, the storage device 2d, and the processing device 2e are interconnected by a bus 2f for communicating information.
  • the bus 2f may be configured using a single bus, or may be configured using different buses for each element such as a device.
  • the display section 2a is a transmissive display section that transmits light. Light representing real space is transmitted through the display section 2a.
  • the display unit 2a displays images of virtual objects under the control of the processing device 2e.
  • the display section 2a is located in front of the user's U left and right eyes.
  • the user U wearing the glasses-type display device 20 visually recognizes the real space represented by the light transmitted through the display section 2a and the image of the virtual object displayed on the display section 2a.
  • the display unit 2a includes a left eye lens, a left eye display panel, a left eye optical member, a right eye lens, a right eye display panel, and a right eye optical member.
  • the display panel for the left eye and the display panel for the right eye are, for example, a liquid crystal panel or an organic EL (Electro Luminescence) panel.
  • the display panel for the left eye displays an image represented by image data provided from the processing device e.
  • the left eye optical member is an optical member that guides light emitted from the left eye display panel to the left eye lens.
  • the display panel for the right eye displays an image represented by image data provided from the processing device 2e.
  • the right eye optical member is an optical member that guides light emitted from the right eye display panel to the right eye lens.
  • Each of the left eye lens and the right eye lens has a half mirror.
  • the half mirror included in the left eye lens guides the light representing the real space to the left eye of the user U by transmitting the light representing the real space. Further, the half mirror included in the left eye lens reflects the light guided by the left eye optical member to the user U's left eye.
  • the half mirror included in the right eye lens guides the light representing the real space to the right eye of the user U by transmitting the light representing the real space.
  • the half mirror included in the right eye lens reflects the light guided by the right eye optical member to the user U's right eye.
  • the communication device 2b is hardware (transmission/reception device) for communicating with the mobile device 10 by wire.
  • the communication device 2b may communicate with the mobile device 10 wirelessly.
  • the glasses-type display device 20 has a glasses-shaped frame that supports a left eye lens and a right eye lens, and the imaging device 2c (for example, a camera) is provided on a bridge of the frame.
  • the imaging device 2c captures an image of the real space that the user U wearing the glasses-type display device 20 sees through the glasses-type display device 20, that is, an image of the user's U field of view under the control of the processing device 2e.
  • the imaging device 2c outputs image data representing the captured image to the processing device 2e.
  • the storage device 2d is a recording medium that can be read by the processing device 2e.
  • the storage device 2d like the storage device 17, includes nonvolatile memory and volatile memory.
  • the storage device 2d stores a program PR2.
  • the processing device 2e includes one or more CPUs.
  • the processing device 2e reads the program PR2 from the storage device 2d.
  • the processing device 2e functions as an operation control unit 2e1 by executing the program PR2.
  • the operation control unit 2e1 controls the operation of the eyeglass-type display device 20.
  • the operation control unit 2e1 transmits the image data output from the imaging device 2c to the mobile device 10 using the communication device 2b. Further, the operation control unit 2e1 supplies image data received from the mobile device 10 via the communication device 2b to the display unit 2a.
  • the display section 2a displays an image represented by the image data supplied from the operation control section 2e1.
  • the image represented by the image data transmitted by the mobile device 10 is an image in which a virtual object is placed in the attention area in the user's U field of view. Since this image is displayed on the display unit 2a, the user's U's eyes see an image of the real space in which the virtual object is superimposed on the attention area in the user's U's field of view.
  • FIG. 7 shows an example of the field of view A that the user U sees through the glasses-type display device 20.
  • the voice of the user U is collected by the microphone 13 of the mobile device 10, and based on the recognition result of the voice by the voice recognition unit 181, the area corresponding to "Patients" is specified as the area of interest.
  • Image data representing an image in which a virtual object corresponding to “Patients” Japanese translation of “Patients”
  • the user U's eyes see an image of the real space in which the virtual object VOB is superimposed on the attention area in the field of view A. .
  • the user U specify an attention area that includes the real object that the user U is gazing at, and to display a virtual object overlapping the attention area specified by the user U. .
  • the user U can specify the area of interest by voice, and there is no need to operate the mouse or keyboard when specifying the area of interest. Since there is no need to operate the mouse or keyboard when specifying the region of interest, convenience is not reduced. In this way, according to the present embodiment, it is possible to have the user U specify an attention area in the field of view and to display a virtual object in the attention area specified by the user U, without reducing the convenience of the user U. Become.
  • B Modification The present disclosure is not limited to the embodiments illustrated above. Specific aspects of the modification are as follows. Two or more aspects arbitrarily selected from the examples below may be combined.
  • B-1 Modification 1
  • the display control unit 183 may display a translucent image in a portion of the display area of the display unit 2a other than the attention area.
  • a translucent image is displayed in a part of the display area of the display unit 2a other than the attention area, a part of the light that passes through the part other than the attention area from the real space and enters the eyes of the user U is blocked. Ru.
  • the portion of the user's field of vision seen through the glasses-type display device 20 other than the area of interest becomes blurred, and the user U's sense of immersion is improved.
  • blurring in areas other than the region of interest is represented by diagonal hatching.
  • the glasses-type display device 20 may be configured to be able to partially control the transmittance of the left eye lens and the right eye lens. In this case, by controlling the transmittance, blocking of light that passes through a portion other than the region of interest may be realized.
  • the identification information may be a series of numbers assigned to each of the plurality of real objects.
  • the display control unit 183 may cause the glasses-type display device 20 to display identification information in association with each of the plurality of real objects, as shown in FIG. 10.
  • a series of numbers (circled numbers in the illustrated example) are assigned to each of a plurality of real objects.
  • the identification information is a series of numbers
  • the user U specifies the area of interest by speaking the number assigned to the real object of interest. For example, if the English word of interest is "Department", the number given to "Department” is "3", so user U can say "san” in Japanese or "three” in English. to specify the area of interest.
  • the identification unit 182 identifies the region of interest based on the recognized number.
  • the display control unit 183 may cause the glasses-type display device 20 to display a virtual object for instructing the user U to start a dictionary application (for example, an English-Japanese dictionary application).
  • a dictionary application for example, an English-Japanese dictionary application
  • the processing device 18 starts the English-Japanese dictionary application when sound data indicating the virtual object is provided from the microphone 13.
  • the audio that specifies the English word of interest is not limited to the audio that pronounces the English word, but may be the audio that reads out the alphabets that make up the English word in the order in which they are arranged in the English word. For example, if the English word of interest is "World", user U specifies the area of interest by saying "Double o r l di”.
  • the identification information may be character string information in which the pronunciations of the alphabets constituting the English word are arranged in the order in which they are arranged in the English word.
  • real objects in the present disclosure are not limited to English words, but may be words in other languages such as French, German, or Chinese.
  • the placement destination of the real object in the present disclosure is not limited to a poster, but may be a magazine, book, or newspaper, or may be a menu board at a restaurant, an instruction manual for equipment, various documents, or a signboard. There may be.
  • the glasses-type display device 20 may include a detection device that detects the user's U line of sight.
  • the specifying unit 182 roughly specifies the range occupied by the attention area in the field of view of the user U based on the line of sight detected by the detection device, and the display control unit 183 performs display control to emphasize the range.
  • the display control unit 183 may cause the display unit 2a to display an image in which the range occupied by the region of interest is expanded, or may cause the display unit 2a to display an image of a frame line surrounding the range.
  • the eyeglass-type display device 20 may be configured to allow the user U to specify an area of interest by touching a part of the field of view A in FIG. 7 with a fingertip of the user U.
  • the display control unit 183 may detect the area designation by the user U based on the image captured by the imaging device 2c, and highlight the detected area as a candidate for translation.
  • the program PR1 is stored in the storage device 17 of the mobile device 10, but the program PR1 may be manufactured or sold separately.
  • the provider of the program PR1 may write the program PR1 on a computer-readable recording medium such as a flash ROM and distribute it, or may distribute it by downloading it via a telecommunications line. .
  • the speech recognition unit 181, identification unit 182, and display control unit 183 in the above embodiment were software modules. However, any one, a plurality, or all of the voice recognition unit 181, the identification unit 182, and the display control unit 183 may be a hardware module. Specific examples of the hardware module include DSP (Digital Signal Processor), ASIC (Application Specific Integrated Circuit), PLD (Programmable Logic Device), FPGA (Field Programmable Gate Array), and the like.
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • PLD Programmable Logic Device
  • FPGA Field Programmable Gate Array
  • the mobile device 10 has the voice recognition section 181, the identification section 182, and the display control section 183, but the glasses-type display device 20 or the management device 30 has the voice recognition section 181, the identification section 182, and the display control section 183. It may also include a control section 183.
  • the voice recognition unit 181, the identification unit 182, and the display control unit 183 may be distributed and provided in any two or all of the glasses-type display device 20, the mobile device 10, and the management device 30. Note that since the voice recognition process by the voice recognition unit 181 is a process with a high processing load, it is preferable that the voice recognition unit 181 be provided in the management device 30 or the mobile device 10 rather than the glasses-type display device 20. Furthermore, the voice recognition unit 181 is preferably provided in the management device 30 rather than the mobile device 10.
  • the storage device 17 and the storage device 2d are exemplified as ROM, RAM, etc., but the storage device 17 and the storage device 2d are flexible disks, magneto-optical disks (for example, compact disks). , digital versatile discs, Blu-ray discs), smart cards, flash memory devices (e.g. cards, sticks, key drives), CD-ROMs (Compact Disc-ROMs), registers, removable disks, hard disks, It may be a floppy disk, magnetic strip, database, server or other suitable storage medium.
  • the information, signals, etc. described may be represented using any of a variety of different technologies.
  • data, instructions, commands, information, signals, bits, symbols, chips, etc. may refer to voltages, currents, electromagnetic waves, magnetic fields or magnetic particles, light fields or photons, or any of these. It may also be represented by a combination of
  • the input/output information may be stored in a specific location (for example, memory) or may be managed using a management table. Information etc. to be input/output may be overwritten, updated, or additionally written. The output information etc. may be deleted. The input information etc. may be transmitted to other devices.
  • the determination may be made based on a value represented by 1 bit (0 or 1), or may be made based on a truth value (Boolean: true or false). , may be performed by numerical comparison (for example, comparison with a predetermined value).
  • each function illustrated in FIG. 4 is realized by an arbitrary combination of at least one of hardware and software.
  • the method for realizing each functional block is not particularly limited. That is, each functional block may be realized using one physically or logically coupled device, or may be realized using two or more physically or logically separated devices directly or indirectly (e.g. , wired, wireless, etc.) and may be realized using a plurality of these devices.
  • the functional block may be realized by combining software with the one device or the plurality of devices.
  • the programs exemplified in the embodiments described above may include instructions, instruction sets, codes, software, firmware, middleware, microcode, hardware description language, or other names. Should be broadly construed to mean a code segment, program code, program, subprogram, software module, application, software application, software package, routine, subroutine, object, executable, thread of execution, procedure, function, etc. .
  • software, instructions, information, etc. may be sent and received via a transmission medium.
  • a transmission medium For example, if the software uses wired technology (coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), etc.) and/or wireless technology (infrared, microwave, etc.) to create a website, When transmitted from a server or other remote source, these wired and/or wireless technologies are included within the definition of transmission medium.
  • wired technology coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), etc.
  • wireless technology infrared, microwave, etc.
  • the information, parameters, etc. described in this disclosure may be expressed using absolute values, relative values from a predetermined value, or other corresponding information. It may also be expressed as
  • the mobile device includes a mobile station (MS).
  • MS mobile station
  • a mobile station is defined by a person skilled in the art as a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless It may also be referred to as a terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable terminology. Further, in the present disclosure, terms such as “mobile station,” “user terminal,” “user equipment (UE),” and “terminal” may be used interchangeably.
  • connection refers to direct or indirect connections between two or more elements. Refers to any connection or combination and may include the presence of one or more intermediate elements between two elements that are “connected” or “coupled” to each other.
  • the bonds or connections between elements may be physical, logical, or a combination thereof.
  • connection may be replaced with "access.”
  • two elements may include one or more electrical wires, cables, and/or printed electrical connections, as well as in the radio frequency domain, as some non-limiting and non-inclusive examples. , electromagnetic energy having wavelengths in the microwave and optical (both visible and non-visible) ranges, and the like.
  • determining and “determining” used in this disclosure may encompass a wide variety of operations.
  • “Judgment” and “decision” include, for example, judging, calculating, computing, processing, deriving, investigating, looking up, search, and inquiry. (e.g., searching in a table, database, or other data structure), and regarding an ascertaining as a “judgment” or “decision.”
  • judgment and “decision” refer to receiving (e.g., receiving information), transmitting (e.g., sending information), input, output, and access.
  • (accessing) may include considering something as a “judgment” or “decision.”
  • judgment and “decision” refer to resolving, selecting, choosing, establishing, comparing, etc. as “judgment” and “decision”. may be included.
  • judgment and “decision” may include regarding some action as having been “judged” or “determined.”
  • judgment (decision) may be read as “assuming", “expecting", “considering”, etc.
  • the glasses-type display device including the transmissive display section 2a on which a virtual object is displayed may include a specifying section 182 and a display control section 183.
  • the specifying unit 182 specifies a region of interest that includes a real object that the user U is gazing at in the user's field of view based on the voice uttered by the user U.
  • the display control unit 183 causes a virtual object corresponding to the real object to be displayed in a display area of the display unit 2a that overlaps with the attention area in the user's U field of view.
  • the virtual object can be displayed in the specified attention area without impairing the user U's convenience. I can do it.
  • the field of view of the user U in the second aspect may include a plurality of real objects. Furthermore, in the second aspect, one real object that the user U is gazing at may be specified from a plurality of real objects by the voice uttered by the user U.
  • the glasses-type display device according to the second aspect may further include a voice recognition unit 181 that recognizes the voice uttered by the user U.
  • the identification unit 182 in the glasses-type display device according to the second aspect may identify the attention area based on the recognition result of the voice recognition unit 181.
  • the glasses-type display device according to the second aspect can specify the attention area based on the recognition result of the voice recognition unit 181 for the user's U voice.
  • the display control unit 183 in the glasses-type display device according to the third aspect displays identification information that uniquely identifies each of the plurality of real objects in association with each of the plurality of real objects. It may be displayed in the display area of section 2a. Further, in the glasses-type display device according to the third aspect (an example of the second aspect), the identification unit 182 detects the voice when the voice recognition unit 181 recognizes the identification information corresponding to any one of the plurality of real objects. The region of interest may be specified based on the identification information recognized by the recognition unit 181.
  • the glasses-type display device determines the attention area based on the voice of the user U indicating any of a plurality of pieces of identification information displayed in the display area of the display unit 2a in association with each of a plurality of real objects. Can be identified.
  • the display system includes a glasses-type display device that is attached to the head of the user U and includes a transmissive display section 2a on which a virtual object is displayed, an identification section 182, a display control section 183, including.
  • the specifying unit 182 specifies a region of interest that includes a real object that the user U is gazing at in the user's field of view based on the voice uttered by the user U.
  • the display control unit 183 causes a virtual object corresponding to the real object to be displayed in a display area of the display unit 2a that overlaps with the attention area in the user's U field of view.
  • the virtual object can be displayed in the specified attention area without impairing the user U's convenience. .
  • SYMBOLS 1...Display system 10...Mobile device, 20...Glasses type display device, 11...Input device, 12...Output device, 13...Microphone, 14, 15, 2b...Communication device, 17, 2d...Storage device, 18, 2e . . . Processing device, 181 .

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)

Abstract

This eyeglass-type display device comprises a transparent display unit, a specification unit, and a display control unit. The specification unit specifies a focus area, which includes a real object gazed at by a user within the user's field of view through the display unit, on the basis of voices spoken by the user of the eyeglass-type display device. The display control unit displays a virtual object corresponding to the real object gazed at by the user on a display area that overlaps with the focus area within the user's field of view through the display unit.

Description

眼鏡型表示装置、及び表示システムEyeglass-type display device and display system
 本発明は、眼鏡型表示装置、及び表示システムに関する。 The present invention relates to a glasses-type display device and a display system.
 実空間(real-world view)における物体(以下、実オブジェクト)に重ねて、実オブジェクトに関する説明文等の付加情報を表す仮想オブジェクトを表示するHMD(Head Mounted Display)装置が一般に知られている。HDM装置の中には、AR(Augmented Reality)グラス又はMR(Mixed Reality)グラス等、ユーザの視界を遮蔽せずに実空間に重ねて仮想オブジェクトを表示するものもある(例えば、特許文献1及び特許文献2)。 An HMD (Head Mounted Display) device is generally known that displays a virtual object representing additional information such as explanatory text regarding the real object, superimposed on an object (hereinafter referred to as a real object) in a real-world view. Some HDM devices, such as AR (Augmented Reality) glasses or MR (Mixed Reality) glasses, display virtual objects superimposed on real space without blocking the user's field of view (for example, Patent Document 1 and Patent Document 2).
特開2022-029429号公報JP2022-029429A 特開2014-093050号公報Japanese Patent Application Publication No. 2014-093050
 ユーザの視界を遮蔽せずに仮想オブジェクトを表示するHMD装置には、ユーザが注視する実オブジェクトを含む領域(以下、注目領域)に不要な光があると、没入感又は利便性が損なわれる場合がある、という問題がある。この問題を解決するために、HDM装置越しにユーザが見る視界全体を遮蔽するアタッチメントを用いることが考えられる。しかし、視界全体が常に遮蔽されると、ユーザは周囲の実空間の状況を視覚的に把握できなくなる、という問題がある。ユーザがマウス又はキーボードを用いて指定した注目領域についてのみ実空間からの光を遮蔽して仮想オブジェクトを表示する等の表示制御を行うことも考えられる。しかし、注目領域の指定にマウス又はキーボードを用いたのではユーザの利便性が低下する。 For HMD devices that display virtual objects without blocking the user's field of view, the immersion or convenience may be impaired if there is unnecessary light in the area that includes the real object that the user is gazing at (hereinafter referred to as the attention area). There is a problem. In order to solve this problem, it is conceivable to use an attachment that blocks the entire field of view seen by the user through the HDM device. However, if the entire field of view is always blocked, there is a problem in that the user cannot visually grasp the situation in the surrounding real space. It is also conceivable to perform display control such as blocking light from the real space and displaying a virtual object only in a region of interest specified by the user using a mouse or keyboard. However, using a mouse or keyboard to specify the attention area reduces user convenience.
 本開示の好適な態様に係る、仮想オブジェクトが表示される透過型の表示部を備えた眼鏡型表示装置は、特定部と、表示制御部と、を備える。特定部は、ユーザの視界において前記ユーザが注視する実オブジェクトを含む注目領域を、前記ユーザが発話した音声に基づいて特定する。表示制御部は、前記表示部のうち、前記視界の注目領域と重なる表示領域に、前記実オブジェクトに対応する仮想オブジェクトを表示させる。 According to a preferred aspect of the present disclosure, a glasses-type display device including a transmissive display section on which a virtual object is displayed includes a specifying section and a display control section. The specifying unit specifies an attention area including a real object that the user is gazing at in the user's field of view based on the voice uttered by the user. The display control unit displays a virtual object corresponding to the real object in a display area of the display unit that overlaps the attention area of the field of view.
 本開示の好適な態様に係る表示システムは、ユーザの頭部に装着され、仮想オブジェクトが表示される透過型の表示部を備えた眼鏡型表示装置と、特定部と、表示制御部と、を備える。特定部は、ユーザの視界において前記ユーザが注視する実オブジェクトを含む注目領域を、前記ユーザが発話した音声に基づいて特定する。表示制御部は、前記表示部のうち、前記視界の注目領域と重なる表示領域に、前記実オブジェクトに対応する仮想オブジェクトを、表示させる。 A display system according to a preferred aspect of the present disclosure includes a glasses-type display device that is attached to a user's head and includes a transmissive display section on which a virtual object is displayed, a specific section, and a display control section. Be prepared. The specifying unit specifies an attention area including a real object that the user is gazing at in the user's field of view based on the voice uttered by the user. The display control unit displays a virtual object corresponding to the real object in a display area of the display unit that overlaps the attention area of the field of view.
 本開示によれば、ユーザの音声に基づいて注目領域が特定されるので、ユーザの利便性を損なうことなく、注目領域に重ねて仮想オブジェクトを表示することができる。 According to the present disclosure, since the attention area is specified based on the user's voice, the virtual object can be displayed over the attention area without impairing the user's convenience.
本開示の一実施形態による表示システム1の構成例を示すブロック図である。FIG. 1 is a block diagram illustrating a configuration example of a display system 1 according to an embodiment of the present disclosure. 実オブジェクトが配置されるポスターPの一例を示す図である。It is a diagram showing an example of a poster P on which real objects are arranged. 本開示における管理テーブルTBLの一例を示す図である。It is a figure showing an example of management table TBL in this indication. 携帯機器10の構成例を示すブロック図である。1 is a block diagram showing a configuration example of a mobile device 10. FIG. 携帯機器10の処理装置18がプログラムPR1に従って実行する表示方法の流れを示すフローチャートである。2 is a flowchart showing the flow of a display method executed by the processing device 18 of the mobile device 10 according to the program PR1. 眼鏡型表示装置20の構成例を示すブロック図である。2 is a block diagram showing a configuration example of a glasses-type display device 20. FIG. 本実施形態の動作を説明するための図である。FIG. 3 is a diagram for explaining the operation of this embodiment. 本実施形態の動作を説明するための図である。FIG. 3 is a diagram for explaining the operation of this embodiment. 変形例1においてユーザUが見る画像の一例を示す図である。7 is a diagram illustrating an example of an image viewed by user U in modification example 1. FIG. 変形例2においてユーザUが見る画像の一例を示す図である。7 is a diagram illustrating an example of an image viewed by user U in modification 2. FIG.
A.実施形態
 図1は、本開示の一実施形態による表示システム1の構成例を示すブロック図である。図1に示されるように、表示システム1は、携帯機器10と、眼鏡型表示装置20と、を含む。眼鏡型表示装置20は、ユーザUの頭部に装着される。眼鏡型表示装置20は、眼鏡型表示装置20を装着したユーザUの視界を遮蔽せずに、実空間には存在しない仮想オブジェクトを表示するHMD装置である。眼鏡型表示装置20は、撮像機能を備える。ユーザUの頭部に装着された眼鏡型表示装置20は、ユーザUの視界に対応する実空間の画像を撮像機能により撮像する。ユーザUの頭部に装着された眼鏡型表示装置20により撮像された画像には、ユーザUの視界に存在する物体、即ち実オブジェクトが映る。
A. Embodiment FIG. 1 is a block diagram showing a configuration example of a display system 1 according to an embodiment of the present disclosure. As shown in FIG. 1, the display system 1 includes a mobile device 10 and a glasses-type display device 20. The glasses-type display device 20 is attached to the head of the user U. The glasses-type display device 20 is an HMD device that displays virtual objects that do not exist in real space without blocking the field of view of the user U wearing the glasses-type display device 20. The glasses-type display device 20 has an imaging function. The glasses-type display device 20 mounted on the head of the user U uses an imaging function to capture an image of real space corresponding to the field of view of the user U. An object existing in the field of view of the user U, that is, a real object, appears in the image captured by the glasses-type display device 20 worn on the head of the user U.
 携帯機器10は、例えばスマートフォンである。携帯機器10は、ユーザUの身体に装着される。携帯機器10は、ストラップ等を用いて首からぶら下げることでユーザUの身体に装着される。携帯機器10は、収音機能を備える。ユーザの身体に装着された携帯機器10は、ユーザUの発した音声を収音機能により収音する。また、携帯機器10は、頭部に装着される眼鏡型表示装置20と有線にて接続される。携帯機器10は、無線にて眼鏡型表示装置20と接続されてもよい。携帯機器10は、眼鏡型表示装置20により撮像された画像を表す画像データを眼鏡型表示装置20から取得する。なお、携帯機器10は、スマートフォンに限らず、例えば、タブレット又はノート型パーソナルコンピュータでもよい。 The mobile device 10 is, for example, a smartphone. The mobile device 10 is worn on the user U's body. The mobile device 10 is attached to the body of the user U by hanging from the neck using a strap or the like. The mobile device 10 has a sound collection function. The mobile device 10 worn on the user's body collects the voice emitted by the user U using a sound collection function. Furthermore, the mobile device 10 is connected by wire to a glasses-type display device 20 that is worn on the head. The mobile device 10 may be connected to the eyeglass-type display device 20 wirelessly. The mobile device 10 acquires image data representing an image captured by the glasses-type display device 20 from the glasses-type display device 20 . Note that the mobile device 10 is not limited to a smartphone, and may be, for example, a tablet or a notebook personal computer.
 また、携帯機器10は、通信網NWを介して管理装置30と通信する。携帯機器10は、眼鏡型表示装置20から取得した画像データを管理装置30へ送信する。管理装置30は、ARにおける位置認識サービス及びコンテンツ管理サービスを提供するサーバ装置である。 Furthermore, the mobile device 10 communicates with the management device 30 via the communication network NW. The mobile device 10 transmits the image data acquired from the glasses-type display device 20 to the management device 30. The management device 30 is a server device that provides a location recognition service and a content management service in AR.
 位置認識サービスとは、眼鏡型表示装置20の撮像機能により撮像された画像に基づいて、グローバル座標系における当該眼鏡型表示装置20の位置を特定するサービスである。位置認識サービスの具体的な実現態様は、ARタグを用いる態様、又はSLAM(Simultaneous Localization and Mapping)のように画像から抽出される特徴点の分布を利用する態様が挙げられる。コンテンツ管理サービスとは、グローバル座標系における眼鏡型表示装置20の位置から見える1又は複数の実オブジェクトに対応する仮想オブジェクトに関する情報を、眼鏡型表示装置20に配信するサービスである。 The position recognition service is a service that specifies the position of the glasses-type display device 20 in the global coordinate system based on an image captured by the imaging function of the glasses-type display device 20. Specific implementation modes of the location recognition service include a mode using an AR tag or a mode using a distribution of feature points extracted from an image, such as SLAM (Simultaneous Localization and Mapping). The content management service is a service that distributes information regarding virtual objects corresponding to one or more real objects visible from the position of the glasses-type display device 20 in the global coordinate system to the glasses-type display device 20.
 管理装置30には、グローバル座標系における位置に対応付けて当該位置から見える1又は複数の実オブジェクトの各々に対応する仮想オブジェクトについて、当該仮想オブジェクトの画像を表す仮想オブジェクト情報と、当該仮想オブジェクトを表示する表示領域の位置及び大きさを示す領域情報とが予め記憶されている。管理装置30は、通信網NWを介して携帯機器10から受信した画像データに基づいて特定した眼鏡型表示装置20の位置から、1又は複数の実オブジェクトを特定する。そして、特定した1又は複数の実オブジェクトの各々に対応する仮想オブジェクト情報及び領域情報を携帯機器10へ返信する。携帯機器10は、管理装置30から受信した仮想オブジェクト情報及び領域情報に従って、仮想オブジェクトの画像を眼鏡型表示装置20に表示させる。これにより、ユーザUの眼には、実空間に重ねて仮想オブジェクトが映る。 The management device 30 stores virtual object information representing an image of the virtual object, and information about the virtual object corresponding to each of one or more real objects visible from the position in association with the position in the global coordinate system. Area information indicating the position and size of the display area to be displayed is stored in advance. The management device 30 specifies one or more real objects from the position of the glasses-type display device 20 specified based on the image data received from the mobile device 10 via the communication network NW. Then, virtual object information and area information corresponding to each of the identified one or more real objects are sent back to the mobile device 10. The mobile device 10 causes the glasses-type display device 20 to display an image of the virtual object according to the virtual object information and area information received from the management device 30. As a result, the virtual object appears superimposed on the real space in the eyes of the user U.
 本実施形態における実空間は、例えば学会等の研究発表会におけるポスターセッションの会場である。本実施形態における実オブジェクトは、例えばポスターセッションの会場内に掲げられ、研究内容が英文で記載されたポスターにおける各英単語である。図2は、ポスターセッションの会場内に掲げられるポスターPの一例を示す図である。本実施形態における仮想オブジェクトは、ポスターPに記載された英単語の日本語訳を表す文字列である。本実施形態では、管理装置30は、図3に示される管理テーブルTBLを携帯機器10に送信することにより、仮想オブジェクト情報及び領域情報を携帯機器10へ送信する。 The real space in this embodiment is, for example, a venue for a poster session at a research presentation such as an academic conference. The real objects in this embodiment are, for example, each English word on a poster displayed in a poster session venue and in which research content is written in English. FIG. 2 is a diagram showing an example of posters P displayed in the poster session venue. The virtual object in this embodiment is a character string representing a Japanese translation of an English word written on the poster P. In this embodiment, the management device 30 transmits virtual object information and area information to the mobile device 10 by transmitting the management table TBL shown in FIG. 3 to the mobile device 10.
 図3に示されるように、管理テーブルTBLには、実オブジェクトを識別する識別情報に対応付けて、当該実オブジェクトに対応する仮想オブジェクトについての仮想オブジェクト情報と領域情報とが格納されている。本実施形態における識別情報は、当該識別情報により識別される実オブジェクトの発音を表す文字列のデータである。例えば、英単語「Patients」が実オブジェクトであるとする。この場合の識別情報は、カタカナの「ペイシェンツ」である。カタカナは、日本語において外国語の単語の発音を表記する際に用いられる発音文字(phonetic characters)である。この識別情報に対応する仮想オブジェクト情報は、英単語「Patients」の日本語訳を表す「病人」という日本語文字列の画像を表す。仮想オブジェクト情報の表す画像では、実空間からの光を遮蔽するために日本語文字列の背景が白色等の所定の色で塗り潰されている。また、「ペイシェンツ」という識別情報に対応付けて管理テーブルTBLに格納されている領域情報は、実オブジェクト「Patients」と重なる領域を表す。 As shown in FIG. 3, the management table TBL stores virtual object information and area information about a virtual object corresponding to the real object in association with identification information for identifying the real object. The identification information in this embodiment is character string data representing the pronunciation of the real object identified by the identification information. For example, assume that the English word "Patients" is a real object. The identification information in this case is "patients" in katakana. Katakana are phonetic characters used in Japanese to represent the pronunciation of foreign words. The virtual object information corresponding to this identification information represents an image of the Japanese character string "Kikin" which represents the Japanese translation of the English word "Patients". In the image representing the virtual object information, the background of the Japanese character string is painted over with a predetermined color such as white to block light from real space. Further, the area information stored in the management table TBL in association with the identification information "Patients" represents an area that overlaps with the real object "Patients".
 図4は、携帯機器10の構成例を示すブロック図である。図4に示されるように、携帯機器10は、入力装置11と、出力装置12と、マイクロホン13と、通信装置14と、通信装置15と、記憶装置17と、処理装置18と、バス19と、を含む。入力装置11、出力装置12、マイクロホン13、通信装置14、通信装置15、及び記憶装置17と、処理装置18とは、情報を通信するためのバス19によって相互に接続される。バス19は、単一のバスを用いて構成されてもよいし、装置間ごとに異なるバスを用いて構成されてもよい。 FIG. 4 is a block diagram showing a configuration example of the mobile device 10. As shown in FIG. 4, the mobile device 10 includes an input device 11, an output device 12, a microphone 13, a communication device 14, a communication device 15, a storage device 17, a processing device 18, and a bus 19. ,including. The input device 11, the output device 12, the microphone 13, the communication device 14, the communication device 15, the storage device 17, and the processing device 18 are interconnected by a bus 19 for communicating information. The bus 19 may be configured using a single bus, or may be configured using different buses for each device.
 入力装置11は、タッチパネルを含む。入力装置11は、タッチパネルに加えて、複数の操作キーを含んでもよい。入力装置11は、タッチパネルを含まずに、複数の操作キーを含んでもよい。入力装置11は、ユーザUが行う操作を受け付ける。出力装置12は、ディスプレイを含む。出力装置12のディスプレイの上には入力装置11のタッチパネルが積層されている。出力装置12は、種々の情報を表示する。 The input device 11 includes a touch panel. The input device 11 may include a plurality of operation keys in addition to a touch panel. The input device 11 may include a plurality of operation keys without including a touch panel. The input device 11 receives operations performed by the user U. Output device 12 includes a display. A touch panel of the input device 11 is stacked on the display of the output device 12 . The output device 12 displays various information.
 マイクロホン13は、ユーザUの音声を収音する。マイクロホン13は、収音した音の波形を示す音データを生成し、処理装置18に出力する。詳細については後述するが、本実施形態では、マイクロホン13により収音されるユーザUの音声に基づいてユーザUの注目領域が特定される。 The microphone 13 picks up user U's voice. The microphone 13 generates sound data indicating the waveform of the collected sound and outputs it to the processing device 18 . Although details will be described later, in this embodiment, the user U's attention area is specified based on the user's U voice picked up by the microphone 13.
 通信装置14は、通信網NWを介して管理装置30と通信するためのハードウェア(送受信デバイス)である。通信装置14は、例えば、ネットワークデバイス、ネットワークコントローラ、ネットワークカード、又は通信モジュール等とも呼ばれる。通信装置14は、処理装置18から与えられた画像データを管理装置30へ送信する。また、通信装置14は、管理装置30から受信した管理テーブルTBLを処理装置18へ供給する。なお、通信装置14は、通信網NWを介さずに管理装置30と通信してもよい。 The communication device 14 is hardware (transmission/reception device) for communicating with the management device 30 via the communication network NW. The communication device 14 is also called, for example, a network device, a network controller, a network card, a communication module, or the like. The communication device 14 transmits the image data given from the processing device 18 to the management device 30. Further, the communication device 14 supplies the management table TBL received from the management device 30 to the processing device 18. Note that the communication device 14 may communicate with the management device 30 without going through the communication network NW.
 通信装置15は、有線にて眼鏡型表示装置20と通信するためのハードウェア(送受信デバイス)である。通信装置15は、眼鏡型表示装置20から受信した画像データを処理装置18へ供給する。また、通信装置15は、処理装置18から与えられる画像データを眼鏡型表示装置20に送信する。なお、通信装置15は、無線にて眼鏡型表示装置20と通信してもよい。 The communication device 15 is hardware (transmission/reception device) for communicating with the eyeglass-type display device 20 by wire. The communication device 15 supplies the image data received from the glasses-type display device 20 to the processing device 18 . Furthermore, the communication device 15 transmits image data provided from the processing device 18 to the glasses-type display device 20. Note that the communication device 15 may communicate with the glasses-type display device 20 wirelessly.
 記憶装置17は、処理装置18が読み取り可能な記録媒体である。記憶装置17は、例えば、不揮発性メモリと揮発性メモリとを含む。不揮発性メモリは、例えば、ROM(Read Only Memory)、EPROM(Erasable Programmable Read Only Memory)及びEEPROM(Electrically Erasable Programmable Read Only Memory)である。揮発性メモリは、例えば、RAM(Random Access Memory)である。記憶装置17には、本開示の注目領域の特定方法を処理装置18に実行させるプログラムPR1が予め記憶されている。また、記憶装置17には、管理装置30から受信した管理テーブルTBLが処理装置18によって書き込まれる。 The storage device 17 is a recording medium that can be read by the processing device 18. The storage device 17 includes, for example, nonvolatile memory and volatile memory. Nonvolatile memories include, for example, ROM (Read Only Memory), EPROM (Erasable Programmable Read Only Memory), and EEPROM (Electrically Erasable Programmable Read Only Memory). The volatile memory is, for example, RAM (Random Access Memory). The storage device 17 stores in advance a program PR1 that causes the processing device 18 to execute the method of specifying a region of interest according to the present disclosure. Furthermore, the management table TBL received from the management device 30 is written into the storage device 17 by the processing device 18 .
 処理装置18は、1又は複数のCPU(Central Processing Unit)を含む。1又は複数のCPUは、1又は複数のプロセッサの一例である。プロセッサ及びCPUの各々は、コンピュータの一例である。処理装置18は、記憶装置17からプログラムPR1を読み取る。プログラムPR1に従って作動している処理装置18は、通信装置15を用いて眼鏡型表示装置20から受信した画像データを、通信装置14を用いて管理装置30へ送信する。また、プログラムPR1に従って作動している処理装置18は、通信装置14を用いて管理装置30から受信した管理テーブルTBLを記憶装置17に書き込む。 The processing device 18 includes one or more CPUs (Central Processing Units). One or more CPUs are an example of one or more processors. Each of the processor and CPU is an example of a computer. The processing device 18 reads the program PR1 from the storage device 17. The processing device 18 operating according to the program PR1 transmits the image data received from the glasses-type display device 20 using the communication device 15 to the management device 30 using the communication device 14. Further, the processing device 18 operating according to the program PR1 writes the management table TBL received from the management device 30 into the storage device 17 using the communication device 14.
 また、プログラムPR1に従って作動している処理装置18は、図4に示される音声認識部181、特定部182、及び表示制御部183として機能する。つまり、図4における音声認識部181、特定部182、及び表示制御部183は処理装置18をソフトウェアに従って作動させることにより実現されるソフトウェアモジュールである。 Furthermore, the processing device 18 operating according to the program PR1 functions as the speech recognition section 181, the identification section 182, and the display control section 183 shown in FIG. That is, the speech recognition section 181, the identification section 182, and the display control section 183 in FIG. 4 are software modules realized by operating the processing device 18 according to software.
 音声認識部181は、マイクロホン13が生成した音データの表す音声を文字列に変換する。即ち、音声認識部181は、ユーザUの音声に対して所定の音声認識アルゴリズムに従った音声認識を施す。音声認識アルゴリズムについては既存技術が適宜採用されればよい。音声認識部181は、ユーザUの音声に対する音声認識の結果、即ちユーザが発話した1又は複数の単語の文字列を表す認識文字列データを生成する。 The speech recognition unit 181 converts the speech represented by the sound data generated by the microphone 13 into a character string. That is, the speech recognition unit 181 performs speech recognition on the user U's speech according to a predetermined speech recognition algorithm. Existing technology may be adopted as appropriate for the speech recognition algorithm. The voice recognition unit 181 generates recognized character string data representing the result of voice recognition of the user U's voice, that is, a character string of one or more words uttered by the user.
 特定部182は、音声認識部181による認識結果に基づいて注目領域を特定する。より詳細には、特定部182は、音声認識部181により生成された認識文字列データの表す文字列が、管理テーブルTBLに格納されている何れかの識別情報と一致するか否かを、管理テーブルTBLを参照して判定する。認識文字列データの表す文字列が何れかの識別情報と一致する場合、特定部182は、認識文字列データの表す文字列と一致する識別情報に対応する領域情報の示す領域を注目領域として特定する。 The identification unit 182 identifies the region of interest based on the recognition result by the voice recognition unit 181. More specifically, the identification unit 182 manages whether or not the character string represented by the recognized character string data generated by the speech recognition unit 181 matches any identification information stored in the management table TBL. The determination is made with reference to the table TBL. When the character string represented by the recognized character string data matches any identification information, the identification unit 182 identifies the area represented by the area information corresponding to the identification information that matches the character string represented by the recognized character string data as a region of interest. do.
 表示制御部183は、眼鏡型表示装置20の表示部のうち、特定部182により特定された注目領域と重なる表示領域に、当該注目領域に対応する仮想オブジェクトを表示させる。より詳細には、表示制御部183は、認識文字列データの表す文字列と一致する識別情報に対応する仮想オブジェクト情報の表す画像を、当該識別情報に対応する領域情報の示す領域に配置した画像を表す画像データを生成する。そして、表示制御部183は、当該画像データを、通信装置15を用いて眼鏡型表示装置20に送信することにより、当該画像データの表す画像を眼鏡型表示装置20に表示させる。 The display control unit 183 causes a display area of the display unit of the glasses-type display device 20 that overlaps with the area of interest specified by the specifying unit 182 to display a virtual object corresponding to the area of interest. More specifically, the display control unit 183 displays an image in which an image representing virtual object information corresponding to identification information that matches the character string represented by the recognized character string data is arranged in an area indicated by area information corresponding to the identification information. Generate image data representing. Then, the display control unit 183 transmits the image data to the eyeglass-type display device 20 using the communication device 15, thereby causing the eyeglass-type display device 20 to display an image represented by the image data.
 また、プログラムPR1に従って作動している処理装置18は、マイクロホン13から音データが出力される毎に、図5に示される表示方法を実行する。図5に示されるように、この表示方法は、ステップSA110~ステップSA140の各処理を含む。 Furthermore, the processing device 18 operating according to the program PR1 executes the display method shown in FIG. 5 every time sound data is output from the microphone 13. As shown in FIG. 5, this display method includes each process of step SA110 to step SA140.
 ステップSA110では、処理装置18は、音声認識部181として機能する。ステップSA110では、処理装置18は、マイクロホン13から出力された音データの表す音声に対して音声認識を施すことにより、認識文字列データを生成する。 In step SA110, the processing device 18 functions as the speech recognition unit 181. In step SA110, the processing device 18 generates recognized character string data by performing voice recognition on the voice represented by the sound data output from the microphone 13.
 ステップSA120及びステップSA130では、処理装置18は、特定部182として機能する。ステップSA120では、処理装置18は、ステップSA110にて生成された認識文字列データの表す文字列が、管理テーブルTBLに格納されている何れかの識別情報と一致するか否かを、管理テーブルTBLを参照して判定する。 In step SA120 and step SA130, the processing device 18 functions as the specifying unit 182. In step SA120, the processing device 18 checks whether the character string represented by the recognized character string data generated in step SA110 matches any identification information stored in the management table TBL. Determine by referring to.
 ステップSA120の判定結果が“Yes”である場合、即ち認識文字列データの表す文字列が何れかの識別情報と一致した場合、処理装置18はステップSA130の処理を実行する。ステップSA130では、処理装置18は、ステップSA110にて生成した認識文字列データの表す文字列と一致する識別情報に対応付けて管理テーブルTBLに格納されている領域情報の示す領域を注目領域として特定する。ステップSA120の判定結果が“No”である場合、即ち認識文字列データの表す文字列が何れの識別情報とも一致しなかった場合には、処理装置18はステップSA130以降の処理を実行することなく、本表示方法を終了する。 If the determination result in step SA120 is "Yes", that is, if the character string represented by the recognized character string data matches any identification information, the processing device 18 executes the process in step SA130. In step SA130, the processing device 18 specifies, as a region of interest, the region indicated by the region information stored in the management table TBL in association with the identification information that matches the character string represented by the recognized character string data generated in step SA110. do. If the determination result in step SA120 is "No", that is, if the character string represented by the recognized character string data does not match any identification information, the processing device 18 does not execute the processing from step SA130 onwards. , this display method ends.
 ステップSA130に後続するステップSA140では、処理装置18は、表示制御部183として機能する。ステップSA140では、処理装置18は、ステップSA110にて生成した認識文字列データの表す文字列と一致する識別情報に対応付けて管理テーブルTBLに格納されている仮想オブジェクト情報を取得し、当該仮想オブジェクト情報の表す仮想オブジェクトを、ステップSA130にて特定した注目領域に配置した画像を表す画像データを生成する。そして、処理装置18は、生成した画像データを眼鏡型表示装置20に供給することにより、当該画像データの表す画像を眼鏡型表示装置20に表示させる。 In step SA140 following step SA130, the processing device 18 functions as the display control unit 183. In step SA140, the processing device 18 acquires the virtual object information stored in the management table TBL in association with the identification information that matches the character string represented by the recognized character string data generated in step SA110, and Image data representing an image in which a virtual object representing information is placed in the region of interest specified in step SA130 is generated. Then, the processing device 18 supplies the generated image data to the eyeglass-type display device 20, thereby causing the eyeglass-type display device 20 to display an image represented by the image data.
 図6は、眼鏡型表示装置20の構成例を示すブロック図である。眼鏡型表示装置20は、表示部2aと、通信装置2bと、撮像装置2cと、記憶装置2dと、処理装置2eと、バス2fと、を含む。表示部2a、通信装置2b、撮像装置2c、及び記憶装置2dと、処理装置2eとは、情報を通信するためのバス2fによって相互に接続される。バス2fは、単一のバスを用いて構成されてもよいし、装置等の要素間ごとに異なるバスを用いて構成されてもよい。 FIG. 6 is a block diagram showing a configuration example of the eyeglass-type display device 20. The eyeglass-type display device 20 includes a display section 2a, a communication device 2b, an imaging device 2c, a storage device 2d, a processing device 2e, and a bus 2f. The display unit 2a, the communication device 2b, the imaging device 2c, the storage device 2d, and the processing device 2e are interconnected by a bus 2f for communicating information. The bus 2f may be configured using a single bus, or may be configured using different buses for each element such as a device.
 表示部2aは、光を透過させる透過型の表示部である。実空間を表す光は、表示部2aを透過する。表示部2aは、処理装置2eによる制御の下、仮想オブジェクトの画像を表示する。ユーザUが眼鏡型表示装置20を装着したとき、表示部2aは、ユーザUの左眼及び右眼の前に位置する。眼鏡型表示装置20を装着したユーザUは、表示部2aを透過した光が表す実空間と、表示部2aに表示される仮想オブジェクトの画像と、を視認する。 The display section 2a is a transmissive display section that transmits light. Light representing real space is transmitted through the display section 2a. The display unit 2a displays images of virtual objects under the control of the processing device 2e. When the user U wears the glasses-type display device 20, the display section 2a is located in front of the user's U left and right eyes. The user U wearing the glasses-type display device 20 visually recognizes the real space represented by the light transmitted through the display section 2a and the image of the virtual object displayed on the display section 2a.
 より詳細に説明すると、表示部2aは、左眼用のレンズ、左眼用の表示パネル、左眼用の光学部材、右眼用のレンズ、右眼用の表示パネル、及び右眼用の光学部材を含む。左眼用の表示パネル及び右眼用の表示パネルは、例えば、液晶パネル又は有機EL(Electro Luminescence)パネルである。左眼用の表示パネルは、処理装置eから与えられる画像データの表す画像を表示する。左眼用の光学部材は、左眼用の表示パネルから射出された光を左眼用のレンズに導光する光学部材である。同様に、右眼用の表示パネルは、処理装置2eから与えられる画像データの表す画像を表示する。右眼用の光学部材は、右眼用の表示パネルから射出された光を右眼用のレンズに導光する光学部材である。 To explain in more detail, the display unit 2a includes a left eye lens, a left eye display panel, a left eye optical member, a right eye lens, a right eye display panel, and a right eye optical member. Contains parts. The display panel for the left eye and the display panel for the right eye are, for example, a liquid crystal panel or an organic EL (Electro Luminescence) panel. The display panel for the left eye displays an image represented by image data provided from the processing device e. The left eye optical member is an optical member that guides light emitted from the left eye display panel to the left eye lens. Similarly, the display panel for the right eye displays an image represented by image data provided from the processing device 2e. The right eye optical member is an optical member that guides light emitted from the right eye display panel to the right eye lens.
 左眼用のレンズ及び右眼用のレンズの各々は、ハーフミラーを有する。左眼用のレンズが有するハーフミラーは、実空間を表す光を透過させることによって、実空間を表す光をユーザUの左眼に導く。また、左眼用のレンズが有するハーフミラーは、左眼用の光学部材によって導光された光をユーザUの左眼に反射する。右眼用のレンズが有するハーフミラーは、実空間を表す光を透過させることによって、実空間を表す光をユーザUの右眼に導く。右眼用のレンズが有するハーフミラーは、右眼用の光学部材によって導光された光をユーザUの右眼に反射する。 Each of the left eye lens and the right eye lens has a half mirror. The half mirror included in the left eye lens guides the light representing the real space to the left eye of the user U by transmitting the light representing the real space. Further, the half mirror included in the left eye lens reflects the light guided by the left eye optical member to the user U's left eye. The half mirror included in the right eye lens guides the light representing the real space to the right eye of the user U by transmitting the light representing the real space. The half mirror included in the right eye lens reflects the light guided by the right eye optical member to the user U's right eye.
 通信装置2bは、有線にて携帯機器10と通信するためのハードウェア(送受信デバイス)である。通信装置2bは、無線にて携帯機器10と通信してもよい。 The communication device 2b is hardware (transmission/reception device) for communicating with the mobile device 10 by wire. The communication device 2b may communicate with the mobile device 10 wirelessly.
 眼鏡型表示装置20は、左眼用のレンズ及び右眼用のレンズを支持する眼鏡型のフレームを有し、撮像装置2c(例えば、カメラ)は、当該フレームにおけるブリッジに設けられる。撮像装置2cは、処理装置2eによる制御の下、眼鏡型表示装置20を装着したユーザUが眼鏡型表示装置20越しに見る実空間の画像、即ちユーザUの視界の画像を撮像する。撮像装置2cは、撮像した画像を表す画像データを処理装置2eへ出力する。 The glasses-type display device 20 has a glasses-shaped frame that supports a left eye lens and a right eye lens, and the imaging device 2c (for example, a camera) is provided on a bridge of the frame. The imaging device 2c captures an image of the real space that the user U wearing the glasses-type display device 20 sees through the glasses-type display device 20, that is, an image of the user's U field of view under the control of the processing device 2e. The imaging device 2c outputs image data representing the captured image to the processing device 2e.
 記憶装置2dは、処理装置2eが読み取り可能な記録媒体である。記憶装置2dは、記憶装置17と同様に不揮発性メモリと揮発性メモリとを含む。記憶装置2dは、プログラムPR2を記憶する。処理装置2eは、1又は複数のCPUを含む。処理装置2eは、記憶装置2dからプログラムPR2を読み取る。処理装置2eは、プログラムPR2を実行することによって、動作制御部2e1として機能する。 The storage device 2d is a recording medium that can be read by the processing device 2e. The storage device 2d, like the storage device 17, includes nonvolatile memory and volatile memory. The storage device 2d stores a program PR2. The processing device 2e includes one or more CPUs. The processing device 2e reads the program PR2 from the storage device 2d. The processing device 2e functions as an operation control unit 2e1 by executing the program PR2.
 動作制御部2e1は、眼鏡型表示装置20の動作を制御する。動作制御部2e1は、撮像装置2cから出力された画像データを、通信装置2bを用いて携帯機器10へ送信する。また、動作制御部2e1は、通信装置2bにより携帯機器10から受信した画像データを表示部2aに供給する。表示部2aは、動作制御部2e1から供給される画像データの表す画像を表示する。前述したように、携帯機器10が送信する画像データの表す画像は、ユーザUの視界における注目領域に仮想オブジェクトを配置した画像である。この画像が表示部2aに表示されるので、ユーザUの眼には、ユーザUの視界における注目領域に仮想オブジェクトが重ねられた実空間の画像が映る。 The operation control unit 2e1 controls the operation of the eyeglass-type display device 20. The operation control unit 2e1 transmits the image data output from the imaging device 2c to the mobile device 10 using the communication device 2b. Further, the operation control unit 2e1 supplies image data received from the mobile device 10 via the communication device 2b to the display unit 2a. The display section 2a displays an image represented by the image data supplied from the operation control section 2e1. As described above, the image represented by the image data transmitted by the mobile device 10 is an image in which a virtual object is placed in the attention area in the user's U field of view. Since this image is displayed on the display unit 2a, the user's U's eyes see an image of the real space in which the virtual object is superimposed on the attention area in the user's U's field of view.
 例えば、眼鏡型表示装置20を頭部に装着し、携帯機器10を身体に装着したユーザUが、眼鏡型表示装置20越しに、図2に示されるポスターPを見ており、携帯機器10の記憶装置17には、図3に示される管理テーブルTBLが記憶されていたとする。図7は、ユーザUが眼鏡型表示装置20越しに見る視界Aの一例を示す。図7に示される視界Aを見ているユーザUの注目先の英単語が「Patients」である場合、ユーザUは「ペイシェンツ」と発音する。 For example, a user U wearing the glasses-type display device 20 on his head and the mobile device 10 on his body is looking at a poster P shown in FIG. It is assumed that the storage device 17 stores a management table TBL shown in FIG. FIG. 7 shows an example of the field of view A that the user U sees through the glasses-type display device 20. When the English word to which the user U, who is looking at the field of view A shown in FIG. 7, is paying attention is "Patients", the user U pronounces "Patients".
 ユーザUの音声は携帯機器10のマイクロホン13によって収音され、この音声に対する音声認識部181による認識結果に基づいて、「Patients」に対応する領域が注目領域として特定される。この注目領域に「Patients」に対応する仮想オブジェクト(「Patients」の日本語訳)を配置した画像を表す画像データが携帯機器10から眼鏡型表示装置20へ送信される。この画像データの表す画像が表示部2aに表示される結果、図8に示されるように、ユーザUの眼には、視界Aにおける注目領域に仮想オブジェクトVOBが重ねられた実空間の画像が映る。 The voice of the user U is collected by the microphone 13 of the mobile device 10, and based on the recognition result of the voice by the voice recognition unit 181, the area corresponding to "Patients" is specified as the area of interest. Image data representing an image in which a virtual object corresponding to “Patients” (Japanese translation of “Patients”) is placed in this attention area is transmitted from the mobile device 10 to the glasses-type display device 20. As a result of displaying the image represented by this image data on the display unit 2a, as shown in FIG. 8, the user U's eyes see an image of the real space in which the virtual object VOB is superimposed on the attention area in the field of view A. .
 以上説明したように、本実施形態によれば、ユーザUが注視する実オブジェクトを含む注目領域をユーザUに指定させ、ユーザUにより指定された注目領域に重ねて仮想オブジェクトを表示することができる。 As described above, according to the present embodiment, it is possible to have the user U specify an attention area that includes the real object that the user U is gazing at, and to display a virtual object overlapping the attention area specified by the user U. .
 加えて、本実施形態では、ユーザUは注目領域を音声により指定することができ、注目領域の指定の際にマウス又はキーボードを操作する必要はない。注目領域の指定の際にマウス又はキーボードを操作する必要はないので、利便性が低下することはない。このように本実施形態によれば、ユーザUの利便性を低下させることなく、視界における注目領域をユーザUに指定させ、ユーザUにより指定された注目領域に仮想オブジェクトを表示することが可能になる。 In addition, in this embodiment, the user U can specify the area of interest by voice, and there is no need to operate the mouse or keyboard when specifying the area of interest. Since there is no need to operate the mouse or keyboard when specifying the region of interest, convenience is not reduced. In this way, according to the present embodiment, it is possible to have the user U specify an attention area in the field of view and to display a virtual object in the attention area specified by the user U, without reducing the convenience of the user U. Become.
B:変形
 本開示は、以上に例示した実施形態に限定されない。具体的な変形の態様は以下の通りである。以下の例示から任意に選択された2以上の態様が併合されてもよい。
B-1:変形例1
 注目領域に対するユーザUの注意を逸らさないため、表示制御部183は、表示部2aにおける表示領域のうち注目領域以外の部分に半透明の画像を表示させてもよい。表示部2aにおける表示領域のうち注目領域以外の部分に半透明の画像が表示されると、実空間から注目領域以外の部分を透過してユーザUの眼に入射する光の一部が遮蔽される。この結果、図9に示されるように、眼鏡型表示装置20越しに見るユーザの視界のうち注目領域以外の部分がぼやけ、ユーザUの没入感が向上する。図9では、注目領域以外の部分のぼやけが、斜め線のハッチングで表されている。なお、眼鏡型表示装置20は、左眼用のレンズ及び右眼用のレンズの透過率を部分的に制御可能に構成されてもよい。この場合、透過率の制御により、注目領域以外の部分を透過する光の遮蔽が実現されてもよい。
B: Modification The present disclosure is not limited to the embodiments illustrated above. Specific aspects of the modification are as follows. Two or more aspects arbitrarily selected from the examples below may be combined.
B-1: Modification 1
In order not to divert user U's attention to the attention area, the display control unit 183 may display a translucent image in a portion of the display area of the display unit 2a other than the attention area. When a translucent image is displayed in a part of the display area of the display unit 2a other than the attention area, a part of the light that passes through the part other than the attention area from the real space and enters the eyes of the user U is blocked. Ru. As a result, as shown in FIG. 9, the portion of the user's field of vision seen through the glasses-type display device 20 other than the area of interest becomes blurred, and the user U's sense of immersion is improved. In FIG. 9, blurring in areas other than the region of interest is represented by diagonal hatching. Note that the glasses-type display device 20 may be configured to be able to partially control the transmittance of the left eye lens and the right eye lens. In this case, by controlling the transmittance, blocking of light that passes through a portion other than the region of interest may be realized.
B-2:変形例2
 識別情報は、複数の実オブジェクトの各々に対して一つずつ割り当てられた一連の番号であってもよい。この場合、表示制御部183は、図10に示されるように、複数の実オブジェクトの各々に対応付けて識別情報を眼鏡型表示装置20に表示させてもよい。図10に示される例では、一連の番号(図示の例では丸付き数字)が、複数の実オブジェクトにそれぞれ付与されている。識別情報が一連の番号である場合、ユーザUは注目先の実オブジェクトに付与されている番号を発話することで注目領域を指定する。例えば、注目先の英単語が「Department」である場合、「Department」に付与された番号は「3」であるため、ユーザUは日本語で「サン」又は英語で「スリー」と発話することで、注目領域を指定する。この場合、特定部182は、複数の実オブジェクトの何れかに対応する識別情報(番号)が音声認識部181により認識された場合に、認識された番号に基づいて注目領域を特定する。
B-2: Modification 2
The identification information may be a series of numbers assigned to each of the plurality of real objects. In this case, the display control unit 183 may cause the glasses-type display device 20 to display identification information in association with each of the plurality of real objects, as shown in FIG. 10. In the example shown in FIG. 10, a series of numbers (circled numbers in the illustrated example) are assigned to each of a plurality of real objects. If the identification information is a series of numbers, the user U specifies the area of interest by speaking the number assigned to the real object of interest. For example, if the English word of interest is "Department", the number given to "Department" is "3", so user U can say "san" in Japanese or "three" in English. to specify the area of interest. In this case, when the identification information (number) corresponding to any of the plurality of real objects is recognized by the voice recognition unit 181, the identification unit 182 identifies the region of interest based on the recognized number.
B-3:変形例3
 表示制御部183は、辞書アプリケーション(例えば、英和辞書アプリケーション)の起動をユーザUに指示させるための仮想オブジェクトを眼鏡型表示装置20に表示させてもよい。この場合、処理装置18は、当該仮想オブジェクトを示す音データがマイクロホン13から与えられた場合に当該英和辞書アプリケーションを起動する。また、注目先の英単語を指定する音声は、英単語を発音する音声には限定されず、当該英単語を構成するアルファベットを英単語における配列順に読み上げる音声であってもよい。例えば、注目先の英単語が「World」である場合、ユーザUは「ダブリュ オー アール エル ディ」と発話することで、注目領域を指定する。この場合における識別情報は当該英単語を構成するアルファベットの発音を英単語における配列順に並べた文字列情報であればよい。なお、本開示における実オブジェクトは英単語には限定されず、フランス語、ドイツ語、又は中国語等の他の言語における単語であってもよい。また、本開示における実オブジェクトの配置先はポスターには限定されず、雑誌、本、又は新聞であってもよく、飲食店におけるメニュー表、機器等の取り扱い説明書、各種書類、又は看板等であってもよい。
B-3: Modification 3
The display control unit 183 may cause the glasses-type display device 20 to display a virtual object for instructing the user U to start a dictionary application (for example, an English-Japanese dictionary application). In this case, the processing device 18 starts the English-Japanese dictionary application when sound data indicating the virtual object is provided from the microphone 13. Furthermore, the audio that specifies the English word of interest is not limited to the audio that pronounces the English word, but may be the audio that reads out the alphabets that make up the English word in the order in which they are arranged in the English word. For example, if the English word of interest is "World", user U specifies the area of interest by saying "Double o r l di". In this case, the identification information may be character string information in which the pronunciations of the alphabets constituting the English word are arranged in the order in which they are arranged in the English word. Note that real objects in the present disclosure are not limited to English words, but may be words in other languages such as French, German, or Chinese. In addition, the placement destination of the real object in the present disclosure is not limited to a poster, but may be a magazine, book, or newspaper, or may be a menu board at a restaurant, an instruction manual for equipment, various documents, or a signboard. There may be.
B-4:変形例4
 眼鏡型表示装置20は、ユーザUの視線を検出する検出装置を備えていてもよい。この場合、特定部182は、検出装置により検出された視線に基づいて、ユーザUの視界において注目領域が占める範囲を大まかに特定し、表示制御部183は当該範囲を強調する表示制御を行なってもよい。例えば、表示制御部183は、注目領域が占める範囲を拡大した画像を表示部2aに表示させてもよいし、あるいは、当該範囲を囲む枠線の画像を表示部2aに表示させてもよい。
 また、眼鏡型表示装置20は、図7における視界Aの一部の領域をユーザUが指先でタッチする操作により、注目領域をユーザUが指定可能に構成されていてもよい。この場合、表示制御部183は、ユーザUによる領域の指定を撮像装置2cによる撮像画像に基づいて検出し、検出した領域を翻訳対象の候補として強調表示してもよい。
B-4: Modification 4
The glasses-type display device 20 may include a detection device that detects the user's U line of sight. In this case, the specifying unit 182 roughly specifies the range occupied by the attention area in the field of view of the user U based on the line of sight detected by the detection device, and the display control unit 183 performs display control to emphasize the range. Good too. For example, the display control unit 183 may cause the display unit 2a to display an image in which the range occupied by the region of interest is expanded, or may cause the display unit 2a to display an image of a frame line surrounding the range.
Further, the eyeglass-type display device 20 may be configured to allow the user U to specify an area of interest by touching a part of the field of view A in FIG. 7 with a fingertip of the user U. In this case, the display control unit 183 may detect the area designation by the user U based on the image captured by the imaging device 2c, and highlight the detected area as a candidate for translation.
B-5:変形例5
 上記実施形態では、携帯機器10の記憶装置17にプログラムPR1が記憶されていたが、プログラムPR1が単体で製造又は販売されてもよい。プログラムPR1を販売する場合、プログラムPR1の提供者は、フラッシュROM等のコンピュータ読み取り可能な記録媒体にプログラムPR1を書き込んで配布してもよいし、又は電気通信回線経由のダウンロードにより配布してもよい。
B-5: Modification 5
In the above embodiment, the program PR1 is stored in the storage device 17 of the mobile device 10, but the program PR1 may be manufactured or sold separately. When selling the program PR1, the provider of the program PR1 may write the program PR1 on a computer-readable recording medium such as a flash ROM and distribute it, or may distribute it by downloading it via a telecommunications line. .
B-6:変形例6
 上記実施形態における音声認識部181、特定部182、及び表示制御部183はソフトウェアモジュールであった。しかし、音声認識部181、特定部182、及び表示制御部183のうちの何れか一つ、複数、又は全部はハードウェアモジュールであってもよい。ハードウェアモジュールの具体例としては、DSP(Digital Signal Processor)、ASIC(Application Specific Integrated Circuit)、PLD(Programmable Logic Device)、FPGA(Field Programmable Gate Array)等が挙げられる。
B-6: Modification 6
The speech recognition unit 181, identification unit 182, and display control unit 183 in the above embodiment were software modules. However, any one, a plurality, or all of the voice recognition unit 181, the identification unit 182, and the display control unit 183 may be a hardware module. Specific examples of the hardware module include DSP (Digital Signal Processor), ASIC (Application Specific Integrated Circuit), PLD (Programmable Logic Device), FPGA (Field Programmable Gate Array), and the like.
B-7:変形例7
 上記実施形態では、携帯機器10が音声認識部181、特定部182、及び表示制御部183を有したが、眼鏡型表示装置20、又は管理装置30が音声認識部181、特定部182、及び表示制御部183を有してもよい。音声認識部181、特定部182、及び表示制御部183が、眼鏡型表示装置20、携帯機器10、及び管理装置30のうちの何れか2つ、又は全部に分散して設けられてもよい。なお、音声認識部181による音声認識処理は処理負荷の高い処理であるため、音声認識部181は眼鏡型表示装置20よりも管理装置30又は携帯機器10に設けられることが好ましい。更に、音声認識部181は、携帯機器10よりも管理装置30に設けられることが好ましい。
B-7: Modification 7
In the above embodiment, the mobile device 10 has the voice recognition section 181, the identification section 182, and the display control section 183, but the glasses-type display device 20 or the management device 30 has the voice recognition section 181, the identification section 182, and the display control section 183. It may also include a control section 183. The voice recognition unit 181, the identification unit 182, and the display control unit 183 may be distributed and provided in any two or all of the glasses-type display device 20, the mobile device 10, and the management device 30. Note that since the voice recognition process by the voice recognition unit 181 is a process with a high processing load, it is preferable that the voice recognition unit 181 be provided in the management device 30 or the mobile device 10 rather than the glasses-type display device 20. Furthermore, the voice recognition unit 181 is preferably provided in the management device 30 rather than the mobile device 10.
C:その他
(1)上述した実施形態では、記憶装置17及び記憶装置2dとしてROM及びRAM等が例示されたが、記憶装置17及び記憶装置2dは、フレキシブルディスク、光磁気ディスク(例えば、コンパクトディスク、デジタル多用途ディスク、Blu-ray(登録商標)ディスク)、スマートカード、フラッシュメモリデバイス(例えば、カード、スティック、キードライブ)、CD-ROM(Compact Disc-ROM)、レジスタ、リムーバブルディスク、ハードディスク、フロッピー(登録商標)ディスク、磁気ストリップ、データベース、サーバその他の適切な記憶媒体であってもよい。
C: Others (1) In the embodiment described above, the storage device 17 and the storage device 2d are exemplified as ROM, RAM, etc., but the storage device 17 and the storage device 2d are flexible disks, magneto-optical disks (for example, compact disks). , digital versatile discs, Blu-ray discs), smart cards, flash memory devices (e.g. cards, sticks, key drives), CD-ROMs (Compact Disc-ROMs), registers, removable disks, hard disks, It may be a floppy disk, magnetic strip, database, server or other suitable storage medium.
(2)上述した実施形態において、説明した情報、信号等は、様々な異なる技術の何れかを使用して表されてもよい。例えば、上記の説明全体に渡って言及され得るデータ、命令、コマンド、情報、信号、ビット、シンボル、チップ等は、電圧、電流、電磁波、磁界若しくは磁性粒子、光場若しくは光子、又はこれらの任意の組み合わせによって表されてもよい。 (2) In the embodiments described above, the information, signals, etc. described may be represented using any of a variety of different technologies. For example, data, instructions, commands, information, signals, bits, symbols, chips, etc. that may be referred to throughout the above description may refer to voltages, currents, electromagnetic waves, magnetic fields or magnetic particles, light fields or photons, or any of these. It may also be represented by a combination of
(3)上述した実施形態において、入出力された情報等は特定の場所(例えば、メモリ)に保存されてもよいし、管理テーブルを用いて管理してもよい。入出力される情報等は、上書き、更新、又は追記され得る。出力された情報等は削除されてもよい。入力された情報等は他の装置へ送信されてもよい。 (3) In the embodiments described above, the input/output information may be stored in a specific location (for example, memory) or may be managed using a management table. Information etc. to be input/output may be overwritten, updated, or additionally written. The output information etc. may be deleted. The input information etc. may be transmitted to other devices.
(4)上述した実施形態において、判定は、1ビットで表される値(0か1か)によって行われてもよいし、真偽値(Boolean:true又はfalse)によって行われてもよいし、数値の比較(例えば、所定の値との比較)によって行われてもよい。 (4) In the embodiments described above, the determination may be made based on a value represented by 1 bit (0 or 1), or may be made based on a truth value (Boolean: true or false). , may be performed by numerical comparison (for example, comparison with a predetermined value).
(5)上述した実施形態において例示した処理手順、シーケンス、フローチャート等は、矛盾の無い限り、順序が入れ替えられてもよい。例えば、本開示において説明した方法については、例示的な順序を用いて様々なステップの要素が提示されており、提示された特定の順序に限定されない。 (5) The order of the processing procedures, sequences, flowcharts, etc. illustrated in the embodiments described above may be changed as long as there is no contradiction. For example, for the methods described in this disclosure, elements of the various steps are presented using an example order and are not limited to the particular order presented.
(6)図4に例示された各機能は、ハードウェア及びソフトウェアの少なくとも一方の任意の組み合わせによって実現される。また、各機能ブロックの実現方法は特に限定されない。すなわち、各機能ブロックは、物理的又は論理的に結合した1つの装置を用いて実現されてもよいし、物理的又は論理的に分離した2つ以上の装置を直接的又は間接的に(例えば、有線、無線等を用いて)接続し、これら複数の装置を用いて実現されてもよい。機能ブロックは、上記1つの装置又は上記複数の装置にソフトウェアを組み合わせて実現されてもよい。 (6) Each function illustrated in FIG. 4 is realized by an arbitrary combination of at least one of hardware and software. Furthermore, the method for realizing each functional block is not particularly limited. That is, each functional block may be realized using one physically or logically coupled device, or may be realized using two or more physically or logically separated devices directly or indirectly (e.g. , wired, wireless, etc.) and may be realized using a plurality of these devices. The functional block may be realized by combining software with the one device or the plurality of devices.
(7)上述した実施形態で例示したプログラムは、ソフトウェアは、ソフトウェア、ファームウェア、ミドルウェア、マイクロコード、ハードウェア記述言語と呼ばれるか、他の名称で呼ばれるかを問わず、命令、命令セット、コード、コードセグメント、プログラムコード、プログラム、サブプログラム、ソフトウェアモジュール、アプリケーション、ソフトウェアアプリケーション、ソフトウェアパッケージ、ルーチン、サブルーチン、オブジェクト、実行可能ファイル、実行スレッド、手順、機能等を意味するよう広く解釈されるべきである。 (7) The programs exemplified in the embodiments described above may include instructions, instruction sets, codes, software, firmware, middleware, microcode, hardware description language, or other names. Should be broadly construed to mean a code segment, program code, program, subprogram, software module, application, software application, software package, routine, subroutine, object, executable, thread of execution, procedure, function, etc. .
 また、ソフトウェア、命令、情報等は、伝送媒体を介して送受信されてもよい。例えば、ソフトウェアが、有線技術(同軸ケーブル、光ファイバケーブル、ツイストペア、デジタル加入者回線(DSL:Digital Subscriber Line)等)及び無線技術(赤外線、マイクロ波等)の少なくとも一方を使用してウェブサイト、サーバ、又は他のリモートソースから送信される場合、これらの有線技術及び無線技術の少なくとも一方は、伝送媒体の定義内に含まれる。 Additionally, software, instructions, information, etc. may be sent and received via a transmission medium. For example, if the software uses wired technology (coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), etc.) and/or wireless technology (infrared, microwave, etc.) to create a website, When transmitted from a server or other remote source, these wired and/or wireless technologies are included within the definition of transmission medium.
(8)前述の各形態において、「システム」及び「ネットワーク」という用語は、互換的に使用される。 (8) In each of the above embodiments, the terms "system" and "network" are used interchangeably.
(9)本開示において説明した情報、パラメータ等は、絶対値を用いて表されてもよいし、所定の値からの相対値を用いて表されてもよいし、対応する別の情報を用いて表されてもよい。 (9) The information, parameters, etc. described in this disclosure may be expressed using absolute values, relative values from a predetermined value, or other corresponding information. It may also be expressed as
(10)上述した実施形態において、携帯機器には、移動局(MS:Mobile Station)である場合が含まれる。移動局は、当業者によって、加入者局、モバイルユニット、加入者ユニット、ワイヤレスユニット、リモートユニット、モバイルデバイス、ワイヤレスデバイス、ワイヤレス通信デバイス、リモートデバイス、モバイル加入者局、アクセス端末、モバイル端末、ワイヤレス端末、リモート端末、ハンドセット、ユーザエージェント、モバイルクライアント、クライアント、又はいくつかの他の適切な用語で呼ばれる場合もある。また、本開示においては、「移動局」、「ユーザ端末(user terminal)」、「ユーザ装置(UE:User Equipment)」、「端末」等の用語は、互換的に使用され得る。 (10) In the embodiments described above, the mobile device includes a mobile station (MS). A mobile station is defined by a person skilled in the art as a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless It may also be referred to as a terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable terminology. Further, in the present disclosure, terms such as "mobile station," "user terminal," "user equipment (UE)," and "terminal" may be used interchangeably.
(11)上述した実施形態において、「接続された(connected)」、「結合された(coupled)」という用語、又はこれらのあらゆる変形は、2又はそれ以上の要素間の直接的又は間接的なあらゆる接続又は結合を意味し、互いに「接続」又は「結合」された2つの要素間に1又はそれ以上の中間要素が存在することを含むことができる。要素間の結合又は接続は、物理的なものであっても、論理的なものであっても、或いはこれらの組み合わせであってもよい。例えば、「接続」は、「アクセス」で読み替えられてもよい。本開示で使用する場合、2つの要素は、1又はそれ以上の電線、ケーブル及びプリント電気接続の少なくとも一つを用いて、並びにいくつかの非限定的かつ非包括的な例として、無線周波数領域、マイクロ波領域及び光(可視及び不可視の両方)領域の波長を有する電磁エネルギー等を用いて、互いに「接続」又は「結合」されると考えることができる。 (11) In the embodiments described above, the terms "connected", "coupled", or any variations thereof refer to direct or indirect connections between two or more elements. Refers to any connection or combination and may include the presence of one or more intermediate elements between two elements that are "connected" or "coupled" to each other. The bonds or connections between elements may be physical, logical, or a combination thereof. For example, "connection" may be replaced with "access." As used in this disclosure, two elements may include one or more electrical wires, cables, and/or printed electrical connections, as well as in the radio frequency domain, as some non-limiting and non-inclusive examples. , electromagnetic energy having wavelengths in the microwave and optical (both visible and non-visible) ranges, and the like.
(12)上述した実施形態において、「に基づいて」という記載は、別段に明記されていない限り、「のみに基づいて」を意味しない。言い換えれば、「に基づいて」という記載は、「のみに基づいて」と「に少なくとも基づいて」の両方を意味する。 (12) In the embodiments described above, the statement "based on" does not mean "based solely on" unless specified otherwise. In other words, the phrase "based on" means both "based only on" and "based at least on."
(13)本開示で使用する「判断(determining)」、「決定(determining)」という用語は、多種多様な動作を包含する場合がある。「判断」、「決定」は、例えば、判定(judging)、計算(calculating)、算出(computing)、処理(processing)、導出(deriving)、調査(investigating)、探索(looking up、search、inquiry)(例えば、テーブル、データベース又は別のデータ構造での探索)、確認(ascertaining)した事を「判断」「決定」したとみなす事等を含み得る。また、「判断」、「決定」は、受信(receiving)(例えば、情報を受信すること)、送信(transmitting)(例えば、情報を送信すること)、入力(input)、出力(output)、アクセス(accessing)(例えば、メモリ中のデータにアクセスすること)した事を「判断」「決定」したとみなす事等を含み得る。また、「判断」、「決定」は、解決(resolving)、選択(selecting)、選定(choosing)、確立(establishing)、比較(comparing)等した事を「判断」「決定」したとみなす事を含み得る。つまり、「判断」「決定」は、何らかの動作を「判断」「決定」したとみなす事を含み得る。また、「判断(決定)」は、「想定する(assuming)」、「期待する(expecting)」、「みなす(considering)」等で読み替えられてもよい。 (13) The terms "determining" and "determining" used in this disclosure may encompass a wide variety of operations. "Judgment" and "decision" include, for example, judging, calculating, computing, processing, deriving, investigating, looking up, search, and inquiry. (e.g., searching in a table, database, or other data structure), and regarding an ascertaining as a "judgment" or "decision." In addition, "judgment" and "decision" refer to receiving (e.g., receiving information), transmitting (e.g., sending information), input, output, and access. (accessing) (e.g., accessing data in memory) may include considering something as a "judgment" or "decision." In addition, "judgment" and "decision" refer to resolving, selecting, choosing, establishing, comparing, etc. as "judgment" and "decision". may be included. In other words, "judgment" and "decision" may include regarding some action as having been "judged" or "determined." Further, "judgment (decision)" may be read as "assuming", "expecting", "considering", etc.
(14)上述した実施形態において、「含む(include)」、「含んでいる(including)」及びそれらの変形が使用されている場合、これらの用語は、用語「備える(comprising)」と同様に、包括的であることが意図される。更に、本開示において使用されている用語「又は(or)」は、排他的論理和ではないことが意図される。 (14) In the embodiments described above, when “include”, “including” and variations thereof are used, these terms are used in the same manner as the term “comprising”. , is intended to be comprehensive. Furthermore, the term "or" as used in this disclosure is not intended to be exclusive or.
(15)本開示において、例えば、英語でのa, an及びtheのように、翻訳により冠詞が追加された場合、本開示は、これらの冠詞の後に続く名詞が複数形であることを含んでもよい。 (15) In the present disclosure, when articles are added by translation, such as a, an, and the in English, the present disclosure does not include that the nouns following these articles are plural. good.
(16)本開示において、「AとBが異なる」という用語は、「AとBが互いに異なる」ことを意味してもよい。なお、当該用語は、「AとBがそれぞれCと異なる」ことを意味してもよい。「離れる」、「結合される」等の用語も、「異なる」と同様に解釈されてもよい。 (16) In the present disclosure, the term "A and B are different" may mean "A and B are different from each other." Note that the term may also mean that "A and B are each different from C". Terms such as "separate", "coupled", etc. may also be interpreted similarly to "different".
(17)本開示において説明した各態様/実施形態は単独で用いてもよいし、組み合わせて用いてもよいし、実行に伴って切り替えて用いてもよい。また、所定の情報の通知(例えば、「Xであること」の通知)は、明示的に行うものに限られず、暗黙的(例えば、当該所定の情報の通知を行わない)ことによって行われてもよい。 (17) Each aspect/embodiment described in the present disclosure may be used alone, in combination, or may be switched and used in accordance with execution. In addition, notification of prescribed information (for example, notification of "X") is not limited to being done explicitly, but may also be done implicitly (for example, not notifying the prescribed information). Good too.
D:上述の形態又は変形例から把握される態様
 以上、本開示について詳細に説明したが、当業者にとっては、本開示が本開示中に説明した実施形態に限定されるものではないということは明らかである。本開示は、請求の範囲の記載により定まる本開示の趣旨及び範囲を逸脱することなく修正及び変更態様として実施することができる。従って、本開示の記載は、例示説明を目的とするものであり、本開示に対して何ら制限的な意味を有するものではない。上述の実施形態又は変形例の少なくとも1つから以下の態様が把握される。
D: Aspects understood from the above embodiments or modified examples Although the present disclosure has been described in detail above, it is clear to those skilled in the art that the present disclosure is not limited to the embodiments described in the present disclosure. it is obvious. The present disclosure can be implemented as modifications and variations without departing from the spirit and scope of the present disclosure as determined by the claims. Therefore, the description of the present disclosure is for the purpose of illustrative explanation and is not intended to have any limiting meaning on the present disclosure. The following aspects can be understood from at least one of the above-described embodiments or modifications.
 第1の態様による、仮想オブジェクトが表示される透過型の表示部2aを備えた眼鏡型表示装置は、特定部182と、表示制御部183と、を備えてもよい。特定部182は、ユーザUの視界においてユーザUが注視する実オブジェクトを含む注目領域を、ユーザUが発話した音声に基づいて特定する。表示制御部183は、ユーザUの視界における注目領域と重なる表示部2aの表示領域に、実オブジェクトに対応する仮想オブジェクトを、表示させる。第1の態様による眼鏡型表示装置によれば、ユーザUは、注視する実オブジェクトを音声により指定できるので、ユーザUの利便性を損なうことなく、指定された注目領域に仮想オブジェクトを表示することができる。 The glasses-type display device according to the first aspect including the transmissive display section 2a on which a virtual object is displayed may include a specifying section 182 and a display control section 183. The specifying unit 182 specifies a region of interest that includes a real object that the user U is gazing at in the user's field of view based on the voice uttered by the user U. The display control unit 183 causes a virtual object corresponding to the real object to be displayed in a display area of the display unit 2a that overlaps with the attention area in the user's U field of view. According to the glasses-type display device according to the first aspect, since the user U can specify the real object to be gazed at by voice, the virtual object can be displayed in the specified attention area without impairing the user U's convenience. I can do it.
 第2の態様(第1の態様の例)におけるユーザUの視界には、複数の実オブジェクトが含まれてもよい。また、第2の態様では、ユーザUが注視する一の実オブジェクトは、ユーザUが発話した音声によって、複数の実オブジェクトから指定されてもよい。第2の態様による眼鏡型表示装置は、ユーザUが発話した音声を認識する音声認識部181を更に備えてもよい。第2の態様による眼鏡型表示装置における特定部182は、音声認識部181の認識結果に基づいて注目領域を特定してもよい。第2の態様による眼鏡型表示装置は、ユーザUの音声に対する音声認識部181による認識結果に基づいて注目領域を特定できる。 The field of view of the user U in the second aspect (an example of the first aspect) may include a plurality of real objects. Furthermore, in the second aspect, one real object that the user U is gazing at may be specified from a plurality of real objects by the voice uttered by the user U. The glasses-type display device according to the second aspect may further include a voice recognition unit 181 that recognizes the voice uttered by the user U. The identification unit 182 in the glasses-type display device according to the second aspect may identify the attention area based on the recognition result of the voice recognition unit 181. The glasses-type display device according to the second aspect can specify the attention area based on the recognition result of the voice recognition unit 181 for the user's U voice.
 第3の態様(第2の態様の例)による眼鏡型表示装置における表示制御部183は、複数の実オブジェクトの各々を一意に識別する識別情報を、複数の実オブジェクトの各々に対応付けて表示部2aの表示領域に表示させてもよい。また、第3の態様(第2の態様の例)による眼鏡型表示装置における特定部182は、複数の実オブジェクトの何れかに対応する識別情報が音声認識部181により認識された場合に、音声認識部181により認識された識別情報に基づいて注目領域を特定してもよい。第3の態様による眼鏡型表示装置は、複数の実オブジェクトの各々に対応付けて表示部2aの表示領域に表示させた複数の識別情報の何れかを示すユーザUの音声に基づいて注目領域を特定できる。 The display control unit 183 in the glasses-type display device according to the third aspect (an example of the second aspect) displays identification information that uniquely identifies each of the plurality of real objects in association with each of the plurality of real objects. It may be displayed in the display area of section 2a. Further, in the glasses-type display device according to the third aspect (an example of the second aspect), the identification unit 182 detects the voice when the voice recognition unit 181 recognizes the identification information corresponding to any one of the plurality of real objects. The region of interest may be specified based on the identification information recognized by the recognition unit 181. The glasses-type display device according to the third aspect determines the attention area based on the voice of the user U indicating any of a plurality of pieces of identification information displayed in the display area of the display unit 2a in association with each of a plurality of real objects. Can be identified.
 第4の態様による表示システムは、ユーザUの頭部に装着され、仮想オブジェクトが表示される透過型の表示部2aを備えた眼鏡型表示装置と、特定部182と、表示制御部183と、を含む。特定部182は、ユーザUの視界においてユーザUが注視する実オブジェクトを含む注目領域を、ユーザUが発話した音声に基づいて特定する。表示制御部183は、ユーザUの視界における注目領域と重なる表示部2aの表示領域に、実オブジェクトに対応する仮想オブジェクトを、表示させる。第4の態様による表示システムによれば、ユーザUは、注視する実オブジェクトを音声により指定できるので、ユーザUの利便性を損なうことなく、指定された注目領域に仮想オブジェクトを表示することができる。 The display system according to the fourth aspect includes a glasses-type display device that is attached to the head of the user U and includes a transmissive display section 2a on which a virtual object is displayed, an identification section 182, a display control section 183, including. The specifying unit 182 specifies a region of interest that includes a real object that the user U is gazing at in the user's field of view based on the voice uttered by the user U. The display control unit 183 causes a virtual object corresponding to the real object to be displayed in a display area of the display unit 2a that overlaps with the attention area in the user's U field of view. According to the display system according to the fourth aspect, since the user U can specify the real object to be gazed at by voice, the virtual object can be displayed in the specified attention area without impairing the user U's convenience. .
1…表示システム、10…携帯機器、20…眼鏡型表示装置、11…入力装置、12…出力装置、13…マイクロホン、14,15,2b…通信装置、17、2d…記憶装置、18,2e…処理装置、181…音声認識部、182…特定部、183…表示制御部、19,2f…バス、2a…表示部、2e1…動作制御部、PR1,PR2…プログラム。 DESCRIPTION OF SYMBOLS 1...Display system, 10...Mobile device, 20...Glasses type display device, 11...Input device, 12...Output device, 13...Microphone, 14, 15, 2b...Communication device, 17, 2d...Storage device, 18, 2e . . . Processing device, 181 .

Claims (4)

  1.  仮想オブジェクトが表示される透過型の表示部を備えた眼鏡型表示装置であって、
     ユーザの視界において前記ユーザが注視する実オブジェクトを含む注目領域を、前記ユーザが発話した音声に基づいて特定する特定部と、
     前記表示部のうち、前記視界の注目領域と重なる表示領域に、前記実オブジェクトに対応する仮想オブジェクトを、表示させる表示制御部と、
     を備える眼鏡型表示装置。
    A glasses-type display device including a transmissive display section on which a virtual object is displayed,
    a specifying unit that specifies an attention area including a real object that the user is gazing at in the user's field of vision based on a voice uttered by the user;
    a display control unit that displays a virtual object corresponding to the real object in a display area of the display unit that overlaps with the attention area of the field of view;
    A glasses-type display device comprising:
  2.  前記視界には、複数の実オブジェクトが含まれ、
     前記ユーザが注視する実オブジェクトは、前記ユーザが発話した音声によって前記複数の実オブジェクトから指定され、
     当該眼鏡型表示装置は、
     前記ユーザが発話した音声を認識する音声認識部を更に備え、
     前記特定部は、前記音声認識部の認識結果に基づいて、前記注目領域を特定する、
     ことを特徴とする請求項1に記載の眼鏡型表示装置。
    The field of view includes a plurality of real objects,
    The real object that the user focuses on is specified from the plurality of real objects by a voice uttered by the user,
    The glasses-type display device is
    further comprising a voice recognition unit that recognizes the voice uttered by the user,
    The identification unit identifies the region of interest based on the recognition result of the voice recognition unit.
    The glasses-type display device according to claim 1, characterized in that:
  3.  前記表示制御部は、前記複数の実オブジェクトの各々を一意に識別する識別情報を、前記複数の実オブジェクトの各々に対応付けて前記表示領域に表示させ、
     前記特定部は、前記複数の実オブジェクトの何れかに対応する識別情報が前記音声認識部により認識された場合に、前記音声認識部により認識された識別情報に基づいて前記注目領域を特定する、
     ことを特徴とする請求項2に記載の眼鏡型表示装置。
    The display control unit causes identification information that uniquely identifies each of the plurality of real objects to be displayed in the display area in association with each of the plurality of real objects,
    The specifying unit specifies the region of interest based on the identification information recognized by the speech recognition unit when identification information corresponding to any of the plurality of real objects is recognized by the speech recognition unit.
    The glasses-type display device according to claim 2, characterized in that:
  4.  ユーザの頭部に装着され、仮想オブジェクトが表示される透過型の表示部を備えた眼鏡型表示装置と、
     前記ユーザの視界において前記ユーザが注視する実オブジェクトを含む注目領域を、前記ユーザが発話した音声に基づいて特定する特定部と、
     前記表示部のうち、前記視界の注目領域と重なる表示領域に、前記実オブジェクトに対応する仮想オブジェクトを、表示させる表示制御部と、
     を備える表示システム。
    a glasses-type display device that is attached to a user's head and includes a transparent display section on which a virtual object is displayed;
    a specifying unit that specifies an attention area including a real object that the user is gazing at in the user's field of vision based on a voice uttered by the user;
    a display control unit that displays a virtual object corresponding to the real object in a display area of the display unit that overlaps with the attention area of the field of view;
    A display system equipped with.
PCT/JP2023/007812 2022-04-28 2023-03-02 Eyeglass-type display device and display system WO2023210158A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-074563 2022-04-28
JP2022074563 2022-04-28

Publications (1)

Publication Number Publication Date
WO2023210158A1 true WO2023210158A1 (en) 2023-11-02

Family

ID=88518461

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/007812 WO2023210158A1 (en) 2022-04-28 2023-03-02 Eyeglass-type display device and display system

Country Status (1)

Country Link
WO (1) WO2023210158A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000194532A (en) * 1998-12-24 2000-07-14 Casio Comput Co Ltd Object selection processor and storage medium
JP2014093050A (en) * 2012-11-06 2014-05-19 Sony Corp Image display device, image display method, and computer program
JP2016218868A (en) * 2015-05-22 2016-12-22 富士通株式会社 Display control method, information processor, and display control program
JP2017091433A (en) * 2015-11-17 2017-05-25 セイコーエプソン株式会社 Head-mounted type display device, method of controlling head-mounted type display device, and computer program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000194532A (en) * 1998-12-24 2000-07-14 Casio Comput Co Ltd Object selection processor and storage medium
JP2014093050A (en) * 2012-11-06 2014-05-19 Sony Corp Image display device, image display method, and computer program
JP2016218868A (en) * 2015-05-22 2016-12-22 富士通株式会社 Display control method, information processor, and display control program
JP2017091433A (en) * 2015-11-17 2017-05-25 セイコーエプソン株式会社 Head-mounted type display device, method of controlling head-mounted type display device, and computer program

Similar Documents

Publication Publication Date Title
US10319382B2 (en) Multi-level voice menu
US11462213B2 (en) Information processing apparatus, information processing method, and program
US20190158927A1 (en) Smart closed caption positioning system for video content
JP6392374B2 (en) Head mounted display system and method for operating head mounted display device
EP3193328A1 (en) Method and device for performing voice recognition using grammar model
EP2940556A1 (en) Command displaying method and command displaying device
EP3479588A1 (en) Augmented reality device and operation thereof
US9336779B1 (en) Dynamic image-based voice entry of unlock sequence
KR20150058286A (en) Leveraging head mounted displays to enable person-to-person interactions
KR102193029B1 (en) Display apparatus and method for performing videotelephony using the same
US20180107651A1 (en) Unsupported character code detection mechanism
US20150339855A1 (en) Laser pointer selection for augmented reality devices
US20160277707A1 (en) Message transmission system, message transmission method, and program for wearable terminal
US10761694B2 (en) Extended reality content exclusion
JPWO2013077110A1 (en) Translation apparatus, translation system, translation method and program
US20190155617A1 (en) Automated setting customization using real-time user data
US11120219B2 (en) User-customized computer-automated translation
CN106228191A (en) A kind of control text overlength detection device and method
WO2023210158A1 (en) Eyeglass-type display device and display system
JP6869809B2 (en) Image estimator
US20230048330A1 (en) In-Vehicle Speech Interaction Method and Device
WO2020075358A1 (en) Information processing device, information processing method, and program
CN110991431A (en) Face recognition method, device, equipment and storage medium
CN107241548B (en) Cursor control method, cursor control device, terminal and storage medium
US20220108624A1 (en) Reader assistance method and system for comprehension checks

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23795906

Country of ref document: EP

Kind code of ref document: A1