WO2015125626A1 - 表示制御装置、表示制御方法およびコンピュータプログラム - Google Patents
表示制御装置、表示制御方法およびコンピュータプログラム Download PDFInfo
- Publication number
- WO2015125626A1 WO2015125626A1 PCT/JP2015/053312 JP2015053312W WO2015125626A1 WO 2015125626 A1 WO2015125626 A1 WO 2015125626A1 JP 2015053312 W JP2015053312 W JP 2015053312W WO 2015125626 A1 WO2015125626 A1 WO 2015125626A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- display
- display control
- unit
- information
- transmissive
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 29
- 238000004590 computer program Methods 0.000 title claims description 9
- 238000002834 transmittance Methods 0.000 claims description 30
- 210000003128 head Anatomy 0.000 claims description 16
- 230000000694 effects Effects 0.000 claims description 14
- 230000004886 head movement Effects 0.000 claims description 4
- 238000012545 processing Methods 0.000 description 78
- 238000003384 imaging method Methods 0.000 description 69
- 238000010586 diagram Methods 0.000 description 25
- 230000008859 change Effects 0.000 description 21
- 230000008569 process Effects 0.000 description 19
- 238000004891 communication Methods 0.000 description 16
- 230000005236 sound signal Effects 0.000 description 14
- 210000000695 crystalline len Anatomy 0.000 description 10
- 230000006870 function Effects 0.000 description 8
- 230000035945 sensitivity Effects 0.000 description 8
- 238000001514 detection method Methods 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 4
- 238000012937 correction Methods 0.000 description 4
- 230000000007 visual effect Effects 0.000 description 4
- 230000006399 behavior Effects 0.000 description 3
- 230000004424 eye movement Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 230000004397 blinking Effects 0.000 description 2
- 210000000744 eyelid Anatomy 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000001151 other effect Effects 0.000 description 2
- 230000035699 permeability Effects 0.000 description 2
- 210000001747 pupil Anatomy 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- 208000003443 Unconsciousness Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 210000000988 bone and bone Anatomy 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/02—Viewing or reading apparatus
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G3/00—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
- G09G3/001—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes using specific devices not provided for in groups G09G3/02 - G09G3/36, e.g. using an intermediate record carrier such as a film slide; Projection systems; Display of non-alphanumerical information, solely or in combination with alphanumerical information, e.g. digital display on projected diapositive as background
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/10—Intensity circuits
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
- H04N5/7475—Constructional details of television projection apparatus
- H04N5/7491—Constructional details of television projection apparatus of head mounted projectors
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0112—Head-up displays characterised by optical features comprising device for genereting colour display
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0132—Head-up displays characterised by optical features comprising binocular systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0141—Head-up displays characterised by optical features characterised by the informative content of the display
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/0613—The adjustment depending on the type of the information to be displayed
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/0626—Adjustment of display parameters for control of overall brightness
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/066—Adjustment of display parameters for control of contrast
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/08—Arrangements within a display terminal for setting, manually or automatically, display parameters of the display terminal
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2360/00—Aspects of the architecture of display systems
- G09G2360/14—Detecting light within display terminals, e.g. using a single or a plurality of photosensors
- G09G2360/144—Detecting light within display terminals, e.g. using a single or a plurality of photosensors the light being ambient light
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4122—Peripherals receiving signals from specially adapted client devices additional display device, e.g. video projector
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/64—Constructional details of receivers, e.g. cabinets or dust covers
Definitions
- the present disclosure relates to a display control device, a display control method, and a computer program.
- a display device that is mounted on a head and views an image that is, a head-mounted display (HMD) is widely known (see, for example, Patent Document 1).
- the head-mounted display has an optical unit for each of the left and right eyes, and is configured to be used in combination with headphones to control vision and hearing.
- Recently, a technique for a so-called transmissive head-mounted display that can present the user in front of the user's line of sight at the time of wearing has also been promoted.
- notification information for notifying an incoming mail or telephone call may be displayed on the display at a timing not intended by the user. If the notification information is suddenly displayed on the display, a part of the user's view is covered with the notification information, which is dangerous.
- the notification information when the user is using a transmissive head-mounted display, if the notification information is blinked or moved in order to pay attention to the notification information, the user's attention will be directed to the blinking or movement, which is dangerous. is there. Furthermore, even when some information is always displayed in a part of the field of view, a part of the user's field of view is covered with the information, which is dangerous.
- the present disclosure provides a new and improved display control apparatus, display control method, and computer program capable of ensuring the user's field of view safely when the user uses a transmissive head mounted display. .
- a situation acquisition unit that acquires information on a situation in which the transmissive display is used, and the transmissive display to the transmissive display using the information acquired by the situation acquisition unit.
- a display control unit that performs display control in which the information display gradually becomes clear, A display control device is provided.
- the present disclosure it is possible to acquire information on a situation in which the transmissive display is used, and to gradually display information on the transmissive display with respect to the transmissive display using the acquired information. Performing a display control to be clarified.
- the computer acquires information on a situation in which the transmissive display is used, and displays information on the transmissive display with respect to the transmissive display using the acquired information.
- a computer program is provided that executes display control that gradually becomes clear.
- a new and improved display control apparatus and display control method capable of ensuring a user's field of view safely. And a computer program can be provided.
- FIG. 3 is an explanatory diagram illustrating an external appearance example of a head mounted display 100 according to an embodiment of the present disclosure.
- 3 is an explanatory diagram illustrating an appearance example of a head mounted display 100 ′ according to an embodiment of the present disclosure.
- FIG. 3 is an explanatory diagram illustrating a functional configuration example of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. It is explanatory drawing which shows an example of the change of the transparency (or the brightness
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 4 is an explanatory diagram illustrating an example of an image displayed on a display unit 110 of a head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 5 is a flowchart illustrating an operation example of the head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 1 is an explanatory diagram illustrating an appearance example of a head mounted display 100 according to an embodiment of the present disclosure.
- the head mounted display 100 is a glasses-type display device as shown in FIG. 1, and includes a display unit 110 including a right eye display unit 110R and a left eye display unit 110L.
- the head-mounted display 100 is a so-called transmissive head-mounted display, and allows the user to check the screen in an actual visual field.
- the nose pad 101 also serves as an attachment sensor, and the attachment is detected when the nasal head of the user (viewer) contacts the nose pad.
- the display unit 110 may be composed of an organic EL display, for example.
- the display unit 110 displays an image in a state where no image is displayed (through state), a state where an image is displayed so as to overlap the actual visual field of the user (image display state), and an image only on one display unit.
- Display control is performed to switch between a state (one-eye display state) and a display state.
- the head mounted display 100 receives an image signal wired or wirelessly from an external device (such as a personal computer, a smartphone, or a tablet portable terminal), and displays an image on the display unit 110 based on the received image signal.
- an external device such as a personal computer, a smartphone, or a tablet portable terminal
- FIG. 1 has described the appearance example of the head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 2 is an explanatory diagram illustrating an appearance example of the head mounted display 100 ′ according to an embodiment of the present disclosure.
- the head-mounted display 100 ′ is a glasses-type display device as shown in FIG. 2, and includes a mounting unit having a frame structure that makes a half turn from the head on both sides to the back of the head.
- the head mounted display 100 ' includes a display unit 110' including a right eye display unit 110R 'and a left eye display unit 110L'.
- the head-mounted display 100 shown in FIG. 1 and the head-mounted display 100 ′ shown in FIG. 2 can allow the user to check the screen while overlapping the actual visual field.
- the head-mounted display is always worn and the display is in front of the user's eyes. Accordingly, notification information for notifying an incoming mail or telephone call may be displayed on the display at a timing not intended by the user. If the notification information is suddenly displayed on the display, a part of the user's view is covered with the notification information, which is dangerous.
- the notification information when the user is using a transmissive head-mounted display, if the notification information is blinked or moved in order to pay attention to the notification information, the user's attention will be directed to the blinking or movement, which is dangerous. is there. Furthermore, even when some information is always displayed in a part of the field of view, a part of the user's field of view is covered with the information, which is dangerous.
- a head mounted display that secures the user's field of view will be described.
- FIG. 3 is an explanatory diagram illustrating a functional configuration example of the head mounted display 100 according to an embodiment of the present disclosure.
- a functional configuration example of the head mounted display 100 according to an embodiment of the present disclosure will be described with reference to FIG.
- the head mounted display 100 includes a sensor unit 101, a system controller 102, a display unit 110, a sensor signal processing unit 111, a sensor control unit 112, Imaging unit 113, imaging signal processing unit 114, imaging control unit 115, audio input unit 116, audio output unit 117, audio signal processing unit 118, audio generation unit 119, image generation unit 120, and display
- the image processing unit 121, the display control unit 122, the display driving unit 123, the communication unit 124, and the storage unit 125 are configured.
- the system controller 102 includes, for example, a microcomputer including a CPU (Central Processing Unit), a ROM (Read Only Memory), a RAM (Random Access Memory), a nonvolatile memory unit, and an interface unit, and controls the entire embodiment. It can function as a control unit.
- the system controller 102 controls each unit based on an internal operation program, and causes the display unit 110 to execute a required image display. Further, the system controller 102 can function as an example of a situation acquisition unit of the present disclosure. That is, the system controller 102 obtains information on the situation in which the head mounted display 100 is used based on the contents of each signal sent from a sensor signal processing unit 111, an imaging signal processing unit 114, and an audio signal processing unit 118 described later. get.
- the sensor unit 101 is a sensor mounted mainly in the vicinity of the display unit 110.
- the sensor unit 101 includes, for example, an acceleration sensor, a gyro sensor, a magnetic sensor, an atmospheric pressure sensor, and the like, and detects, for example, a head movement, a neck movement, a head mounted display 100 movement, and the like as a signal corresponding to a user movement. can do.
- the sensor unit 101 sends a sensor signal to the sensor signal processing unit 111.
- the sensor unit 101 also processes a sensor system signal used by the user to operate the head mounted display 100, such as a capacitance sensor, a button, and a GPS (Global Positioning System). Note that the mounting position of the input device such as the capacitance sensor, the button, and the GPS is not limited to the vicinity of the display unit 110.
- the imaging unit 113 is arranged, for example, in the vicinity of the display unit 110 as a sensor for detecting the user's vision, and images the user's eyes.
- the head mounted display 100 can detect the user's line-of-sight direction, focal length, pupil opening, fundus pattern, eyelid opening and closing, and the like.
- the user's line-of-sight direction, focal length, pupil opening, fundus pattern, eyelid opening / closing, and the like are detected in the vicinity of the display unit 110 to emit light to the user's eye, It may be performed by a light receiving unit that receives reflected light from the unit. For example, it is possible to detect the thickness of the user's crystalline lens from the light reception signal of the light receiving unit.
- the sensor control unit 112 performs sensor control such as which sensor is driven at which timing or in what driving method in accordance with an instruction from the system controller 102.
- the sensor signal processing unit 111 performs various numerical processes such as averaging and variance as preprocessing for sending the sensor signal from the sensor unit 101 to the system controller 102.
- the user's operation is performed for power on / off, display start / end of various information images, change of image contents, display adjustment such as brightness level and hue, change of display area on the display screen, and the like. It may be necessary.
- processing operation trigger detection for example, an operation key operated by a user, an operation element as an operation dial, and an operation detection for detecting the operation of the operation element.
- a mechanism may be provided so that the system controller 102 can detect a user operation.
- the system controller 102 determines whether or not the user's operation intention is appropriate based on the user's situation (for example, eye movement, body behavior or state) detected by the sensor unit 101.
- the operation process may be determined and the process corresponding to the operation process may be executed.
- the head mounted display 100 is configured to be able to detect external environment information (detection information such as the situation around the display device, the location, the date / time, and the subject status) by the sensor unit 101, and the system controller 102 can detect the external environment. An appropriate operation process may be determined based on the information, and the process may be executed.
- external environment information detection information such as the situation around the display device, the location, the date / time, and the subject status
- the voice generation unit 119 performs voice synthesis processing based on an instruction from the system controller 102 to generate a voice signal such as a message voice, or generates an electronic sound or other voice signal to be presented to the user.
- the audio output unit 117 includes, for example, a speaker or earphone speaker (not shown) mounted on the head mounted display 100, and an amplifier circuit for the speaker. By supplying the audio signal generated by the audio generation unit 119 to the audio output unit 117, the user can listen to message audio, electronic sound, and the like.
- the sound output unit 117 may be configured as a so-called bone conduction speaker.
- the audio input unit 116 includes a microphone amplifier unit and an A / D converter that amplifies an audio signal obtained by the microphone, and outputs audio data to the audio signal processing unit 118.
- the audio signal processing unit 118 includes, for example, a digital signal processor, a D / A converter, and the like.
- the audio signal processing unit 118 is supplied with audio data obtained by the audio input unit 116 and audio data generated by the audio generation unit 119.
- the audio signal processing unit 118 performs processing such as volume adjustment, sound quality adjustment, and acoustic effect on the supplied audio data according to the control of the system controller 102.
- the audio signal processing unit 118 converts the processed audio data into an analog signal and supplies the analog signal to the audio output unit 117.
- the audio signal processing unit 118 is not limited to a configuration that performs digital signal processing, and may perform signal processing using an analog amplifier or an analog filter.
- the audio signal output from the audio signal processing unit 118 is output as audio from the earphone speaker of the audio output unit 117. With such a configuration, the user can hear the external sound collected by the sound input unit 116 or can hear the sound generated by the sound generation unit 119.
- the image generation unit 120 generates an image signal based on the control of the system controller 102.
- the system controller 102 causes the image generation unit 120 to generate an image signal, which is an image to be presented to the user, according to the contents and numerical values obtained from each unit, and presents an image image, a graph image, a character display image, and a warning to the user For this purpose, an image signal as an image or other image is generated.
- the image generation unit 120 supplies the generated image signal to the display image processing unit 121.
- the display image processing unit 121 is configured by, for example, a video processor, and executes various display processes on the image signal supplied from the image generation unit 120.
- the display image processing unit 121 can perform brightness level adjustment, color correction, contrast adjustment, sharpness (contour emphasis) adjustment of an image signal, and the like.
- the display image processing unit 121 also sets a display position on the display screen in the display unit 110. Further, the display image processing unit 121 generates an enlarged image obtained by enlarging a part of the image signal, generation of a reduced reduced image, soft focus, mosaic, inversion of luminance, and highlight display (highlighted display) of a part of the image.
- Image effect processing such as change in the atmosphere of the entire color, separation and synthesis of image signals for divided display of images, generation of character images and image images, and generated images supplied from the image generation unit 120 A process of combining with a signal may be performed.
- the display image processing unit 121 is supplied with the image signal generated by the image generation unit 120.
- the display signal processing unit 121 performs necessary signal processing for display on the image signal and supplies the image signal to the display driving unit 123.
- the display driving unit 123 includes a pixel driving circuit for causing the display unit 110 to display the image signal supplied from the display image processing unit 121. That is, the display driving unit 123 applies a driving signal based on the image signal to each pixel arranged in a matrix in the display unit 110 at a predetermined horizontal / vertical driving timing, and causes the display unit 110 to perform display.
- the display driving unit 123 can also control the luminance of each pixel of the display unit 110 so that the entire screen and a part of the screen are in a through state.
- the display control unit 122 displays the processing operation of the display image processing unit 121, the operation of the display driving unit 123, and the display unit 110 (including processing for the left and right display units 110R and 110L) based on an instruction from the system controller 102.
- the display image processing unit 121 is instructed to perform signal processing. Further, the display control unit 122 controls the display driving unit 123 so as to be switched between the through state, the image display state, and the one-eye display state.
- the imaging unit 113 is obtained by using a lens system including an imaging lens, an aperture, a zoom lens, a focus lens, a drive system for causing the lens system to perform a focus operation and a zoom operation, and a lens system.
- a solid-state imaging device array or the like that generates an imaging signal by detecting captured imaging light and performing photoelectric conversion is provided.
- the solid-state imaging device array is, for example, a CCD (Charge Coupled Device) sensor array or a CMOS (Complementary Metal Oxide Semiconductor) sensor array.
- the imaging unit 113 images a scene in front of the user wearing the head mounted display 100. Of course, a scene in another direction such as behind the user may be captured by the imaging unit 113.
- the imaging signal processing unit 114 includes a sample hold / AGC (Automatic Gain Control) circuit that performs gain adjustment and waveform shaping on a signal obtained by the solid-state imaging device of the imaging unit 113, and a video A / D converter. An imaging signal is obtained.
- the imaging signal processing unit 114 may also perform white balance processing, luminance processing, color signal processing, blur correction processing, and the like on the imaging signal.
- the imaging control unit 115 controls the operations of the imaging unit 113 and the imaging signal processing unit 114 based on instructions from the system controller 102.
- the imaging control unit 115 controls on / off of operations of the imaging unit 113 and the imaging signal processing unit 114, for example.
- the imaging control unit 115 may perform control (motor control) for causing the imaging unit 113 to perform operations such as auto focus, automatic exposure adjustment, aperture adjustment, zoom, and focus change.
- control unit 115 controls the operation of the movable mechanism based on an instruction from the system controller 102 to perform imaging.
- the direction of the imaging lens in the unit 113 may be changed.
- the imaging control unit 115 may include a timing generator.
- the imaging control unit 115 may control the signal processing operation by the timing signal generated by the timing generator with respect to the solid-state imaging device and the sample hold / AGC circuit of the imaging signal processing unit 114 and the video A / D converter. . Further, the imaging frame rate may be variably controlled by this timing control.
- the imaging control unit 115 may control imaging sensitivity and signal processing in the solid-state imaging device and the imaging signal processing unit 114.
- the imaging control unit 115 performs gain control of a signal read from the solid-state imaging device, for example, as imaging sensitivity control, black level setting control, various coefficient control of imaging signal processing at the digital data stage, and correction amount in blur correction processing Control or the like may be performed.
- the imaging control unit 115 performs overall sensitivity adjustment that does not particularly consider the wavelength band, and sensitivity adjustment that adjusts imaging sensitivity in a specific wavelength band such as an infrared region or an ultraviolet region (for example, a specific wavelength band). Imaging that cuts the image) is also possible.
- Sensitivity adjustment according to the wavelength can be performed by inserting a wavelength filter in the imaging lens system or wavelength filter calculation processing on the imaging signal.
- the imaging control unit 115 can perform sensitivity control by insertion control of a wavelength filter, designation of a filter calculation coefficient, or the like.
- the captured image signal obtained by the imaging unit 113 and the imaging signal processing unit 114 is supplied to the display image processing unit 121 together with the information image signal generated by the image generation unit 120.
- the display image processing unit 121 performs the above-described various signal processing on the captured image signal and the information image signal, and causes the display unit 110 to simultaneously display the captured image and the information image for the captured image signal and the information image signal. Then, signal processing (image composition processing) as screen division is performed.
- the image signal synthesized by the display image processing unit 121 is supplied to the display driving unit 123 and displayed on the display unit 110, so that the captured image and other images are simultaneously displayed on the display unit 110. It will be. In other words, the user can view various images even when viewing the image captured by the imaging unit 113.
- User operation may be required for the start and end of the imaging operation, zoom operation, focus operation, and captured image adjustment.
- user operations are required to turn on / off the power, start and end display of various information images, change the image contents, adjust the display of the brightness level and hue, change the display area on the display screen, and the like.
- the head mounted display 100 may be provided with an operation element such as an operation key.
- the user's situation for example, eye movement and body behavior
- the system controller 102 may determine the user's intention to operate and appropriate operation processing from the state and the like, and execute processing corresponding to the determination.
- the head-mounted display 100 is configured to be able to detect external world information (detection information such as the situation around the display device, location, date and time, and subject status), and the system controller 102 is based on the external world information. An appropriate operation process may be determined and the process may be executed.
- external world information detection information such as the situation around the display device, location, date and time, and subject status
- the storage unit 125 is a part that records and reproduces data on a predetermined recording medium, and is realized as, for example, an HDD (Hard Disk Drive) or an SSD (Solid State Drive).
- various recording media such as a solid-state memory such as a flash memory, a memory card with a built-in fixed memory, an optical disk, a magneto-optical disk, and a hologram memory are conceivable.
- the storage unit 125 performs recording / reproduction according to the recording medium employed. It suffices if the configuration is such that can be executed.
- image data as an imaging signal captured by the imaging unit 113 and processed by the imaging signal processing unit 114, image data received by the communication unit 124, and various information image signals generated by the image generation unit 120 are The storage unit 125 may be supplied.
- voice data obtained by the voice input unit 116, voice data received by the communication unit 124, and voice data generated by the voice generation unit 119 may also be supplied to the storage unit 125.
- the storage unit 125 performs an encoding process for recording on a predetermined recording medium on the supplied image data and audio data according to the control of the system controller 102 and records the encoded data on the recording medium.
- the storage unit 125 reproduces and outputs image data and audio data recorded on the recording medium according to the control of the system controller 102.
- Examples of data reproduced by the storage unit 125 include moving image content such as movies and video clips, still image content captured by a digital still camera or the like and recorded on a recording medium, data such as an electronic book, and a user using a personal computer or the like. Any data to be displayed such as image data, text data, computer use data such as spreadsheet data, game images, and the like created in the above is assumed.
- the communication unit 124 transmits / receives data to / from an external device.
- an imaging device such as a video camera having a communication function or a digital still camera, a computer device, a smartphone, a smart watch, an AV device such as a video storage device or a television receiver, a network server device, or the like is assumed.
- the communication unit 124 may be configured to perform network communication via short-range wireless communication with a network access point using a method such as a wireless LAN or Bluetooth (registered trademark), or an external device having a corresponding communication function. It is good also as a structure which performs direct wireless communication between.
- the external device As data transmitted from the external device to the head mounted display 100, when the external device is an imaging device, image data captured by the imaging device is assumed, and the external device is a device that is a content source device.
- Video content such as movies and video clips, still image content captured by a digital still camera and recorded on a recording medium, data such as electronic books, image data created by a user using a personal computer, text data, Any data to be displayed such as computer use data such as spreadsheet data and game images is assumed.
- the audio data obtained by the audio input unit 116, the audio data reproduced by the storage unit 125, and the audio data received by the communication unit 124 are supplied to the audio signal processing unit 118 according to an instruction from the system controller 102. Therefore, the user can watch the external sound when the captured image is captured, the image and sound reproduced by the storage unit 125, and the image and sound received by the communication unit 124 with the device attached. Can be watched.
- the image generated by the image generation unit 120 is supplied to the display image processing unit 121 together with the captured image, the reproduced image, or the received image, so that various images can be obtained together with the captured image, the reproduced image, or the received image.
- An information image is displayed on the display unit 110.
- the generated audio data is supplied to the audio signal processing unit 118, so that the user is listening to, for example, external audio, reproduced audio, or received audio. At this time, it is possible to listen to message voices, warning sounds, and the like generated by the voice generation unit 119.
- the system controller 102 may make a determination. Also in this case, an operator such as an operation key operated by the user is provided, and processing corresponding to the operation is executed, or the user's situation (for example, eye movement, body behavior, state, etc.) detected by the sensor unit 101 ), The system controller may determine the operation intention of 102 users and appropriate operation processing, and execute processing corresponding to the determination.
- the head-mounted display 100 is configured to be able to detect external world information (detection information such as the situation around the display device, location, date and time, and subject status), and the system controller 102 is based on the external world information. An appropriate operation process may be determined and the process may be executed.
- external world information detection information such as the situation around the display device, location, date and time, and subject status
- the system controller 102 executes a process of displaying an image indicating that the incoming call has been received, for example, a predetermined icon, on the display unit 110. At this time, the system controller 102 does not cause the display unit 110 to display an image indicating that there is an incoming call, but displays the image on the display unit 110 while changing the transmittance of the icon (or the brightness of the display unit 110). .
- the head-mounted display 100 displays the image indicating that an incoming call has been received on the display unit 110 while changing the transmittance of the icon (or the brightness of the display unit 110).
- the field of view of the user wearing the mount display 100 can be secured safely.
- 4A to 4D are explanatory diagrams illustrating an example of an image displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure.
- An icon when an interruption such as an incoming mail occurs is illustrated. It is explanatory drawing which shows the example of a display.
- 4A to 4D show scenes that the user of the head mounted display 100 visually recognizes through the display unit 110 in addition to the image displayed on the display unit 110 by the head mounted display 100.
- FIG. 4A is an example of an image displayed on the display unit 110 immediately after an interruption such as an incoming mail occurs.
- the system controller 102 displays an icon indicating that an interrupt has occurred on the display unit 110.
- the system controller 102 instructs the display control unit 122 to increase the transparency of the icon 131 displayed on the display unit 110 or decrease the luminance of the display unit 110. Instruct. That is, the system controller 102 executes control that makes it difficult for the icon 131 to be recognized as a display immediately after an interruption such as an incoming mail occurs.
- FIG. 4B is an example of an image displayed on the display unit 110 after a predetermined time (for example, after 100 milliseconds to 200 milliseconds) from the state of FIG. 4A.
- FIG. 4C is an example of an image displayed on the display unit 110 after a predetermined time (for example, after 100 milliseconds to 200 milliseconds) from the state of FIG. 4B.
- the system controller 102 controls the display so that the icon 131 can be recognized as a display by gradually decreasing the transparency of the icon 131 or increasing the brightness of the display unit 110.
- the unit 122 is instructed.
- FIG. 4D is an example of an image displayed on the display unit 110 after a predetermined time (for example, after 100 milliseconds to 200 milliseconds) from the state of FIG. 4C.
- the system controller 102 sets the display control unit 122 so that the icon 131 can be completely recognized as a display by minimizing the transparency of the icon 131 or maximizing the luminance of the display unit 110.
- the display control unit 122 executes control to minimize the transparency of the icon 131 or maximize the luminance of the display unit 110 based on an instruction from the system controller 102.
- FIG. 5 is an explanatory diagram illustrating an example of a change in the transparency of the icon 131 (or the luminance of the display unit 110) displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure.
- the horizontal axis of the graph shown in FIG. 5 indicates time, and the vertical axis indicates the transparency of the icon 131 (or the luminance of the display unit 110). It is assumed that time t1 is the time of the state of FIG. 4A, time t2 is the time of the state of FIG. 4B, time t3 is the time of the state of FIG. 4C, and time t4 is the time of the state of FIG.
- the head-mounted display 100 slowly changes the transmittance of the icon 131 (or the luminance of the display unit 110) indicating that there is an incoming call, thereby making the user unconscious. It avoids attracting attention to the animation.
- the system controller 102 does not display the icon 131 again like the solid line of the graph shown in FIG.
- the brightness of the icon 131 or the brightness of the display unit 110 may not be changed as indicated by the broken line in the graph shown in FIG.
- the system controller 102 When the system controller 102 receives another incoming call after displaying the icon 131 on the display unit 110 as shown in FIGS. 4A to 4C, the system controller 102 displays the icon 131 corresponding to the other incoming call as shown in FIG. 4A. As shown in FIGS. 4C to 4C, the icon 131 may be newly displayed on the display unit 110. Further, after the icon 131 displayed on the display unit 110 is deleted from the display unit 110, the icon 131 corresponding to another incoming call is displayed. May be displayed on the display unit 110 as shown in FIGS. 4A to 4C.
- the head-mounted display 100 gradually changes the icon 131 by changing the transmittance of the icon 131 (or the brightness of the display unit 110) indicating that there is an incoming call with time.
- the system controller 102 may change either the transmittance of the icon 131 or the luminance of the display unit 110.
- the system controller 102 changes the brightness of the display unit 110 with time and changes only a specific icon 131 when changing the icons 131 uniformly. In the case of making it, the transmittance of the icon 131 may be changed with time.
- the head mounted display 100 includes a display process for displaying the icon 131 so that the icon 131 gradually becomes clear as illustrated in FIGS. 4A to 4C, and the icon 131 from the beginning.
- the display process to be clearly displayed on the display unit 110 may be switchable. Such switching may be set by the user, or may be automatically set according to a situation in which the head mounted display 100 is used. For example, in a situation where there is a problem in suddenly hindering the user's field of view, such as when there is a user outside the room, the head mounted display 100 is controlled by the user's own setting or automatic setting by the head mounted display 100, as shown in FIGS. Display processing for displaying on the display unit 110 may be executed so that the icon 131 gradually becomes clear as shown in FIG.
- the head mounted display 100 may change the transmittance of the image (or the luminance of the display unit 110) indicating that there is an incoming call, depending on the intensity of external light. That is, when the outside light is weak, the user can fully recognize the change even if the change amount of the image transmittance (or the luminance of the display unit 110) is small. Change is difficult to recognize. Therefore, the head mounted display 100 according to an embodiment of the present disclosure measures the external light intensity using a sensor that can measure the intensity of external light, such as an illuminance sensor or a camera, and transmits the sensor unit 101 depending on the intensity of external light. The amount of change in brightness and brightness may be adjusted.
- FIG. 6A is an explanatory diagram illustrating an example of a change in transmittance (or luminance of the display unit 110) of the icon 131 displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure.
- a reference numeral 141 in FIG. 6A indicates a display transition of the transmittance (or the luminance of the display unit 110) when the external light intensity is weak, and a reference numeral 142 indicates a transmission when the external light intensity is stronger than the reference numeral 141.
- the reference numeral 143 indicates the display transition of the transmittance (or the luminance of the display unit 110) when the external light intensity is higher than that of the reference numeral 142. ing.
- the head-mounted display 100 slowly reduces the transmittance of the icon 131 (or the luminance of the display unit 110) indicating that there is an incoming call according to the strength of the external light intensity. This prevents the user from being unconsciously attracted to the animation.
- the system controller 102 adjusts the transparency of the icon 131 or the brightness of the display unit 110 so that the icon 131 is not displayed again.
- the transparency of the icon 131 or the luminance of the display unit 110 may not be changed.
- the time until the icon 131 can be completely recognized as a display is the same regardless of the strength of the external light intensity.
- the example is limited to this example. is not.
- the head mounted display 100 performs display control that increases the time until the icon 131 can be completely recognized as a display as the intensity of external light increases. May be.
- FIG. 6B is an explanatory diagram illustrating an example of a change in transmittance (or luminance of the display unit 110) of the icon 131 displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 6B shows the transmittance of the icon 131 when executing the display control for increasing the time until the icon 131 can be completely recognized as a display as the intensity of the external light intensity increases.
- an example of a change in luminance of the display unit 110 is shown.
- the method of changing the transmittance of icons and display information is combined with sensor values obtained by detecting the user's line of sight, head movement, and the like with the sensor unit 101.
- sensor values obtained by detecting the user's line of sight, head movement, and the like with the sensor unit 101.
- 7A to 7C are explanatory diagrams illustrating an example of an image displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure, and a table of icons when an interruption such as an incoming mail occurs. It is explanatory drawing which shows an example.
- 7A to 7C indicate the position of the line of sight of the user wearing the head mounted display 100.
- the system controller 102 displays the icon 131 so that the transparency of the icon 131 is high or the luminance of the display unit 110 is low.
- the control unit 122 is instructed.
- the system controller 102 changes the transmittance of the icon 131 (or the brightness of the display unit 110), so that it is clear rather than transparent.
- the display control unit 122 is instructed to display on the display unit 110.
- the head-mounted display 100 performs display control of icons according to the location of the user's line of sight, so that the display is not usually displayed in front of the eyes, but the head-mounted display 100 is mounted.
- the icon 131 can be displayed on the display unit 110 only when the user wants to see the information and performs an operation to change the line of sight.
- FIG. 8 is an explanatory diagram showing an example of a specific calculation method when the display control of the icon 131 is executed according to the user's line of sight.
- the coordinates of the display area of the icon 131 are (X1, Y1), and the coordinates of the line of sight are (X2, Y2).
- the system controller 102 has a high transmittance (or low brightness) and looks transparent, and when A is small, the transmittance is low (or high brightness) and clearly appears.
- the display control unit 122 may be instructed.
- the head mounted display 100 detects the angle of the user's head using the sensor unit 101 instead of the user's line of sight, and similarly controls display of icons using the angle. You can do it.
- 9A to 9C are explanatory diagrams illustrating an example of an image displayed on the display unit 110 of the head mounted display 100 according to an embodiment of the present disclosure, and a table of icons when an interruption such as an incoming mail occurs. It is explanatory drawing which shows an example.
- the system controller 102 increases the transparency of the icon 131 or the display unit 110.
- the display control unit 122 is instructed to display on the display unit 110 at a low luminance.
- the system controller 102 displays the icon 131.
- the display control unit 122 is instructed to change the transmittance (or the luminance of the display unit 110) to be clearly displayed on the display unit 110 instead of being transparent.
- the head mounted display 100 executes display control that clearly displays the icon 131 according to the angle of the user's head, but the present disclosure is limited to such an example. is not.
- the system controller 102 may execute display control for clearly displaying the icon 131 when the amount of movement of the user's head exceeds a predetermined threshold.
- the head mounted display 100 may detect the user's moving speed using the sensor unit 101, and similarly perform icon display control using the moving speed. That is, if an icon is displayed on the display unit 110 while the user is walking and the user's attention is directed to the icon, the attention to the front is neglected, which is dangerous. Therefore, the head mounted display 100 according to the embodiment of the present disclosure performs display control so that the icon is displayed lightly when the user is walking and the icon is clearly displayed when the user is stopped. May be executed.
- the head-mounted display 100 detects the user's moving speed and performs emergency display control using the moving speed even if the user is walking. If it is an icon that means a notification that requires sexuality, display control may be executed so that the icon is clearly displayed even when the user is walking.
- the notification that requires urgency for example, a call from a specific party, an incoming mail, a reminder with high priority, or the like can be assumed. It can be determined as appropriate by the user, and is not limited to specific contents.
- the head mounted display 100 detects a user activity state using the imaging unit 113 and the voice input unit 116 in addition to the sensor unit 101, and uses the detection result of the user activity state. Then, display control of icons displayed on the display unit 110 may be executed.
- the system controller 102 knows that the user is driving the automobile. In the state of driving a car, as in the case where the above-mentioned user is walking, if the icon is displayed on the display unit 110 and the user's attention is directed to the icon, attention to the front is weakened. It is very dangerous. Thus, if the user knows that the user is driving a car, the system controller 102 may cause the icon to be dimmed or not displayed at all while the user is driving the car, When the automobile stops, display control for clearly displaying the icon as described above may be executed.
- the system controller 102 uses the head-mounted display 100 if it is found that the sound collected by the voice input unit 116 includes a human conversation sound. It can be seen that the user is talking with another person. When a user using the head mounted display 100 is talking to another person and displaying an icon on the display unit 110 and directing the user's attention to the icon, the conversation with the other person is negligible. This hinders smooth communication. Therefore, if it is known that the user is in a conversation with another person, the system controller 102 displays the icon lightly while the user is speaking, or does not display the icon at all, Display control for clearly displaying the icon as described above may be executed when the communication is interrupted.
- the head mounted display 100 recognizes the situation of the user who uses the head mounted display 100 based on the image captured by the imaging unit 113 and the sound collected by the audio input unit 116.
- the icon display on the display unit 110 can be controlled in accordance with the user's situation.
- the example in which the icon is gradually displayed on the display unit 110 by changing the transmittance of the icon or the brightness of the display unit 110 has been described.
- the head mounted display according to an embodiment of the present disclosure The icon 100 may be displayed on the display unit 110 gradually and clearly by gradually increasing the size of the icon.
- the function configuration example of the head mounted display 100 according to an embodiment of the present disclosure has been described above. Subsequently, an operation example of the head mounted display 100 according to an embodiment of the present disclosure will be described.
- FIG. 10 is a flowchart illustrating an operation example of the head mounted display 100 according to an embodiment of the present disclosure.
- FIG. 10 shows an image indicating that there is information to be notified to the user, such as an incoming call or mail on the display unit 110, for example, a process for displaying a predetermined icon.
- 3 is an operation example of the head mounted display 100.
- an operation example of the head mounted display 100 according to an embodiment of the present disclosure will be described with reference to FIG.
- Step S101 processing for displaying an icon indicating that there is information to be notified to the user, such as an incoming call or mail, on the display unit 110 occurs.
- step S101 processing for displaying an icon indicating that there is information to be notified to the user, such as an incoming call or mail, on the display unit 110 occurs.
- step S101 processing for displaying the icon on the display unit 110 has occurred.
- step S102 the head mounted display 100 displays the icon while adjusting the transmittance of the icon (or the brightness of the display unit 110).
- the display control unit 122 executes the adjustment and display processing in step S102 according to an instruction from the system controller 102, for example.
- the adjustment of the transmittance of the icon may be such that the icon is clearly displayed over time, and the icon is clearly displayed according to the position of the line of sight.
- the icon may be displayed according to the inclination of the head, or the icon may be clearly displayed according to the moving speed of the user. It may be a thing to make.
- the field of view of the user wearing the head mounted display 100 is safe. Secured.
- the head mounted display 100 may execute display control for gradually and clearly displaying icons using color information in the user's field of view. For example, when there is a user who uses the head mounted display 100 in a room with a white wall surface, the system controller 102 first performs display control for displaying a white icon on the display unit 110. When there is a user who uses the head mounted display 100 in a room with a white wall surface, the head mounted display 100 according to the embodiment of the present disclosure is displayed on the display unit 110 by first displaying a white icon. It is possible to prevent the user from noticing that the icon is displayed at 110.
- the system controller 102 executes display control for gradually changing the icon to the original color or, if white is the original color, to a different color.
- the head-mounted display 100 displays information that causes the user to be conscious of information displayed unintentionally or disturb the visual field. Can be avoided.
- the system controller 102 first executes display control for displaying an icon of a color excluding the green component on the display unit 110. Since green is the most sensitive color for humans, the head-mounted display 100 according to an embodiment of the present disclosure is displayed unconsciously by the user by displaying an icon of a color excluding the green component. It is possible to avoid displaying information that is conscious of information or disturbs the field of view.
- the transmissive head-mounted display 100 when displaying information that covers a part of the user's field of view, the icon transmittance or the luminance of the display unit 110 is displayed. There is provided a head mounted display 100 that clearly displays information while adjusting. When displaying information to be notified to the user on the display unit 110, the head-mounted display 100 according to an embodiment of the present disclosure responds to the tilt of the head according to the position of the line of sight, for example, with time. The information is clearly displayed while adjusting the transmittance or the brightness of the display unit 110 according to the moving speed of the user.
- the head-mounted display 100 is displayed unconsciously by the user by clearly displaying information while adjusting the transmittance or the luminance of the display unit 110. It is possible to avoid displaying information that is conscious of information or disturbs the field of view, and it is possible to reduce the risk of the user when using the head mounted display 100.
- the display control on the display unit 110 is performed by the display control unit 122 according to an instruction from the system controller 102 included in the head mounted display 100.
- the present disclosure is limited to such an example. Is not to be done.
- the display control on the display unit 110 may be executed by a device different from the head mounted display 100 and connected to the head mounted display 100 by wire or wirelessly.
- each step in the processing executed by each device in this specification does not necessarily have to be processed in chronological order in the order described as a sequence diagram or flowchart.
- each step in the processing executed by each device may be processed in an order different from the order described as the flowchart, or may be processed in parallel.
- the software that implements the user interface and application shown in the above embodiment may be realized as a web application used via a network such as the Internet.
- the web application may be realized by, for example, a markup language such as HTML (HyperText Markup Language), SGML (Standard Generalized Markup Language), or XML (Extensible Markup Language).
- a situation acquisition unit that acquires information on a situation in which a transmissive display that is mounted in front of the user's eyes is used; Using the information acquired by the status acquisition unit, a display control unit that executes display control for the transmissive display so that information display on the transmissive display gradually becomes clear, A display control device.
- the situation acquisition unit acquires the state of external light, The display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the state of the external light.
- the situation acquisition unit acquires the state of the line of sight of the user viewing the transmissive display, The display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the state of the line of sight.
- the situation acquisition unit acquires a state of a head movement of a user who visually recognizes the transmissive display, The display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the state of movement of the head.
- the situation acquisition unit acquires the state of the tilt of the head of the user viewing the transmissive display, The display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the magnitude of the tilt of the head.
- the situation acquisition unit acquires a state of an activity state of a user who visually recognizes the transmissive display, The display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the activity state.
- the situation acquisition unit acquires a background of the line of sight of the transmissive display of the user,
- the display control device according to (1), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the state of the background.
- the display control device according to (8), wherein the display control unit executes display control in which information display on the transmissive display gradually becomes clear according to the color of the background.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- Optics & Photonics (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Controls And Circuits For Display Device (AREA)
Abstract
Description
を備える、表示制御装置が提供される。
1.本開示の一実施形態
1.1.ヘッドマウントディスプレイの外観例
1.2.ヘッドマウントディスプレイの機能構成例
1.3.ヘッドマウントディスプレイの動作例
2.まとめ
[1.1.ヘッドマウントディスプレイの外観例]
まず、図面を参照しながら本開示の一実施形態に係るヘッドマウントディスプレイの外観例について説明する。図1は、本開示の一実施形態に係るヘッドマウントディスプレイ100の外観例を示す説明図である。
続いて、本開示の一実施形態に係るヘッドマウントディスプレイの機能構成例について説明する。図3は、本開示の一実施形態に係るヘッドマウントディスプレイ100の機能構成例を示す説明図である。以下、図3を用いて本開示の一実施形態に係るヘッドマウントディスプレイ100の機能構成例について説明する。
図10は、本開示の一実施形態に係るヘッドマウントディスプレイ100の動作例を示す流れ図である。図10に示したのは、表示部110へ、電話やメールの着信等があった等、ユーザに通知すべき情報があることを示す画像、例えば所定のアイコンを表示させる処理を実行する際の、ヘッドマウントディスプレイ100の動作例である。以下、図10を用いて本開示の一実施形態に係るヘッドマウントディスプレイ100の動作例について説明する。
以上説明したように本開示の一実施形態によれば、透過型のヘッドマウントディスプレイ100において、ユーザの視野の一部を覆うような情報の表示に際し、アイコンの透過率、または表示部110の輝度を調整しながらはっきりと情報の表示を行なっていくヘッドマウントディスプレイ100が提供される。本開示の一実施形態に係るヘッドマウントディスプレイ100は、ユーザに通知すべき情報を表示部110に表示する際に、例えば、時間の経過と共に、視線の場所に応じて、頭部の傾きに応じて、またはユーザの移動速度に応じて、透過率、または表示部110の輝度を調整しながらはっきりと情報の表示を行なっていく。
(1)
ユーザの眼前に位置して装着される透過型ディスプレイが使用されている状況の情報を取得する状況取得部と、
前記状況取得部が取得した前記情報を用いて、前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を前記透過型ディスプレイに対して実行する表示制御部と、
を備える、表示制御装置。
(2)
前記状況取得部は、外光の状態を取得し、
前記表示制御部は、前記外光の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(3)
前記表示制御部は、前記外光が強いほど、前記透過型ディスプレイへの情報表示が明確になるまでの時間が長くなる表示制御を実行する、前記(2)に記載の表示制御装置。
(4)
前記状況取得部は、前記透過型ディスプレイを視認するユーザの視線の状態を取得し、
前記表示制御部は、前記視線の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(5)
前記状況取得部は、前記透過型ディスプレイを視認するユーザの頭の動きの状態を取得し、
前記表示制御部は、前記頭の動きの状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(6)
前記状況取得部は、前記透過型ディスプレイを視認するユーザの頭の傾きの状態を取得し、
前記表示制御部は、前記頭の傾きの大きさに応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(7)
前記状況取得部は、前記透過型ディスプレイを視認するユーザの活動状態の状態を取得し、
前記表示制御部は、前記活動状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(8)
前記状況取得部は、前記ユーザの前記透過型ディスプレイの視線の先の背景を取得し、
前記表示制御部は、前記背景の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(9)
前記表示制御部は、前記背景の色に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(8)に記載の表示制御装置。
(10)
前記表示制御部は、前記透過型ディスプレイへの情報表示の透過率を変化させることで前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(11)
前記表示制御部は、前記透過型ディスプレイの輝度を変化させることで前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、前記(1)に記載の表示制御装置。
(12)
透過型ディスプレイが使用されている状況の情報を取得することと、
取得された前記情報を用いて前記透過型ディスプレイに対して前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行することと、
を含む、表示制御方法。
(13)
コンピュータに、
透過型ディスプレイが使用されている状況の情報を取得することと、
取得された前記情報を用いて前記透過型ディスプレイに対して前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行することと、
を実行させる、コンピュータプログラム。
101 センサ部
102 システムコントローラ
110 表示部
111 センサ信号処理部
112 センサ制御部
113 撮像部
114 撮像信号処理部
115 撮像制御部
116 音声入力部
117 音声出力部
118 音声信号処理部
119 音声生成部
120 画像生成部
121 表示画像処理部
122 表示制御部
123 表示駆動部
124 通信部
125 ストレージ部
131 アイコン
Claims (13)
- ユーザの眼前に位置して装着される透過型ディスプレイが使用されている状況の情報を取得する状況取得部と、
前記状況取得部が取得した前記情報を用いて、前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を前記透過型ディスプレイに対して実行する表示制御部と、
を備える、表示制御装置。 - 前記状況取得部は、外光の状態を取得し、
前記表示制御部は、前記外光の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記表示制御部は、前記外光が強いほど、前記透過型ディスプレイへの情報表示が明確になるまでの時間が長くなる表示制御を実行する、請求項2に記載の表示制御装置。
- 前記状況取得部は、前記透過型ディスプレイを視認するユーザの視線の状態を取得し、
前記表示制御部は、前記視線の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記状況取得部は、前記透過型ディスプレイを視認するユーザの頭の動きの状態を取得し、
前記表示制御部は、前記頭の動きの状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記状況取得部は、前記透過型ディスプレイを視認するユーザの頭の傾きの状態を取得し、
前記表示制御部は、前記頭の傾きの大きさに応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記状況取得部は、前記透過型ディスプレイを視認するユーザの活動状態の状態を取得し、
前記表示制御部は、前記活動状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記状況取得部は、前記ユーザの前記透過型ディスプレイの視線の先の背景を取得し、
前記表示制御部は、前記背景の状態に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。 - 前記表示制御部は、前記背景の色に応じて前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項8に記載の表示制御装置。
- 前記表示制御部は、前記透過型ディスプレイへの情報表示の透過率を変化させることで前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。
- 前記表示制御部は、前記透過型ディスプレイの輝度を変化させることで前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行する、請求項1に記載の表示制御装置。
- 透過型ディスプレイが使用されている状況の情報を取得することと、
取得された前記情報を用いて前記透過型ディスプレイに対して前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行することと、
を含む、表示制御方法。 - コンピュータに、
透過型ディスプレイが使用されている状況の情報を取得することと、
取得された前記情報を用いて前記透過型ディスプレイに対して前記透過型ディスプレイへの情報表示が徐々に明確になる表示制御を実行することと、
を実行させる、コンピュータプログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP15752499.2A EP3109854B1 (en) | 2014-02-20 | 2015-02-05 | Display control device, display control method, and computer program |
JP2016504031A JP6743691B2 (ja) | 2014-02-20 | 2015-02-05 | 表示制御装置、表示制御方法およびコンピュータプログラム |
CN201580008449.5A CN106030692B (zh) | 2014-02-20 | 2015-02-05 | 显示控制装置、显示控制方法及计算机程序 |
US15/101,739 US10146057B2 (en) | 2014-02-20 | 2015-02-05 | Display control device, display control method, and computer program |
US16/184,409 US10324294B2 (en) | 2014-02-20 | 2018-11-08 | Display control device, display control method, and computer program |
US16/413,276 US10437069B2 (en) | 2014-02-20 | 2019-05-15 | Display control device, display control method, and computer program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014-030945 | 2014-02-20 | ||
JP2014030945 | 2014-02-20 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/101,739 A-371-Of-International US10146057B2 (en) | 2014-02-20 | 2015-02-05 | Display control device, display control method, and computer program |
US16/184,409 Continuation US10324294B2 (en) | 2014-02-20 | 2018-11-08 | Display control device, display control method, and computer program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015125626A1 true WO2015125626A1 (ja) | 2015-08-27 |
Family
ID=53878135
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/053312 WO2015125626A1 (ja) | 2014-02-20 | 2015-02-05 | 表示制御装置、表示制御方法およびコンピュータプログラム |
Country Status (5)
Country | Link |
---|---|
US (3) | US10146057B2 (ja) |
EP (1) | EP3109854B1 (ja) |
JP (1) | JP6743691B2 (ja) |
CN (1) | CN106030692B (ja) |
WO (1) | WO2015125626A1 (ja) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017081915A1 (ja) * | 2015-11-10 | 2017-05-18 | ソニー株式会社 | 画像処理装置、画像処理方法およびプログラム |
WO2017094557A1 (ja) * | 2015-12-02 | 2017-06-08 | コニカミノルタ株式会社 | 電子機器及びヘッドマウントディスプレイ |
JP2017516196A (ja) * | 2014-04-17 | 2017-06-15 | マイクロソフト テクノロジー ライセンシング,エルエルシー | 会話検出 |
WO2018198503A1 (ja) * | 2017-04-26 | 2018-11-01 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
JP2018534643A (ja) * | 2015-09-16 | 2018-11-22 | グーグル エルエルシー | 拡張および/または仮想現実環境におけるタッチスクリーンホバーの検出 |
JP2018195321A (ja) * | 2018-07-03 | 2018-12-06 | 株式会社東芝 | ウェアラブル端末及び方法 |
JP2019197564A (ja) * | 2019-07-03 | 2019-11-14 | 株式会社東芝 | ウェアラブル端末、システム及び方法 |
US10679648B2 (en) | 2014-04-17 | 2020-06-09 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
JP2020520494A (ja) * | 2017-05-04 | 2020-07-09 | ソニー インタラクティブ エンタテインメント ヨーロッパ リミテッド | ヘッドマウントディスプレイおよび方法 |
US10755545B2 (en) | 2015-09-01 | 2020-08-25 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
WO2020234939A1 (ja) * | 2019-05-17 | 2020-11-26 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
JP2022038495A (ja) * | 2020-08-26 | 2022-03-10 | ソフトバンク株式会社 | 表示制御装置、プログラム、及びシステム |
JP2022038783A (ja) * | 2020-08-27 | 2022-03-10 | Necパーソナルコンピュータ株式会社 | 情報処理装置及びその通知制御方法並びに通知制御プログラム |
Families Citing this family (51)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9298007B2 (en) | 2014-01-21 | 2016-03-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9229233B2 (en) | 2014-02-11 | 2016-01-05 | Osterhout Group, Inc. | Micro Doppler presentations in head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US9594246B2 (en) | 2014-01-21 | 2017-03-14 | Osterhout Group, Inc. | See-through computer display systems |
US20150228119A1 (en) | 2014-02-11 | 2015-08-13 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9575321B2 (en) | 2014-06-09 | 2017-02-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US9299194B2 (en) | 2014-02-14 | 2016-03-29 | Osterhout Group, Inc. | Secure sharing in head worn computing |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US9494800B2 (en) | 2014-01-21 | 2016-11-15 | Osterhout Group, Inc. | See-through computer display systems |
US12093453B2 (en) | 2014-01-21 | 2024-09-17 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9811159B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US12105281B2 (en) | 2014-01-21 | 2024-10-01 | Mentor Acquisition One, Llc | See-through computer display systems |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9852545B2 (en) | 2014-02-11 | 2017-12-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9401540B2 (en) | 2014-02-11 | 2016-07-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US12112089B2 (en) | 2014-02-11 | 2024-10-08 | Mentor Acquisition One, Llc | Spatial location presentation in head worn computing |
WO2015125626A1 (ja) | 2014-02-20 | 2015-08-27 | ソニー株式会社 | 表示制御装置、表示制御方法およびコンピュータプログラム |
US20160187651A1 (en) | 2014-03-28 | 2016-06-30 | Osterhout Group, Inc. | Safety for a vehicle operator with an hmd |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
USD751552S1 (en) | 2014-12-31 | 2016-03-15 | Osterhout Group, Inc. | Computer glasses |
US20160239985A1 (en) | 2015-02-17 | 2016-08-18 | Osterhout Group, Inc. | See-through computer display systems |
US10878775B2 (en) | 2015-02-17 | 2020-12-29 | Mentor Acquisition One, Llc | See-through computer display systems |
US10204596B2 (en) * | 2015-12-21 | 2019-02-12 | Mediatek Inc. | Display control for transparent display |
US10591728B2 (en) | 2016-03-02 | 2020-03-17 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
US10667981B2 (en) | 2016-02-29 | 2020-06-02 | Mentor Acquisition One, Llc | Reading assistance system for visually impaired |
US10699691B1 (en) * | 2017-06-29 | 2020-06-30 | Amazon Technologies, Inc. | Active noise cancellation for bone conduction speaker of a head-mounted wearable device |
CN108717244B (zh) * | 2018-05-18 | 2021-03-26 | 京东方科技集团股份有限公司 | 显示装置及其控制方法、存储介质 |
CN110187855B (zh) * | 2019-05-28 | 2022-09-16 | 幻蝎科技(武汉)有限公司 | 近眼显示设备的避免全息影像阻挡视线的智能调节方法 |
CA3142786A1 (en) * | 2019-07-31 | 2021-02-04 | Xenon-Vr, Inc. | Ophthalmologic testing systems and methods |
JP7485528B2 (ja) * | 2020-03-27 | 2024-05-16 | 株式会社コロプラ | プログラム |
KR20220002800A (ko) * | 2020-06-30 | 2022-01-07 | 현대모비스 주식회사 | 운전자 부주의 경고 장치 및 방법 |
US11747622B2 (en) | 2021-01-04 | 2023-09-05 | Rovi Guides, Inc. | Methods and systems for controlling media content presentation on a smart glasses display |
WO2022146696A1 (en) * | 2021-01-04 | 2022-07-07 | Rovi Guides, Inc. | Methods and systems for controlling media content presentation on a smart glasses display |
US11906737B2 (en) | 2021-01-04 | 2024-02-20 | Rovi Guides, Inc. | Methods and systems for controlling media content presentation on a smart glasses display |
USD1019641S1 (en) | 2022-03-03 | 2024-03-26 | Xenon Ophthalmics Inc. | Headset |
USD1021898S1 (en) | 2022-03-03 | 2024-04-09 | Xenon Ophthalmics Inc. | Module for head mounted display |
USD1005288S1 (en) | 2022-03-03 | 2023-11-21 | Xenon Ophthalmics Inc. | Module for head mounted display |
USD1005289S1 (en) | 2022-03-03 | 2023-11-21 | Xenon Ophthalmics Inc. | Headset |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH09191438A (ja) * | 1996-01-09 | 1997-07-22 | Canon Inc | 視線検出機能を有する表示装置 |
JP2005138801A (ja) * | 2003-11-10 | 2005-06-02 | Calsonic Kansei Corp | ヘッドアップディスプレイ装置 |
WO2005122128A1 (ja) * | 2004-06-10 | 2005-12-22 | Matsushita Electric Industrial Co., Ltd. | 装着型情報提示装置 |
WO2012133379A1 (ja) * | 2011-03-31 | 2012-10-04 | ブラザー工業株式会社 | ヘッドマウントディスプレイ,及びその輝度調整方法 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5003300A (en) * | 1987-07-27 | 1991-03-26 | Reflection Technology, Inc. | Head mounted display for miniature video display system |
WO1996000406A1 (fr) * | 1994-06-23 | 1996-01-04 | Seiko Epson Corporation | Dispositif de presentation monte sur des lunettes |
US20020044152A1 (en) * | 2000-10-16 | 2002-04-18 | Abbott Kenneth H. | Dynamic integration of computer generated and real world images |
WO2004061519A1 (ja) * | 2002-12-24 | 2004-07-22 | Nikon Corporation | ヘッドマウントディスプレイ |
JP2004219664A (ja) * | 2003-01-14 | 2004-08-05 | Sumitomo Electric Ind Ltd | 情報表示システム及び情報表示方法 |
JP2004109995A (ja) * | 2003-08-12 | 2004-04-08 | Toshiba Corp | 装着型情報呈示装置および方法および記憶媒体 |
WO2006064655A1 (ja) * | 2004-12-14 | 2006-06-22 | Matsushita Electric Industrial Co., Ltd. | 情報提示装置及び情報提示方法 |
JP2006171311A (ja) * | 2004-12-15 | 2006-06-29 | Nec Corp | 表示部を備えた電子機器及びその制御方法 |
JP4977525B2 (ja) * | 2006-05-25 | 2012-07-18 | パナソニック株式会社 | 画像制御の装置と方法、及びそのプログラムとそれを記録した記録媒体 |
US8907866B2 (en) * | 2006-09-13 | 2014-12-09 | Nikon Corporation | Head mount display |
JP5228307B2 (ja) * | 2006-10-16 | 2013-07-03 | ソニー株式会社 | 表示装置、表示方法 |
JP2010164782A (ja) * | 2009-01-15 | 2010-07-29 | Brother Ind Ltd | 画像表示装置 |
US9111498B2 (en) * | 2010-08-25 | 2015-08-18 | Eastman Kodak Company | Head-mounted display with environmental state detection |
JP6051605B2 (ja) | 2012-06-13 | 2016-12-27 | ソニー株式会社 | 表示装置、および表示制御方法、並びにプログラム |
US9219901B2 (en) * | 2012-06-19 | 2015-12-22 | Qualcomm Incorporated | Reactive user interface for head-mounted display |
WO2015125626A1 (ja) * | 2014-02-20 | 2015-08-27 | ソニー株式会社 | 表示制御装置、表示制御方法およびコンピュータプログラム |
-
2015
- 2015-02-05 WO PCT/JP2015/053312 patent/WO2015125626A1/ja active Application Filing
- 2015-02-05 EP EP15752499.2A patent/EP3109854B1/en active Active
- 2015-02-05 JP JP2016504031A patent/JP6743691B2/ja active Active
- 2015-02-05 CN CN201580008449.5A patent/CN106030692B/zh active Active
- 2015-02-05 US US15/101,739 patent/US10146057B2/en active Active
-
2018
- 2018-11-08 US US16/184,409 patent/US10324294B2/en active Active
-
2019
- 2019-05-15 US US16/413,276 patent/US10437069B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH09191438A (ja) * | 1996-01-09 | 1997-07-22 | Canon Inc | 視線検出機能を有する表示装置 |
JP2005138801A (ja) * | 2003-11-10 | 2005-06-02 | Calsonic Kansei Corp | ヘッドアップディスプレイ装置 |
WO2005122128A1 (ja) * | 2004-06-10 | 2005-12-22 | Matsushita Electric Industrial Co., Ltd. | 装着型情報提示装置 |
WO2012133379A1 (ja) * | 2011-03-31 | 2012-10-04 | ブラザー工業株式会社 | ヘッドマウントディスプレイ,及びその輝度調整方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3109854A4 * |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017516196A (ja) * | 2014-04-17 | 2017-06-15 | マイクロソフト テクノロジー ライセンシング,エルエルシー | 会話検出 |
US10529359B2 (en) | 2014-04-17 | 2020-01-07 | Microsoft Technology Licensing, Llc | Conversation detection |
US10679648B2 (en) | 2014-04-17 | 2020-06-09 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
US11176797B2 (en) | 2015-09-01 | 2021-11-16 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US11741811B2 (en) | 2015-09-01 | 2023-08-29 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
US10755545B2 (en) | 2015-09-01 | 2020-08-25 | Kabushiki Kaisha Toshiba | Electronic apparatus and method |
JP2020113298A (ja) * | 2015-09-16 | 2020-07-27 | グーグル エルエルシー | 拡張および/または仮想現実環境におけるタッチスクリーンホバーの検出 |
JP2018534643A (ja) * | 2015-09-16 | 2018-11-22 | グーグル エルエルシー | 拡張および/または仮想現実環境におけるタッチスクリーンホバーの検出 |
WO2017081915A1 (ja) * | 2015-11-10 | 2017-05-18 | ソニー株式会社 | 画像処理装置、画像処理方法およびプログラム |
WO2017094557A1 (ja) * | 2015-12-02 | 2017-06-08 | コニカミノルタ株式会社 | 電子機器及びヘッドマウントディスプレイ |
WO2018198503A1 (ja) * | 2017-04-26 | 2018-11-01 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
JP2020520494A (ja) * | 2017-05-04 | 2020-07-09 | ソニー インタラクティブ エンタテインメント ヨーロッパ リミテッド | ヘッドマウントディスプレイおよび方法 |
JP7191853B2 (ja) | 2017-05-04 | 2022-12-19 | 株式会社ソニー・インタラクティブエンタテインメント | ヘッドマウントディスプレイおよび方法 |
JP2018195321A (ja) * | 2018-07-03 | 2018-12-06 | 株式会社東芝 | ウェアラブル端末及び方法 |
WO2020234939A1 (ja) * | 2019-05-17 | 2020-11-26 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
US11846783B2 (en) | 2019-05-17 | 2023-12-19 | Sony Group Corporation | Information processing apparatus, information processing method, and program |
JP2021120901A (ja) * | 2019-07-03 | 2021-08-19 | 株式会社東芝 | ウェアラブル端末、システム及び方法 |
JP2019197564A (ja) * | 2019-07-03 | 2019-11-14 | 株式会社東芝 | ウェアラブル端末、システム及び方法 |
JP2022038495A (ja) * | 2020-08-26 | 2022-03-10 | ソフトバンク株式会社 | 表示制御装置、プログラム、及びシステム |
JP2022038783A (ja) * | 2020-08-27 | 2022-03-10 | Necパーソナルコンピュータ株式会社 | 情報処理装置及びその通知制御方法並びに通知制御プログラム |
JP7216053B2 (ja) | 2020-08-27 | 2023-01-31 | Necパーソナルコンピュータ株式会社 | 情報処理装置及びその通知制御方法並びに通知制御プログラム |
Also Published As
Publication number | Publication date |
---|---|
US20190072769A1 (en) | 2019-03-07 |
JPWO2015125626A1 (ja) | 2017-03-30 |
US10146057B2 (en) | 2018-12-04 |
EP3109854A4 (en) | 2017-07-26 |
CN106030692A (zh) | 2016-10-12 |
JP6743691B2 (ja) | 2020-08-19 |
EP3109854A1 (en) | 2016-12-28 |
US20190265485A1 (en) | 2019-08-29 |
US10437069B2 (en) | 2019-10-08 |
US20160306173A1 (en) | 2016-10-20 |
EP3109854B1 (en) | 2024-09-18 |
US10324294B2 (en) | 2019-06-18 |
CN106030692B (zh) | 2019-11-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2015125626A1 (ja) | 表示制御装置、表示制御方法およびコンピュータプログラム | |
JP5309448B2 (ja) | 表示装置、表示方法 | |
US9733701B2 (en) | Display device and display method that determines intention or status of a user | |
US10142618B2 (en) | Imaging apparatus and imaging method | |
US20080062291A1 (en) | Image pickup apparatus and image pickup method | |
JP6094190B2 (ja) | 情報処理装置および記録媒体 | |
JP3954028B2 (ja) | 装着型情報告知装置 | |
CN106454644B (zh) | 音频播放方法及装置 | |
JP2013210643A (ja) | 表示装置、表示方法 | |
US20170083282A1 (en) | Information processing device, control method, and program | |
CN109885368A (zh) | 一种界面显示防抖方法及移动终端 | |
JP6160654B2 (ja) | 表示装置、表示方法、プログラム | |
US20240289088A1 (en) | Volume Adjustment Method, Electronic Device, and System | |
US11037519B2 (en) | Display device having display based on detection value, program, and method of controlling device | |
JP2008288821A (ja) | 撮像装置、撮像方法 | |
EP4383743A1 (en) | Video processing method and terminal | |
JP2019161483A (ja) | 携帯端末および画像切り出し方法 | |
JP2009055080A (ja) | 撮像装置、撮像方法 | |
JP2008129045A (ja) | 映像表示装置 | |
KR101871660B1 (ko) | 웨어러블 기반의 개인용 자동 기록장치의 운용방법 | |
CN114760575A (zh) | 一种助听器个性化调节方法及装置 | |
JP2022167186A (ja) | 撮像装置 | |
JP2021072580A (ja) | 撮像装置、警告制御方法及びコンピュータプログラム | |
JP2015139169A (ja) | 電子機器及び撮像装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15752499 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15101739 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2016504031 Country of ref document: JP Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2015752499 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2015752499 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |