WO2006051743A1 - 映像表示装置 - Google Patents
映像表示装置 Download PDFInfo
- Publication number
- WO2006051743A1 WO2006051743A1 PCT/JP2005/020332 JP2005020332W WO2006051743A1 WO 2006051743 A1 WO2006051743 A1 WO 2006051743A1 JP 2005020332 W JP2005020332 W JP 2005020332W WO 2006051743 A1 WO2006051743 A1 WO 2006051743A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- display device
- wearer
- video display
- function
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
- H04N5/7475—Constructional details of television projection apparatus
- H04N5/7491—Constructional details of television projection apparatus of head mounted projectors
Definitions
- the present invention is an image display device that displays an image on the wearer's eyes while being worn on the head (usually referred to as a head-mounted display, and is sometimes referred to as a head-mounted display in this description). Sekida
- Head-mounted displays are often configured to be mounted on the head in the form of being wound around the head, with a two-eye type in which an image display system is formed at a position corresponding to both eyes and a position corresponding to one eye of the left and right eyes.
- the two-eye type is mainly used for the purpose of enjoying video.
- One-eye type devices are expected to be used as display devices such as wearable personal computers and display devices that display instructions to workers.
- the methods described in Japanese Patent Application Laid-Open Nos. 8-3005 2 98 and 2 0 4 2 0 7 8 4 7 are known.
- One problem with head mount displays is that the edges of the screen are difficult to see. In normal movie theaters, when looking at the edge of the screen projected on the screen, humans shake their eyes by eye movements, and at the same time, turn their necks to change the direction of their eyes. Pointed to the edge of the screen.
- Japanese Patent Application Laid-Open No. 7-95 4 98 discloses a method for detecting the movement of the head and moving the display image accordingly.
- this method has a problem that the device becomes complicated and a part of the screen is cut off when the display image is moved.
- Japanese Patent Application Laid-Open No. 2 00 0-1 7 1 7 5 1 discloses a method of displaying a screen in a part of the field of view and moving the display position of the screen according to the movement of the head. Yes.
- the screen is displayed only in a part of the area that can be displayed originally, so the displayed elements (icons, etc.) must also be made smaller than the size that can be displayed.
- Disclosure of the invention has been made in view of such circumstances, and provides an image display device having a function of performing training so that an end portion of a displayed image can be seen only by eye movements. Let it be an issue.
- a first means for solving the above-mentioned problem is a video display device that displays a video image to the wearer's eye while being worn on the head, and a part or all of the displayed screen is reduced and displayed. And a function of changing the reduction ratio according to the time or number of times the wearer has used the video display device.
- the reduction ratio is increased and the screen is displayed small in the center of the field of view, and the reduction rate is gradually reduced as the usage time and the number of times of use increase.
- the image can be concentrated near the center of the field of view before it can be used, and it can grow to reach the edge of the field of view. Therefore, the wearer can naturally see the edge of the image with only eye movement, and the feeling of discomfort is reduced.
- a second means for solving the above problem is a video display device that displays a video image to the wearer's eye while being worn on the head, and a part of an image constituting the displayed screen is displayed on the screen.
- a video display device characterized in that it has a function of moving to the center side of the image and displaying it, and a function of changing the display position according to the time or number of times the wearer has used the video display device. .
- the third means for solving the problem is the second means, wherein a part of the image is displayed at a position closer to the center of the screen than the position originally displayed, As the user's usage time or number of uses increases, it moves outside the screen, and after a predetermined time or a predetermined number of uses, it has a function of displaying it at a position where it is originally displayed. In this means, a part of the image is displayed at a position closer to the center of the screen than originally displayed. For “part of the image”, it is preferable to select something that the wearer needs to pay attention to, such as an icon. Therefore, unless the wearer is accustomed, a part of these images is displayed at a position closer to the center of the screen than the original display position, so the wearer has less chance to see the peripheral portion of the image. I do n’t feel uncomfortable.
- a fourth means for solving the above-mentioned problem is the second means, wherein a part of the image is set to one end point at a position where it is originally displayed, and the screen is more than a position where it is originally displayed. Reciprocating between the position and the position near the center of the body, and the amplitude of the back-and-forth movement is reduced as the wearer's usage time or number of uses increases.
- a part of the image is reciprocated between a position where the image is originally displayed as one end point and a position closer to the center of the screen than the position where the image is originally displayed. Therefore, the wearer becomes accustomed to seeing the peripheral part of the screen naturally only by eye movement by following a part of these images.
- the reciprocating motion position can be limited to the peripheral edge of the screen according to the wearer's familiarity. .
- the fifth means for solving the problem is any one of the first means to the fourth means, wherein the use time or the number of use is stored for each wearer, Each wearer uses the stored use time or use count.
- the first to fourth means When a single image display device is used by multiple people, simply operating the first to fourth means according to the usage time and number of times of use of the image display device will make it easier for the wearer to get used to it. Cannot display according to.
- the usage time or the number of times of use is stored for each wearer, and the operation of the first means to the fourth means is performed using the stored use time or the number of times of use for each wearer.
- a sixth means for solving the above-described problem is a video display device that displays a video image to the wearer's eye while being worn on the head, and a part or all of the displayed screen is reduced and displayed. And a function of detecting the wearer's line of sight, and a function of changing the reduction ratio in accordance with the detected direction distribution of the line of sight of the wearer.
- a seventh means for solving the above-mentioned problem is a video display device that displays a video image to the wearer's eye while being worn on the head, and a part of an image constituting the displayed screen is displayed on the screen. It has a function of moving to the center side of the display and a function of detecting the wearer's line of sight, and a function of changing the display position according to the detected direction distribution of the wearer's line of sight. It is a video display device.
- the eighth means for solving the above-mentioned problem is the seventh means, and when the direction distribution of the line of sight of the wearer is concentrated near the center of the image, It moves to the center of the screen from the displayed position, and as the proportion of the wearer's line-of-sight direction increases outside the image, it moves to the outside of the screen and the wearer's line-of-sight direction distribution Of outside It has a function to display at the position where it is originally displayed when the value exceeds a predetermined value.
- Ninth means for solving the problem is the seventh means, wherein a part of the image, a part of the image, and a position where the image is originally displayed as one end point, Move back and forth between a position closer to the center of the screen than the originally displayed position, and change the amplitude of the reciprocating movement according to the ratio of the wearer's gaze direction distribution to the outside of the image. It is characterized by this.
- the image reduction ratio and the display position of a part of the image are changed in accordance with the use time or the use frequency in the first to fourth means.
- the difference is that the direction of the gaze of the wearer is changed.
- the wearer's gaze direction distribution is biased toward the center of the field of view, and as you get used to it, the gaze direction distribution reaches the periphery of the field of view. It is averaged as follows. Therefore, using this, when the directional distribution of the line of sight reaches the periphery and is averaged, the use time or the number of use increases in the first to fourth means. Control is performed as is done.
- the tenth means for solving the problem is the first means, and has a function of detecting the wearer's line of sight, and the use time or the number of use of the video display device, and the line-of-sight detection result It is characterized by having a function to change the reduction ratio according to the combination.
- the first means for solving the problem is the second means, which has a function of detecting the wearer's line of sight, and the time or number of times the wearer has used the video display device. It has a function of changing the display position in accordance with the combination of the line-of-sight detection results.
- the first and second means for solving the problem is the first means.
- a part of the image is displayed at a position closer to the center of the screen than originally displayed, and depending on the combination of the wearer's use time or number of times of use and the gaze detection result, It has the function of moving to the outside of the screen and finally displaying it at the position where it is displayed.
- a first means for solving the above-mentioned problem is the first means, wherein a part of the image is set to one end point from a position where the image is originally displayed. Is moved back and forth between the position close to the center of the screen and the amplitude of the reciprocation is changed according to the time or number of times the wearer used the video display device and the combination of the visual line detection results. It is characterized by this.
- the 14th means for solving the above-mentioned problem is any one of the above 1st 0 means to the 1st 3 means, and any one of claims 10 to 1 3
- the image reduction ratio and the display position of a part of the image in the first to fifth means are set according to the use time or the number of use. In addition to this, it is different in that it is changed according to the direction distribution of the wearer's line of sight. Therefore, the first and fifth means to the fifth means and the sixth to ninth means are appropriately combined to provide an operational effect.
- FIG. 1 is a block diagram showing a schematic configuration of a head mount display as an example of an embodiment of the present invention.
- FIG. 2 is a diagram illustrating a first operation of the image control unit.
- FIG. 3 is a diagram illustrating a second operation of the image control unit.
- FIG. 4 is a diagram illustrating a third operation of the image control unit. BEST MODE FOR CARRYING OUT THE INVENTION
- FIG. 1 is a block diagram showing a schematic configuration of a head mount display as an example of an embodiment of the present invention.
- This head-mounted display is composed of a head mounting unit 1 mounted on the head, a control unit 2, an operation unit 3, and a line-of-sight detection unit 4.
- the head mounting unit 1 is provided with an image display unit 5 and an audio output unit 6.
- the control unit 2 receives the wearer's signal input from the operation unit 3, and the number of uses for each wearer.
- An input control unit 7 having a function of storing a usage time, receiving a signal from the line-of-sight detection unit 4, and calculating a direction distribution of the line of sight of the wearer is provided.
- the input control unit 7 has a function of processing an operation signal input from the operation unit 3, but the function is not related to the present invention and is omitted.
- control unit 2 is provided with a storage unit 8 that stores an image to be displayed and a sound to be output with respect to the head mounting unit 1.
- the display image stored in the storage unit 8 is converted into an image that is actually displayed by the image processing program in the image control unit 9, transmitted to the image display unit 5, and displayed.
- the audio signal stored in the storage unit 8 is adjusted by the audio control unit 10, transmitted to the audio output unit 6, and output as an audio signal.
- FIG. 2 is a diagram illustrating a first operation of the image control unit 9.
- 1 1 is originally displayed
- the image to be displayed is shown and is displayed over the entire display area surrounded by the square frame 12.
- the image control unit 9 has a function of reducing and displaying the entire image with the center of the image as the center.
- the scaled-down thumbnail image is shown in 13.
- the image control unit 9 displays a reduced image 13 at the beginning of use of the head mount display, and the number of times or use of the head mount display corresponding to the wearer input from the input control unit 7
- the time (determined which is selected is determined by the switch of the operation unit 3) increases, and the reduction ratio of the reduced image 13 is reduced, the image is enlarged, and the predetermined number of times or the predetermined usage time is reached. After that, the image 1 1 that should be displayed should be displayed.
- the input control unit 7 increments the usage count by 1 each time the head mounted display is turned on, and adds the usage time while the head mounted display is turned on.
- the number of times used and the use time are calculated for each wearer code input from the operation unit 3, and the use corresponding to each wearer is used. The above control is performed using the number of times and the usage time.
- FIG. 3 is a diagram illustrating a second operation of the image control unit 9.
- 1 4 indicates an image that should be displayed, and the image is displayed over the entire display area surrounded by the square frame 1 2.
- the display image includes a background 15 and an icon 16.
- the image control unit 9 has a function of moving the position of the icon 16 that is a part of the display image from the position that should be originally displayed to the center of the image.
- the actual image displayed at the beginning of use of the head-mounted display is as shown in 17 and the background 15 is unchanged, but the position of the icon 16 shifts to the center of the image. ing.
- the image control unit 9 displays the initial image 1 7 in the initial use of the head mount display, and the number of times the head mounted display corresponding to the wearer is used or input from the input control unit 7.
- the time (determined which is selected is determined by the switch of the operation unit 3) increases, and the position of the icon 16 is sequentially shifted to the outside of the screen to reach a predetermined number of times or a predetermined use time. After that, it returns to the position where it should be displayed, and the image 14 that should be displayed is displayed.
- the wearer initially focuses on the icon 16 in the center of the field of view, and expands the field of view by eye movement by looking at the icon 16 that moves outward over time. Can be trained. Therefore, the peripheral part of the image display area can be seen naturally only by eye movement.
- the position where the icon 16 is originally displayed is fixed as one end, and the other end 16 ′ of the reciprocating motion is input from the input control unit 7.
- the number of times or time of use of the head mount display corresponding to the user increases (determining which to select is determined by the switch of operation unit 3), and shifts to the outside of the screen. Predetermined number of times or place After reaching the fixed usage time, return to the original display position and display the original display image 14.
- the wearer first looks at the icon 16 when it reaches 16 6 ', the center of the field of view, but as time passes, the position where the icon is displayed moves outward. Therefore, you can receive training to widen your field of vision through eye movements. Therefore, it becomes possible to naturally see the peripheral portion of the image display area only by eye movement.
- the reduction and movement of the image are controlled according to the number of use times or the use time.
- the reduction and movement of the image may be controlled based on the gaze direction distribution.
- the line-of-sight detection device for example, a known device as described in JP-A-7-3208 can be used.
- the standard deviation of the distribution of the line-of-sight angle centered on the center direction of the screen within a predetermined time may be taken, and this standard deviation may be used instead of the above-mentioned number of uses and use time.
- the display when the standard deviation is small, the display is the same as when the number of uses and the usage time is small, and as the standard deviation increases, the display should be the same as when the usage count and the usage time have increased. That's fine.
- the ratio of the reduced graphic display in the example shown above may be determined by a combination of the number of times of use, the time of use, and the line-of-sight detection result. For example, based on the square root of the sum of squares of the standard deviation of usage time and eye movement The ratio of the reduced display of the figure may be determined based on the square root of the sum of squares of the number of uses and the standard deviation of the line-of-sight variation.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Controls And Circuits For Display Device (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP05799987.2A EP1811493B1 (en) | 2004-11-12 | 2005-10-31 | Video display |
US11/667,279 US20070262917A1 (en) | 2004-11-12 | 2005-10-31 | Video Display |
JP2006544874A JP4925831B2 (ja) | 2004-11-12 | 2005-10-31 | 映像表示装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004-329135 | 2004-11-12 | ||
JP2004329135 | 2004-11-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006051743A1 true WO2006051743A1 (ja) | 2006-05-18 |
Family
ID=36336424
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/020332 WO2006051743A1 (ja) | 2004-11-12 | 2005-10-31 | 映像表示装置 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20070262917A1 (ja) |
EP (1) | EP1811493B1 (ja) |
JP (1) | JP4925831B2 (ja) |
WO (1) | WO2006051743A1 (ja) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5828070B2 (ja) * | 2010-08-20 | 2015-12-02 | パナソニックIpマネジメント株式会社 | 撮像装置および撮像方法 |
GB201310364D0 (en) | 2013-06-11 | 2013-07-24 | Sony Comp Entertainment Europe | Head-mountable apparatus and systems |
EP3057089A4 (en) * | 2014-01-20 | 2017-05-17 | Sony Corporation | Image display device and image display method, image output device and image output method, and image display system |
CN107424583B (zh) * | 2017-07-03 | 2020-05-12 | 威创集团股份有限公司 | 异形图像的显示数据处理方法和系统 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10221639A (ja) * | 1996-12-03 | 1998-08-21 | Sony Corp | 表示装置および表示方法 |
JP2000171751A (ja) * | 1998-12-07 | 2000-06-23 | Hitachi Ltd | 双眼式ディスプレイ |
JP2003076353A (ja) * | 2001-09-04 | 2003-03-14 | Sharp Corp | ヘッドマウント型表示装置 |
JP2003348622A (ja) * | 2002-05-24 | 2003-12-05 | Canon Inc | 立体映像表示方法及び記憶媒体 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07294844A (ja) * | 1994-04-22 | 1995-11-10 | Canon Inc | 表示装置 |
US6388638B2 (en) * | 1994-10-28 | 2002-05-14 | Canon Kabushiki Kaisha | Display apparatus and its control method |
US6127990A (en) * | 1995-11-28 | 2000-10-03 | Vega Vista, Inc. | Wearable display and methods for controlling same |
JP3679506B2 (ja) * | 1996-06-03 | 2005-08-03 | キヤノン株式会社 | 表示装置 |
US6069608A (en) * | 1996-12-03 | 2000-05-30 | Sony Corporation | Display device having perception image for improving depth perception of a virtual image |
US6445363B1 (en) * | 1999-02-12 | 2002-09-03 | Canon Kabushiki Kaisha | Head-mounted display apparatus |
US7002551B2 (en) * | 2002-09-25 | 2006-02-21 | Hrl Laboratories, Llc | Optical see-through augmented reality modified-scale display |
JP3988632B2 (ja) * | 2002-11-28 | 2007-10-10 | 日本電気株式会社 | 眼鏡型ディスプレイの制御方法 |
WO2004061519A1 (ja) * | 2002-12-24 | 2004-07-22 | Nikon Corporation | ヘッドマウントディスプレイ |
JP2004236241A (ja) * | 2003-01-31 | 2004-08-19 | Nikon Corp | 表示装置 |
JP4490074B2 (ja) * | 2003-04-17 | 2010-06-23 | ソニー株式会社 | 立体視画像処理装置、立体視画像表示装置、立体視画像提供方法、および立体視画像処理システム |
JP4174420B2 (ja) * | 2003-12-15 | 2008-10-29 | キヤノン株式会社 | 光偏向器 |
US20060103591A1 (en) * | 2004-11-15 | 2006-05-18 | Canon Kabushiki Kaisha | Information processing apparatus and method for providing observer with information |
JP4916145B2 (ja) * | 2005-08-08 | 2012-04-11 | キヤノン株式会社 | 情報処理装置及び情報処理方法並びにプログラム |
JP4626668B2 (ja) * | 2008-04-03 | 2011-02-09 | ソニー株式会社 | 画像処理装置、表示制御方法、プログラム及び記録媒体 |
JP2010164782A (ja) * | 2009-01-15 | 2010-07-29 | Brother Ind Ltd | 画像表示装置 |
-
2005
- 2005-10-31 WO PCT/JP2005/020332 patent/WO2006051743A1/ja active Application Filing
- 2005-10-31 JP JP2006544874A patent/JP4925831B2/ja not_active Expired - Fee Related
- 2005-10-31 EP EP05799987.2A patent/EP1811493B1/en not_active Expired - Fee Related
- 2005-10-31 US US11/667,279 patent/US20070262917A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10221639A (ja) * | 1996-12-03 | 1998-08-21 | Sony Corp | 表示装置および表示方法 |
JP2000171751A (ja) * | 1998-12-07 | 2000-06-23 | Hitachi Ltd | 双眼式ディスプレイ |
JP2003076353A (ja) * | 2001-09-04 | 2003-03-14 | Sharp Corp | ヘッドマウント型表示装置 |
JP2003348622A (ja) * | 2002-05-24 | 2003-12-05 | Canon Inc | 立体映像表示方法及び記憶媒体 |
Also Published As
Publication number | Publication date |
---|---|
EP1811493A4 (en) | 2010-09-01 |
JP4925831B2 (ja) | 2012-05-09 |
EP1811493B1 (en) | 2013-04-24 |
EP1811493A1 (en) | 2007-07-25 |
JPWO2006051743A1 (ja) | 2008-05-29 |
US20070262917A1 (en) | 2007-11-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11989930B2 (en) | UI for head mounted display system | |
JP4927631B2 (ja) | 表示装置、その制御方法、プログラム、記録媒体および集積回路 | |
US8619005B2 (en) | Switchable head-mounted display transition | |
JP2020515895A (ja) | 操作可能な中心窩ディスプレイ | |
US20230164286A1 (en) | Methods and Apparatuses for Compensating for Retinitis Pigmentosa | |
JP6112878B2 (ja) | ウェアラブル型表示装置及びプログラム | |
WO2017183279A1 (ja) | 画像処理装置および画像処理方法 | |
WO2005122128A1 (ja) | 装着型情報提示装置 | |
JP2006267604A (ja) | 複合情報表示装置 | |
CN111886564A (zh) | 信息处理装置、信息处理方法和程序 | |
CN110082910A (zh) | 用于在显示镜上展示表情符号的方法和设备 | |
JP4925831B2 (ja) | 映像表示装置 | |
JP6160654B2 (ja) | 表示装置、表示方法、プログラム | |
JP6341759B2 (ja) | 頭部装着型情報表示装置及び頭部装着型情報表示装置の制御方法 | |
JP2003334221A (ja) | 画像表示装置および眼疲労解消装置 | |
US11051005B2 (en) | Virtual reality device and operation method of virtual reality device | |
JP2009020180A (ja) | 映像表示装置 | |
KR102029201B1 (ko) | 멀미가 감소된 가상현실 게임 제공 방법 및 가상현실 게임 단말기 | |
KR100846355B1 (ko) | 두부 장착식 디스플레이 장치의 시력 보조 방법 및 그 두부장착식 디스플레이 장치 | |
JP5343304B2 (ja) | ウェアラブルディスプレイ | |
JP2006163009A (ja) | 映像表示方法 | |
JP2000171751A (ja) | 双眼式ディスプレイ | |
JPH07287187A (ja) | 映像表示装置 | |
KR100662248B1 (ko) | 주시거리 조절기능을 가지는 입체영상 디스플레이 시스템 | |
US20240104860A1 (en) | Devices, Methods, and Graphical User Interfaces for Interacting with Three-Dimensional Environments |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KN KP KR KZ LC LK LR LS LT LU LV LY MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006544874 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005799987 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11667279 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2005799987 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 11667279 Country of ref document: US |