US20070262917A1 - Video Display - Google Patents
Video Display Download PDFInfo
- Publication number
- US20070262917A1 US20070262917A1 US11/667,279 US66727905A US2007262917A1 US 20070262917 A1 US20070262917 A1 US 20070262917A1 US 66727905 A US66727905 A US 66727905A US 2007262917 A1 US2007262917 A1 US 2007262917A1
- Authority
- US
- United States
- Prior art keywords
- wearer
- image
- screen
- function
- display device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
- H04N5/7475—Constructional details of television projection apparatus
- H04N5/7491—Constructional details of television projection apparatus of head mounted projectors
Definitions
- the present invention relates to a video display device which displays video images to the eyes of the wearer in a state in which the device is mounted on the head of the wearer (often called a “head-mounted display” and may be referred to as a “head-mounted display” in the present specification as well).
- ocular type devices in which video images displayed on a display device such as a liquid crystal panel (LCD) are observed as false images that are enlarged via an optical system having an ocular lens, half-mirror, or the like; such display devices are called “wearable displays.”
- LCD liquid crystal panel
- wearable displays devices that are used while mounted on the head are particularly referred to as “head-mounted displays.”
- head-mounted displays are constructed so as to be mounted on the head in a form that is wrapped around the head.
- Such displays include displays of the both-eye type in which video display systems are formed in positions corresponding to both eyes, and displays of the one-eye type in which a video display system is formed in a position corresponding to one eye (left or right eye).
- displays of the both-eye type are used mainly for the purpose of enjoying videos.
- displays of the one-eye type for example, methods involving use as the display device of a wearable personal computer or the like, or use as a display device which displays instructions to workers, may be expected.
- Publicly known examples include systems such as those described in Japanese Laid-Open Patent Application No. H8-305298 and Japanese Laid-Open Patent Application No. 2004-207847.
- edge parts of the screen may be difficult to see.
- the person viewing the image moves the line of sight by moving the eyeballs, and also at the same time, directs the line of sight toward the edge parts of the screen by turning the head.
- the present invention was devised in light of such circumstances; it is an object of the present invention to provide a video display device which has the function of performing training so that the edge parts of the image that is displayed can be viewed by moving the eyeballs alone.
- the first means used to solve the problems described above is a video display device which displays video images to the eyes of the wearer in a state in which the device is mounted on the head of the wearer, wherein this device has the function of reducing and displaying all or part of the screen that is displayed, and the function of altering the reduction ratio in accordance with the time or frequency of use of the video display device by the wearer.
- the reduction ratio is set at a large ratio, so that the screen is displayed with a small size in the center of the visual field. Then, as the use time or use frequency increases, the reduction ratio is gradually reduced, so that an ordinary screen is finally shown; thus, while the wearer is still not accustomed to use of the device, the image is concentrated near the center of the visual field, and as the wearer becomes accustomed to the device, the size of the image can be increased so that the image reaches the peripheral parts of the visual field.
- the wearer naturally becomes able to view the edge parts of the image merely by moving the eyeballs, and becomes less likely to experience any uncomfortable feeling.
- the second means used to solve the problems described above is a video display device which displays video images to the eyes of the wearer in a state in which the device is mounted on the head of the wearer, wherein this device has the function of moving a portion of the image constituting the screen that is displayed toward the center of the screen and displaying this portion of the image, and the function of altering the display position in accordance with the time or frequency of use of the video display device by the wearer.
- the third means used to solve the problems described above is the second means, wherein the device has the function of initially displaying a portion of the image in a position that is closer to the center of the screen than the inherent display position, moving this portion of the image toward the outside of the screen as the use time or use frequency by the wearer increases, and displaying this portion of the image in the inherent display position following a specified time or specified number of times of use.
- a portion of the image is initially displayed in a position that is closer to the center of the screen than the inherent display position. It is desirable that elements such as icons to which the wearer must pay attention be selected as this “portion of the image.” Accordingly, while the wearer is still not accustomed to the device, these portions of the image are displayed in a position that is closer to the center of the screen than the inherent display position, so that the wearer has little occasion to view the peripheral portions of the image, and experiences no uncomfortable feeling.
- the fourth means used to solve the problems described above is the second means, wherein the inherent display position is taken as one end point, a portion of the image is caused to undergo a reciprocating movement between this end point and a position that is closer to the center of the screen than the inherent display position, and the amplitude of this reciprocating movement is reduced in accordance with an increase in the use time or use frequency by the wearer.
- the inherent display position is taken as one end point, and a portion of the image is caused to undergo a reciprocating motion between this end point and a position that is closer to the center of the screen than the inherent display position. Accordingly, the wearer follows these portions of the image with his eyes, and thus naturally becomes accustomed to viewing the peripheral portions of the screen by the movement of the eyeballs alone.
- the system is devised so that the amplitude of the reciprocating movement is reduced as the use time or use frequency increases; therefore, as the wearer becomes accustomed to the device, the position of the reciprocating movement can be limited to the peripheral portions of the screen.
- the fifth means used to solve the problems described above is any of the first through fourth means, wherein the use time or use frequency is stored in memory for each wearer, and the stored use time or use frequency is used for each wearer.
- the use time or use frequency is stored in memory for each wearer, and the operation of the first through fourth means is performed using the stored use time or use frequency for each wearer. Accordingly, even in cases where a single video display device is used by a plurality of persons, separate displays can be performed in accordance with the extent to which these respective persons have become accustomed to the device.
- the sixth means used to solve the problems described above is a video display device which displays video images to the eyes of the wearer in a state in which the device is mounted on the head of the wearer, wherein the device has the function of reducing and displaying all or part of the screen that is displayed, the function of detecting the line of sight of the wearer, and the function of altering the reduction ratio in accordance with the directional distribution of the detected line of sight of the wearer.
- the seventh means used to solve the problems described above is a video display device which displays video images to the eyes of the wearer in a state in which the device is mounted on the head of the wearer, wherein the device has the function of moving a portion of the image constituting the screen that is displayed toward the center of the screen and displaying this portion of the image, the function of detecting the line of sight of the wearer, and the function of altering the display position in accordance with the directional distribution of the detected line of sight of the wearer.
- the eighth means used to solve the problems described above is the seventh means, wherein the device has the following function: namely, when the directional distribution of the line of sight of the wearer is concentrated near the center of the image, a portion of the image is moved closer to the central portion of the screen than the inherent display position, and as the extent to which the directional distribution of the line of sight of the wearer reaches the outer portions of the image increases, this portion of the image is moved toward the outside of the screen, so that this portion of the image is displayed in the inherent display position when the extent to which the directional distribution of the line of sight of the wearer reaches the outer portions of the image is equal to or greater than a specified value.
- the ninth means used to solve the problems described above is the seventh means, wherein the inherent display position is taken as one end point, a portion of the image is caused to undergo a reciprocating movement between this end point and a position that is closer to the center of the screen than the inherent display position, and the amplitude of this reciprocating movement is varied in accordance with the extent to which the directional distribution of the line of sight of the wearer reaches the outer portions of the image.
- the device differs as follows: namely, in contrast to the first through fourth means in which the reduction ratio of the image and the display position of a portion of the image are varied in accordance with the use time or use frequency, these values are varied in accordance with the directional distribution of the line of sight of the wearer.
- the directional distribution of the line of sight of the wearer is biased toward the central portion of the visual field.
- the directional distribution of the line of sight begins to reach the peripheral portions of the visual field, and is evened out. Accordingly, by utilizing this, as the directional distribution of the line of sight reaches the peripheral portions and is evened out, control is performed such as that performed in the first through fourth means in cases where the use time or use frequency increases.
- the tenth means used to solve the problems described above is the first means, wherein the device has the function of detecting the line of sight of the wearer, and the function of altering the reduction ratio in accordance with a combination of the use time or use frequency of the video display device and the line of sight detection results.
- the eleventh means used to solve the problems described above is the second means, wherein the device has the function of detecting the line of sight of the wearer, and the function of altering the display position in accordance with a combination of the time or frequency of use of the video display device by the wearer and the line of sight detection results.
- the twelfth means used to solve the problems described above is the eleventh means, wherein the device has the following function: namely, a portion of the image is initially displayed in a position that is closer to the center of the screen than the inherent display position, and this portion of the image is moved toward the outside of the screen in accordance with a combination of an increase in the use time or use frequency by the wearer and the line of sight detection results, so that this portion of the image is finally displayed in the inherent display position.
- the thirteenth means used to solve the problems described above is the eleventh means, wherein the inherent display position is taken as one end point, a portion of the image is caused to undergo a reciprocating movement between this end point and a position that is closer to the center of the screen than the inherent display position, and the amplitude of this reciprocating movement is varied in accordance with a combination of the time or frequency of use of the video display device by the wearer and the line of sight detection results.
- the fourteenth means used to solve the problems described above is any of the tenth through thirteenth means, wherein the use time or use frequency is stored in memory for each wearer, and the stored use time or use frequency is used for each wearer.
- the device differs as follows: namely, in contrast to the first through fifth means in which the reduction ratio of the image and the display position of a portion of the image are varied in accordance with the use time or use frequency, these values are also varied in accordance with the directional distribution of the line of sight of the wearer. Accordingly, an action and effect are obtained which appropriately combine the first through fifth means and the sixth through ninth means.
- FIG. 1 is a block diagram showing the schematic construction of a head-mounted display constituting one working configuration of the present invention.
- FIG. 2 is a diagram showing the first operation of the image control part.
- FIG. 3 is a diagram showing the second operation of the image control part.
- FIG. 4 is a diagram showing the third operation of the image control part.
- FIG. 1 is a block diagram showing the schematic construction of a head-mounted display constituting one working configuration of the present invention.
- This head-mounted display comprises a head mounting part 1 which is mounted on the head, a control part 2 , an operating part 3 , and a line of sight detection part 4 .
- the control part 2 is provided with an input control part 7 which has the function of receiving signals from the wearer that are input via the operating part 3 , and storing the use frequency or use time of respective wearers, and the function of receiving signals from the line of sight detection part 4 , and calculating the directional distribution of the line of sight of the wearer.
- the input control part 7 also has the function of performing the processing of operating signals that are input from the operating part 3 ; however, since this function has no direct bearing on the present invention, a description of this function is omitted here.
- the control part 2 is also provided with a memory part 8 which stores images that are to be displayed and audio that is to be output to the head mounting part 1 .
- the display images that are stored in the memory part 8 are converted by an image processing program in the image control part 9 into the images that are actually displayed, and are transmitted to the image display part 5 and displayed.
- the audio signals that are stored in the memory part 8 are adjusted by the audio control part 10 , and are transmitted to the audio output part 6 and output as audio signals.
- the operation that is selected from the various operations is determined by the operation of the operating switch of the operating part 3 , and the position of the operating switch is transmitted to the input control part 7 .
- FIG. 2 is a diagram showing the first operation of the image control part 9 .
- 11 indicates the image that is to be originally displayed; this image is displayed over the entire display area surrounded by a square frame 12 .
- the image control part 9 has the function of reducing and displaying this overall image with the center of the image as the center.
- the reduced image that is reduced and displayed is indicated by 13 .
- the image control part 9 displays the reduced image 13 during the initial stage of use of the head-mounted display. As the use frequency or use time of the head-mounted display (the selection of which of these is used is determined by the switch of the operating part 3 ) that corresponds to the wearer and that is input from the input control part 7 increases, the reduction ratio of the reduced image 13 is decreased, so that the size of the image is increased. After a specified use frequency or a specified use time is reached, the image 11 that is originally to be displayed is displayed.
- the input control part 7 increases the number of times of use of the device by one increment each time that the power supply of the head-mounted display is switched on, and adds the use time while the power supply of the head-mounted display is switched on.
- the use frequency or use time is calculated for each wearer code that is input from the operating part 3 , and the control described above is performed using the use frequency or use time corresponding to each wearer.
- the wearer initially concentrates his line of sight on the central portion of the visual field, and can receive training in expanding the visual field by the movement of the eyeballs as time passes. Accordingly, the wearer can naturally view the peripheral portions of the image display area by movement of the eyeballs alone.
- FIG. 3 is a diagram showing the second operation of the image control part 9 .
- 14 indicates the image that is to be originally displayed; this image is displayed over the entire image display area surrounded by a square frame 12 .
- the display image includes a background 15 and an icon 16 .
- the image control part 9 has the function of moving the position of the icon 16 , which is a portion of this display image, to the central portion of the image from the inherent display position.
- the actual image that is displayed in the initial stage of use of the head-mounted display is an image such as that indicated by 17 ; the background is invariable, but the position of the icon 16 is shifted toward the center of the image.
- the image control part 9 displays the initial-stage image 17 .
- the use frequency or use time of the head-mounted display (the selection of which of these is used is determined by the switch of the operating part 3 ) that corresponds to the wearer and that is input from the input control part 7 increases, the position of the icon 16 is successively shifted toward the outside of the screen. After a specified use frequency or specified use time is reached, the icon returns to the inherent display position, so that the image 14 that is originally to be displayed is displayed.
- the wearer initially concentrates his line of sight on the icon 16 which is located in the central portion of the visual field.
- the wearer receives training in expanding the visual field by movement of the eyeballs. Accordingly, the wearer can naturally view the peripheral portions of the image display area by movement of the eyeballs alone.
- FIG. 4 is a diagram showing the third operation of the image control part 9 .
- 14 indicates the image that is originally to be displayed; this image is displayed over the entire display area surrounded by a square frame 12 .
- the display image includes a background 15 and an icon 16 .
- the image control part 9 has the function of causing the position of the icon 16 , which constitutes a portion of this display image, to undergo a reciprocating movement between the inherent display position and a position in the central portion of the image (indicated by 16 ′).
- the position in which the icon 16 is inherently displayed is fixed as one end point, and the other end point 16 ′ of the reciprocating movement is shifted toward the outside of the screen as the use frequency or use time of the head-mounted display (the selection of which of these is used is determined by the switch of the operating part 3 ) that corresponds to the wearer and that is input from the input control part 7 increases, and after a specified use frequency or specified use time is reached, the icon returns to the inherent display position, so that the image 14 that is originally to be displayed is displayed.
- the wearer initially directs his attention toward the icon 16 when the icon reaches the position 16 ′, which is in the central portion of the visual field. However, as time passes, the position where the icon is displayed moves toward the outside, so that the wearer can receive training in expanding the visual field by movement of the eyeballs. Consequently, the wearer can naturally view the peripheral portions of the image display area by movement of the eyeballs alone.
- the reduction and movement of the image are controlled in accordance with the use frequency or use time.
- the reduction and movement of the image are controlled on the basis of the directional distribution of the line of sight as described above.
- a publicly known device such as that described in Japanese Laid-Open Patent Application No. H7-32008, for example, can be used as a device for detecting the line of sight.
- the standard deviation of the distribution of the angle of the line of sight centered on the direction of the center of the screen within a specified time may be taken, and this standard deviation may be used instead of the use frequency or use time described above.
- the standard deviation when the standard deviation is small, a display is performed which is the same as that performed when the use frequency or use time is small, and as the standard deviation increases, the display may be varied to a display which is the same as that performed when the use frequency or use time has increased.
- the proportions of the reduced display of the figures in the examples described above may also be determined by a combination of the use frequency or use time and the line of sight detection results.
- the proportions of the reduced display of the figures may also be determined using the square root of the sum of the squares of the use time and the standard deviation of the fluctuation in the line of sight as a reference, or using the square root of the sum of the squares of the use frequency and the standard deviation of the fluctuation in the line of sight as a reference.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Controls And Circuits For Display Device (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004329135 | 2004-11-12 | ||
JP2004-329135 | 2004-11-12 | ||
PCT/JP2005/020332 WO2006051743A1 (ja) | 2004-11-12 | 2005-10-31 | 映像表示装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070262917A1 true US20070262917A1 (en) | 2007-11-15 |
Family
ID=36336424
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/667,279 Abandoned US20070262917A1 (en) | 2004-11-12 | 2005-10-31 | Video Display |
Country Status (4)
Country | Link |
---|---|
US (1) | US20070262917A1 (ja) |
EP (1) | EP1811493B1 (ja) |
JP (1) | JP4925831B2 (ja) |
WO (1) | WO2006051743A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120044347A1 (en) * | 2010-08-20 | 2012-02-23 | Toshiyasu Sugio | Imaging apparatus and imaging method |
CN105684074A (zh) * | 2014-01-20 | 2016-06-15 | 索尼公司 | 图像显示装置和图像显示方法、图像输出装置和图像输出方法与图像显示系统 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB201310364D0 (en) | 2013-06-11 | 2013-07-24 | Sony Comp Entertainment Europe | Head-mountable apparatus and systems |
CN107424583B (zh) * | 2017-07-03 | 2020-05-12 | 威创集团股份有限公司 | 异形图像的显示数据处理方法和系统 |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5635948A (en) * | 1994-04-22 | 1997-06-03 | Canon Kabushiki Kaisha | Display apparatus provided with use-state detecting unit |
US6069608A (en) * | 1996-12-03 | 2000-05-30 | Sony Corporation | Display device having perception image for improving depth perception of a virtual image |
US20010035845A1 (en) * | 1995-11-28 | 2001-11-01 | Zwern Arthur L. | Portable display and method for controlling same with speech |
US20010038360A1 (en) * | 1994-10-28 | 2001-11-08 | Nobuo Fukushima | Display apparatus and its control method |
US6445363B1 (en) * | 1999-02-12 | 2002-09-03 | Canon Kabushiki Kaisha | Head-mounted display apparatus |
US20040025175A1 (en) * | 1996-06-03 | 2004-02-05 | Yasuhiro Tamekuni | Display apparatus |
US20040051680A1 (en) * | 2002-09-25 | 2004-03-18 | Azuma Ronald T. | Optical see-through augmented reality modified-scale display |
US20040104864A1 (en) * | 2002-11-28 | 2004-06-03 | Nec Corporation | Glasses type display and controlling method thereof |
US20050128546A1 (en) * | 2003-12-15 | 2005-06-16 | Canon Kabushiki Kaisha | Optical deflector |
US20060103591A1 (en) * | 2004-11-15 | 2006-05-18 | Canon Kabushiki Kaisha | Information processing apparatus and method for providing observer with information |
US20060119539A1 (en) * | 2002-12-24 | 2006-06-08 | Nikon Corporation | Head mounted display |
US20070030293A1 (en) * | 2005-08-08 | 2007-02-08 | Canon Kabushiki Kaisha | Information processing apparatus and information processing method, and program |
US20070257902A1 (en) * | 2003-04-17 | 2007-11-08 | Sony Corporation | Stereoscopic-Vision Image Processing Apparatus, Stereoscopic-Vision Image Providing Method, and Image Display Method |
US20090251482A1 (en) * | 2008-04-03 | 2009-10-08 | Sony Corporation | Image processing apparatus, display control method, program, and storage medium |
US20100177114A1 (en) * | 2009-01-15 | 2010-07-15 | Brother Kogyo Kabushiki Kaisha | Image display device |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10221639A (ja) * | 1996-12-03 | 1998-08-21 | Sony Corp | 表示装置および表示方法 |
JP2000171751A (ja) * | 1998-12-07 | 2000-06-23 | Hitachi Ltd | 双眼式ディスプレイ |
JP2003076353A (ja) * | 2001-09-04 | 2003-03-14 | Sharp Corp | ヘッドマウント型表示装置 |
JP2003348622A (ja) * | 2002-05-24 | 2003-12-05 | Canon Inc | 立体映像表示方法及び記憶媒体 |
JP2004236241A (ja) * | 2003-01-31 | 2004-08-19 | Nikon Corp | 表示装置 |
-
2005
- 2005-10-31 US US11/667,279 patent/US20070262917A1/en not_active Abandoned
- 2005-10-31 WO PCT/JP2005/020332 patent/WO2006051743A1/ja active Application Filing
- 2005-10-31 EP EP05799987.2A patent/EP1811493B1/en not_active Expired - Fee Related
- 2005-10-31 JP JP2006544874A patent/JP4925831B2/ja not_active Expired - Fee Related
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5635948A (en) * | 1994-04-22 | 1997-06-03 | Canon Kabushiki Kaisha | Display apparatus provided with use-state detecting unit |
US20010038360A1 (en) * | 1994-10-28 | 2001-11-08 | Nobuo Fukushima | Display apparatus and its control method |
US20010035845A1 (en) * | 1995-11-28 | 2001-11-01 | Zwern Arthur L. | Portable display and method for controlling same with speech |
US20040025175A1 (en) * | 1996-06-03 | 2004-02-05 | Yasuhiro Tamekuni | Display apparatus |
US6069608A (en) * | 1996-12-03 | 2000-05-30 | Sony Corporation | Display device having perception image for improving depth perception of a virtual image |
US6445363B1 (en) * | 1999-02-12 | 2002-09-03 | Canon Kabushiki Kaisha | Head-mounted display apparatus |
US20040051680A1 (en) * | 2002-09-25 | 2004-03-18 | Azuma Ronald T. | Optical see-through augmented reality modified-scale display |
US20040104864A1 (en) * | 2002-11-28 | 2004-06-03 | Nec Corporation | Glasses type display and controlling method thereof |
US20060119539A1 (en) * | 2002-12-24 | 2006-06-08 | Nikon Corporation | Head mounted display |
US20070257902A1 (en) * | 2003-04-17 | 2007-11-08 | Sony Corporation | Stereoscopic-Vision Image Processing Apparatus, Stereoscopic-Vision Image Providing Method, and Image Display Method |
US20050128546A1 (en) * | 2003-12-15 | 2005-06-16 | Canon Kabushiki Kaisha | Optical deflector |
US20060103591A1 (en) * | 2004-11-15 | 2006-05-18 | Canon Kabushiki Kaisha | Information processing apparatus and method for providing observer with information |
US20070030293A1 (en) * | 2005-08-08 | 2007-02-08 | Canon Kabushiki Kaisha | Information processing apparatus and information processing method, and program |
US20090251482A1 (en) * | 2008-04-03 | 2009-10-08 | Sony Corporation | Image processing apparatus, display control method, program, and storage medium |
US20100177114A1 (en) * | 2009-01-15 | 2010-07-15 | Brother Kogyo Kabushiki Kaisha | Image display device |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120044347A1 (en) * | 2010-08-20 | 2012-02-23 | Toshiyasu Sugio | Imaging apparatus and imaging method |
US9774830B2 (en) * | 2010-08-20 | 2017-09-26 | Panasonic Intellectual Property Management Co., Ltd. | Imaging apparatus and imaging method |
CN105684074A (zh) * | 2014-01-20 | 2016-06-15 | 索尼公司 | 图像显示装置和图像显示方法、图像输出装置和图像输出方法与图像显示系统 |
US20160246057A1 (en) * | 2014-01-20 | 2016-08-25 | Sony Corporation | Image display device and image display method, image output device and image output method, and image display system |
US10437060B2 (en) * | 2014-01-20 | 2019-10-08 | Sony Corporation | Image display device and image display method, image output device and image output method, and image display system |
Also Published As
Publication number | Publication date |
---|---|
EP1811493A1 (en) | 2007-07-25 |
EP1811493A4 (en) | 2010-09-01 |
JP4925831B2 (ja) | 2012-05-09 |
JPWO2006051743A1 (ja) | 2008-05-29 |
EP1811493B1 (en) | 2013-04-24 |
WO2006051743A1 (ja) | 2006-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10444833B2 (en) | Methods and apparatus for vision enhancement | |
US20200225489A1 (en) | Head-Mounted Device with Active Optical Foveation | |
US10488925B2 (en) | Display control device, control method thereof, and display control system | |
US20210350762A1 (en) | Image processing device and image processing method | |
EP3158531B1 (en) | A method and a display device with pixel repartition optimization | |
JP2023027369A (ja) | 操作可能な中心窩ディスプレイ | |
US11803061B2 (en) | Hybrid see through augmented reality systems and methods for low vision users | |
US11546527B2 (en) | Methods and apparatuses for compensating for retinitis pigmentosa | |
EP3588447B1 (en) | Image processing device, image processing method, and image system | |
CN109791431A (zh) | 视点渲染 | |
JP2006267604A (ja) | 複合情報表示装置 | |
US11144119B2 (en) | Methods and systems for generating a magnification region in output video images | |
US20070262917A1 (en) | Video Display | |
US20210240275A1 (en) | Head-mounted display with user-operated control | |
JP4199818B1 (ja) | 眼精疲労軽減システムおよびプログラム | |
WO2022004130A1 (ja) | 情報処理装置、情報処理方法、および記憶媒体 | |
JP2009020180A (ja) | 映像表示装置 | |
JP5343304B2 (ja) | ウェアラブルディスプレイ | |
JP2006163009A (ja) | 映像表示方法 | |
KR100311462B1 (ko) | 온스크린디스플레이를이용한안구운동제어방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NIKON CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OTSUKI, MASAKI;KATO, SHIGERU;ISAGI, KAZUOU;AND OTHERS;REEL/FRAME:019461/0026;SIGNING DATES FROM 20070525 TO 20070529 Owner name: NIKON SYSTEMS INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OTSUKI, MASAKI;KATO, SHIGERU;ISAGI, KAZUOU;AND OTHERS;REEL/FRAME:019461/0026;SIGNING DATES FROM 20070525 TO 20070529 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO PAY ISSUE FEE |