WO2005124431A1 - 情報表示装置、ナビゲーション装置 - Google Patents
情報表示装置、ナビゲーション装置 Download PDFInfo
- Publication number
- WO2005124431A1 WO2005124431A1 PCT/JP2005/011064 JP2005011064W WO2005124431A1 WO 2005124431 A1 WO2005124431 A1 WO 2005124431A1 JP 2005011064 W JP2005011064 W JP 2005011064W WO 2005124431 A1 WO2005124431 A1 WO 2005124431A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- display
- information
- light
- unit
- transmittance
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/09—Arrangements for giving variable traffic instructions
- G08G1/0962—Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
- G08G1/0968—Systems involving transmission of navigation instructions to the vehicle
- G08G1/0969—Systems involving transmission of navigation instructions to the vehicle having a display in the form of a map
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B29/00—Maps; Plans; Charts; Diagrams, e.g. route diagram
- G09B29/10—Map spot or coordinate position indicators; Map reading aids
- G09B29/106—Map spot or coordinate position indicators; Map reading aids using electronic means
- G09B29/108—Map spot or coordinate position indicators; Map reading aids using electronic means involving the optical projection of slides, microfilm or the like
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2300/00—Aspects of the constitution of display devices
- G09G2300/08—Active matrix structure, i.e. with use of active elements, inclusive of non-linear two terminal elements, in the pixels together with light emitting or modulating elements
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G3/00—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
- G09G3/20—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
- G09G3/34—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters by control of light from an independent source
Definitions
- the present invention relates to an information display device for displaying various information to be presented to a user, and particularly to a see-through type information display device for displaying various information in a manner superimposed on a real scene visually recognized by the user.
- a so-called head-up display which presents various information such as a text indicating a traveling speed and a speedometer to a driver through a windshield for a moving object such as a vehicle or an aircraft. It has been transformed.
- a typical example of this type of head-up display is to irradiate a light beam from a projection device such as a projector to a complier composed of, for example, a half mirror and an HOE (Holographic Optical Element).
- a device that presents various kinds of information to the user by the reflected light at the convener.
- Patent Document 1 JP-A-5-104980
- An object of the present invention is to provide an information display device such as a head-up display and a navigation device capable of improving visibility.
- an information display device is an information display device that displays notification information to be notified to a user
- a display means comprising a base material that transmits light, and a display surface for displaying the notification information provided opposite to the incident surface of the visible light, and disposed so as to overlap the incident surface of the display unit Means for adjusting the luminous intensity of the visible light incident on the incident surface, wherein the area is formed of a transparent base material and is divided into a plurality of areas overlapping the display means.
- Light control means for changing light transmittance, display control means for generating data including the notification information as image data, and controlling display of the notification information on the display means; Broadcast information display position Specifying the area of the light control means corresponding to the display position, and transmitting the light to reduce the light transmittance of the specified area by controlling the light control means. Rate control means.
- a navigation device that displays route guidance information necessary for route guidance accompanying movement of a moving object, and includes a visible light beam.
- a display unit which is formed of a base material that transmits light, and has a display surface for displaying the route guidance information facing the incident surface of the visible light, and a display unit that is provided so as to overlap the incident surface of the display unit.
- Dimming means whose rate is changed, recording means on which image data containing the route guidance information is recorded, and images recorded on the recording means
- a display control means for controlling the display of the route guidance information on the display means; and a display position of the route guidance information on the display means, and the light control means corresponding to the display position. It is characterized by comprising a specifying means for specifying an area, and a transmittance control means for reducing the light transmittance of the specified area by controlling the light control means.
- FIG. 1 is a block diagram showing a configuration of an information display device I according to a first embodiment.
- FIG. 2 is a diagram illustrating a method of dividing the light control panel 15 in the embodiment.
- FIG. 3 is a diagram illustrating a method of dividing the light control panel 15 according to the embodiment.
- FIG. 4 is a diagram showing a configuration example when the light control panel 15 in the embodiment is configured as a segment type liquid crystal panel.
- FIG. 5 is a diagram showing a configuration example when the light control panel 15 in the embodiment is configured as an active matrix type liquid crystal panel.
- FIG. 6 is a sequence diagram showing a process executed when an image for one frame is displayed on the information display device I according to the embodiment.
- FIG. 7 shows a relationship between an object o 3 observed from a viewing position P and a cell when the complier 14 and the dimming panel 15 are arranged at intervals in the information display device I according to the first modification of the first embodiment.
- FIG. 8 is a view showing the positional relationship between the viewing position P and the complier 14 and the dimming panel 15 when the complier 14 and the dimming panel 15 are arranged at intervals in the information display device I according to the second modification of the embodiment.
- FIG. 8 is a view showing the positional relationship between the viewing position P and the complier 14 and the dimming panel 15 when the complier 14 and the dimming panel 15 are arranged at intervals in the information display device I according to the second modification of the embodiment.
- FIG. 9 is a block diagram showing a configuration of an information display device 12 according to a third modification of the embodiment.
- FIG. 10 is a block diagram showing a configuration of an information display device 13 according to a second embodiment.
- FIG. 11 is a block diagram showing a configuration of an information display device 14 according to a modification of the second embodiment.
- FIG. 12 is a block diagram showing a configuration of a navigation device 100 according to application example 1.
- FIG. 13 A process executed by the system control unit 110 of the navigation device 100 according to the application example 1. It is a flowchart which shows a process.
- the information display device I according to the present embodiment will be described with reference to FIG. As shown in the figure, the information display device I according to the present embodiment includes a recording unit 11, a display information generation unit 12, a projector 13, a complier 14, a light control panel 15, and a transmittance control unit 16. Yes.
- the information display device I when projecting an image in which an object indicating information content is drawn from the projector 13 to the complier 14, determines the transmittance of the external light R2 in the object display area.
- the dimming panel 15 is used to reduce the brightness and to ensure the visibility of the object by the user.
- the recording unit 11 is composed of a recording device such as an HD (Hard Disk) drive, and stores data for displaying objects such as text, polygons, and polylines, and actual projection. Data for displaying an image is recorded.
- the display information generating unit 12 generates image data for displaying various objects corresponding to information to be notified to a user, such as arrows and text, based on various data recorded in the recording unit 11. Output to the projector 13.
- the specific contents of the image data generated by the display information generation unit 12 are arbitrary.
- image data for displaying an arrow indicating a blinker operation state or a speedometer is generated in accordance with an on / off of a blinker switch by a user or an input value from a speed sensor. You may do it.
- an arrow indicating the direction of travel along with a character string such as “Please turn right 30m ahead” (hereinafter “route guidance information”). ) May be generated.
- image data corresponding to a real video image may be generated.
- a configuration example in the case of linking with the navigation device will be described later.
- the display position of the arrow may be determined in advance, or the display position may be changed according to the support of the user. You may do it.
- the size of the object to be displayed is arbitrary, and the size of the object to be displayed may be determined in advance, or the user may be able to select the object size.
- the projector 13 has, for example, a light source and a liquid crystal panel, drives the liquid crystal panel based on the image data input from the display information generating unit 12, and generates a light beam R1 corresponding to the image data. Irradiate Connoina 14.
- the complier 14 is composed of, for example, a half mirror or an HOE, and transmits a part of the external light R2 and reflects a light beam R1 emitted by the projector 13. As a result, a part of the external light R2 transmitted through the connector 14 and the reflected light of the light ray R1 emitted by the projector 13 are combined in the complier 14 and visually recognized by the user.
- the dimming panel 15 is configured by a transmissive liquid crystal panel, and the transmittance changes when a voltage is applied or removed.
- the dimming panel 15 is formed integrally with the compiler 14, is installed on the incident surface side of the external light R 2 with respect to the compiler 14, and changes the transmittance so that the light incident on the compiler 14 is changed. Adjust the amount of light R2.
- the reflected light of the light ray R1 emitted from the projector 13 and the transmitted light of the external light R2 are combined in the combiner 14 and are visually recognized by the user. If the transmittance of the external light R2 in the optical panel 15 is reduced uniformly over the entire panel, it becomes difficult to secure the user's view through the complier 14. Therefore, in the present embodiment, the light control panel 15 is divided into a plurality of regions (hereinafter, this region is referred to as “cell”), and the transmittance can be changed for each cell.
- the transmittance of only the cell including the projection area of the object is reduced, and the user's view through the combiner 14 is secured while maintaining the visibility. It is configured to improve the visibility of the object.
- the “region” in the claims corresponds to, for example, this cell.
- the light control panel 15 is divided into dice-shaped cells of numbers “1” to “15” as shown in FIG. 2 (hereinafter, this number is referred to as “cell number”).
- this number is referred to as “cell number”.
- the dimming panel 15 reduces the transmittance of the “8” cell including the arrow ol, and the arrow ol is visually recognized. It improves the quality.
- the shape and display position of the object based on the image data generated by the display information generating unit 12 change. For example, in the example shown in FIG.
- the cells including the arrow o2 are “9”, “10”, “14”, and “15”. Change to number four. As described above, when the display position or shape of the object changes, the transmittance of the cell “8” is increased and the transmittance of the four cells is reduced.
- the specific division form of the cell is arbitrary.
- the light control panel 15 is divided into smaller cells and includes information (objects) to be notified to the user.
- the transmittance of a plurality of cells may be reduced at the same time.
- the dimming panel 15 may have the same resolution as the projector 13 so that one pixel in the projector 13 and the dimming panel One cell in the rule 15 may be made to correspond.
- the specific configuration of the light control panel 15 for realizing a powerful function will be described later.
- the three components of the projector 13, the compiler 14, and the light control panel 15 are used. It is necessary that the correspondence between the coordinates is established between the persons. For example, the coordinates of the liquid crystal panel mounted on the projector 13 are (x ', y'), the coordinates within the projectable range of the complier 14 are (x, y), and the coordinates of the light control panel 15 are (X, y). Consider the case of Y).
- the display information generating unit 12 generates image data based on the coordinates (x, y) on the complier 14, but ( ⁇ ′, y ′) Since the relationship between (x, y) changes depending on the distance between the compiler 14 and the projector 13, (x ', y') and (x, y) are associated in advance, and if not, The display position of the object on the liquid crystal panel does not correspond to the display position on the compiler 14.
- the coordinates (x, y) on the complier 14 and the coordinates (X, Y) on the light control panel 15 do not match. In other cases, the light control panel corresponding to the display position of the object It would be impossible to identify the 15 positions. Furthermore, if the value of the coordinates (X, Y) contained in each cell of the light control panel 15 cannot be specified, it becomes impossible to specify the cell to which the specified coordinates (X, Y) belong.
- the display information generating unit 12 when the display information generating unit 12 generates image data based on the data recorded in the recording unit 11, the display information generating unit 12 includes the image data in the image corresponding to the data.
- Information to be presented to the user that is, a group of coordinates (x, y) on the compiler 14 included in the area where the object is to be displayed is calculated.
- the recording unit 11 specifies coordinates ( ⁇ ′, y ′) and coordinates (X, Y) based on the coordinates ( x , y ).
- a table (hereinafter referred to as “xx ′ conversion table” and “xX conversion table”, respectively) and a cell specifying table for specifying the cell number on the light control panel 15 based on the coordinates (X, Y) are provided. It is recorded.
- (x ′, y ′) and (X, Y) corresponding to the group of coordinates (x, y) calculated by the display information generating unit 12 are These tables
- the cell number is specified based on the table, and a cell number specifying table is searched based on the specified coordinates (X, Y) to specify a cell whose transmittance is to be reduced.
- the form of the coordinate group generated by the display information generating unit 12 is arbitrary.
- all coordinate values included in the display area of the object are generated by tabulating.
- a feature point (for example, each vertex of a corner of an arrow) may be determined in advance for the object, and a co-ordinate value of each feature point may be calculated.
- FIG. 4 is a perspective view showing a configuration example of the light control panel 15 when the segment method is adopted
- FIG. 5 is a light control panel when the active matrix method using a TFT (Thin Film Transistor) is adopted.
- 15 is a perspective view showing a configuration example of the fifteenth embodiment.
- the light control panel 15 is configured such that the element substrate 1501 and the opposing substrate 1502 are bonded to each other with a gap therebetween, and a liquid crystal 1503 is provided between the substrates 1501 and 1502. Is filled.
- Both substrates 1501 and 1502 are made of a transparent material such as Si02 having high transmittance to secure the transmittance of external light R2, and are provided with polarizing films 1504 and 1505.
- transparent electrodes 1506 and 1507 having a shape corresponding to each cell are formed on the opposing surfaces of both substrates 1501 and 1502, and the transparent electrodes 1506 and 1507 are arranged to face each other with the liquid crystal 1503 layer therebetween.
- the electrode 1507 on the counter substrate 1502 side is connected to the reference voltage
- the electrode 1506 on the element substrate 1501 side is connected to the reference voltage via the switch 1508.
- an electric potential difference is generated between the electrodes 1506 and 1507 of the substrates 1501 and 1502 on the both substrates in the switch off state, the liquid crystal 1503 existing between the electrodes 1506 and 1507 is oriented in a certain direction, and the polarizing films 1504 and 1505
- the transmittance increases due to the interaction with the liquid crystal, and the alignment of the liquid crystal becomes random when the switch 1508 is turned on, so that the transmittance decreases.
- the switches 1508 provided corresponding to the electrodes 1506 and 1507 are turned on and off by the transmittance control unit 16, and as a result, in the area corresponding to each cell on the light control panel 15.
- the transmittance changes.
- the light control panel 15 has a gap between the element substrate 1511 provided with the polarizing films 1514 and 1515 and the opposing substrate 1512 as shown in FIG.
- the two substrates 1511 and 1512 are laminated and filled with a liquid crystal 1513.
- a transparent common electrode 1517 is formed, and on the element substrate 1511, transparent element electrodes 1516 corresponding to each cell are arranged in a matrix.
- Each element electrode 1516 is connected to the drain of the MOSFET 1518, the gate is connected to the scanning line, and the source is connected to the data line.
- the MOSFET 1518 is switched on and off in accordance with the input signal. It is held between the element electrode 1516 and the common electrode 1517.
- the orientation of the liquid crystal existing between the electrodes changes, and the interaction with the polarizing films 1514 and 1515 changes the transmittance of the region corresponding to each cell.
- the transmittance control unit 16 changes the transmittance of the external light R2 in a partial area of the light control panel 15 by controlling signals for each scanning line, data line, and common electrode. .
- the configuration and the driving method of the light control panel 15 are the same as those of the conventional liquid crystal panel except that each transparent electrode is formed in a shape corresponding to the cell, and thus the details are omitted.
- the liquid crystal panel constituting the light control panel 15 may employ another method such as a simple matrix drive method in addition to the above two methods.
- FIG. 6 is a sequence diagram showing a process executed when displaying an image for one frame in the information display device I according to the present embodiment.
- the display information generating unit 12 reads data such as polygons recorded in the recording unit 11 and draws an object to be displayed on the complier 14 based on the read data. Generated image data (Step Sl). After generating the image data, the display information generating unit 12 calculates a group of coordinates (x, y) for specifying the display position of the object drawn in the image data (step S2). The calculated coordinates (x, y) group together with the image data generated in step SI is used for the projector. And the output to the transmittance control unit 16 (steps S3 and S4). At this time, the form of the group of coordinates (x, y) calculated in the display information generation unit 12 is arbitrary as described above.
- the projector 13 causes the coordinate on the liquid crystal panel to correspond to the coordinate (x, y) group.
- the (X ′, y ′) group is recorded in the recording unit 11, and the xx ′ conversion table is read out (step S5).
- the projector 13 drives the liquid crystal panel based on the image data supplied from the display information generating unit 12 and the specified coordinates (x, y) (Step S6).
- the light beam R1 emitted from the projector 13 to the complier 14 includes an object drawn on the image data generated by the display information generation unit 12, that is, an object corresponding to the information to be presented to the user.
- the object is displayed at a position on the complier 14 corresponding to the group of coordinates (x, y).
- the transmittance control unit 16 which has been supplied with the image data and the coordinate (x, y) group from the display information generating unit 12, controls the light control panel 15 corresponding to the coordinate (x, y) group.
- the coordinates (X, Y) of the xX conversion table recorded in the recording unit 11 are also read (step S7), and further, a cell identification table is searched based on the read coordinates (X, Y), and the coordinates (X, Y) are searched.
- the cell number corresponding to (X, Y) is specified (step S8).
- the transmittance control unit 16 reads out the coordinates (X, Y) corresponding to all the coordinate values included in the coordinate (x, y) group supplied from the display information generating unit 12, and reads the coordinates (X , Y).
- the transmittance control unit 16 controls the light control panel 15 to reduce the transmittance of cells corresponding to all the cell numbers specified in step S8.
- the light ray R1 corresponding to the image of the frame to be displayed is projected on the compiler 14 and the transmittance of the cell corresponding to the display position of the object on the compiler 14 Is reduced.
- the light intensity of the external light R2 transmitted through the light control panel 15 and incident on the connector 14 is reduced, and the visibility of an image corresponding to the light ray R1 is improved.
- the information display device I provides information to be notified to the user, that is, information display for displaying an object, on a light-transmitting panel provided in the field of view of the user.
- the apparatus I which is composed of a base material that transmits the external light R2 and has a display surface for displaying an object facing the entrance surface of the external light R2, and an external light R2 of the compiler 14.
- the cell of the light control panel 15 corresponding to the display position is specified, and the light transmittance of the specified cell is reduced by controlling the light control panel 15.
- the light control panel 15 includes two or more transparent substrates and a plurality of transparent substrates formed in a shape corresponding to a cell on a facing surface of each substrate.
- the transmissivity control unit 16 controls the application of a voltage to the transparent electrode to control each cell of the light control panel 15 by controlling a voltage applied to the transparent electrode and a liquid crystal material filled between the two substrates. Since the configuration is such that the light transmittance is reduced, it is possible to arbitrarily determine the range of each cell and adjust the area for controlling the transmittance.
- the light ray R1 including the object is reflected on the reflection surface of the complier 14 based on the image data.
- the projector further includes a projector 13 for irradiating, and the complier 14 is configured to display an object by reflecting the light ray R1 on the reflecting surface. This With this configuration, the light beam Rl emitted from the projector 13 is projected on the complier 14 and visually recognized by the user.
- the information display device I which is useful in the present embodiment, associates the object display position on the complier 14 with the cell on the light control panel 15 that is superimposed on the display position and visually recognized by the user.
- An xX conversion table and a cell identification table storing information for the object are prepared in advance, and the transmittance control unit 16 determines the cell of the light control panel corresponding to the display position of the object in the combiner 14 based on this table. Since the display position of the object is changed, the cell corresponding to the position can be specified at any time even when the display position of the object changes, and the transmittance of the cell can be reduced.
- the force described in the case where a liquid crystal panel is used as the light control panel 15 The light control panel 15 has, for example, a transmittance of the external light R2 depending on an applied voltage.
- the transmittance control unit 16 uses the xX conversion table and the cell identification table recorded in the recording unit 11 to store cells corresponding to the display position of the object.
- a specific configuration was employed. However, it is also possible to directly specify the cell corresponding to the display position of the object based on the coordinate (x, y) group supplied from the display information generating unit 12.
- a table for specifying a cell may be recorded in the recording unit 11 based on the coordinates (x, y), and the cell may be specified using the table. Good,.
- the complier 14 and the dimming panel 15 are integrally configured. But dimming with these Compina 14 It is also possible to create the panel 15 as a separate body and arrange it with an interval between them.
- the coordinate group (yl to y2) corresponding to the display position of the object o3 on the compiler 14 is represented by the coordinates (Y1 to Y2) corresponding to the "2" -th cell of the light control panel 15. It has become. However, if there is an interval between the complier 14 and the light control panel 15, this object ⁇ 3 is visually recognized so as to be superimposed on the cell “1” of the light control panel 15 from the actual viewing position ⁇ of the user. It will be. The same applies to the X and X axis directions.
- the coordinates on the combiner 14 are determined based on the positional relationship between the actual visual recognition position ⁇ ⁇ ⁇ from the user and the arrangement positions of the compina 14 and the light control panel 15. It is necessary to determine the correspondence between (x, y) and the coordinates (X, Y) on the light control panel 15.
- the method of determining the actual correspondence is arbitrary, and the coordinates (X, Y) of the light control panel 15 corresponding to each coordinate (x, y) on the complier 14 may be determined visually. good.
- the cell corresponding to the object display position of the combiner 14 is reliably specified and transmitted. It is possible to reduce the rate and thereby improve the visibility of the object.
- the coordinates (x, y) on the complier 14 and the coordinates (X, Y) on the dimming panel 15 match.
- the combiner 14 and the dimming panel 15 are installed, for example, on the windshield of the vehicle, and the driving This is not a problem when viewing from a seat, but if the user's viewing position fluctuates significantly, the two coordinates may not match.
- the viewing position P of the user is determined in advance and registered in the information display device I. Then, when displaying the object, a cell whose transmittance is to be reduced is specified using the xX conversion table and the cell specifying table corresponding to the registered viewing position P.
- This method is a method of creating an xX conversion table at the start of use of the information display device I, utilizing the fact that the relative positional relationship between the complier 14 and the light control panel 15 is fixed.
- the user when using the information display device I, the user measures the distance between the viewing position ⁇ and the complier 14 in advance, and registers the measured value in the information display device I. Then, a ⁇ conversion table is created by substituting the registered measurement value into the above (Equation 1), and a cell whose transmittance is to be reduced is specified using the ⁇ conversion table.
- a conversion table that stores information for associating cells on the light control panel 15 and the object display position that are visually recognized by the user is prepared for each different visual position, and the user's visual position is determined. After the identification, the cell corresponding to the display position of the object in the combiner 14 is identified using the conversion table corresponding to the visual recognition position.
- the visual position of the user also depends on the distance to the dimming panel 15 and the complier 14, and the information for associating the object display position with the cell that is superimposed on the display position and visually recognized by the user is also used.
- a configuration is generated based on the above (Equation 1) and identifies the cell based on the information.
- FIG. 9 shows a configuration of an information display device 12 according to the present modification.
- the same elements as those in FIG. 1 described above are denoted by the same reference numerals. Therefore, unless otherwise specified, elements denoted by the same reference numerals as those shown in FIG. 1 have the same configuration as in the first embodiment and perform the same operations.
- the information display device 12 includes a complier 14 and a projector. 13, information to be notified to the user, that is, a configuration for projecting an object has been adopted.
- an information display device 12 according to the present modification is provided with a display panel 17 in place of the complier 14 and the projector 13, and intends to display various objects on the display panel 17.
- the display panel 17 is configured by a self-luminous display device that transmits a part of the external light R2 like a transmissive organic EL panel, for example.
- the display panel 17 outputs display light R3 corresponding to the image data.
- the transmitted light of the external light R2 and the display light R3 are combined via the display panel 17, and are visually recognized by the user.
- the transmittance of the cell corresponding to the display area of the object is reduced. Function is realized.
- the specific configuration of the light control panel 15 is the same as that in FIGS. 4 and 5 described above.
- the display panel 17 is driven based on the image data generated by the display information generation unit 12, and thus the first embodiment is performed. It is not necessary to consider the correspondence between the coordinates (X ′, y ′) on the liquid crystal panel mounted on the projector 13 and the coordinates ( x , y) on the complier 14 as in the embodiment. On the other hand, even in the case of employing a small configuration, it is necessary to specify a cell corresponding to the display position of the object.
- the recording unit 11 does not record the xx 'conversion table, but records only the xX conversion table and the cell identification table. Note that when displaying an image corresponding to each frame based on the image data, The method for specifying the file is the same as in the first embodiment.
- the light control panel 15 By reducing the transmittance, the visibility of the object displayed on the display panel 17 can be ensured.
- the display panel 17 may require a light source, such as a transmissive liquid crystal panel.
- a light source such as a transmissive liquid crystal panel.
- a half mirror is formed on one surface of the light control panel 15, and the display panel 17 faces the surface on which the mirror is formed.
- FIG. 10 is a block diagram illustrating a configuration of the information display device 13 according to the present embodiment.
- the same elements as those in FIG. 1 described above are denoted by the same reference numerals. Therefore, unless otherwise specified, elements denoted by the same reference numerals as those shown in FIG. 1 have the same configuration as in the first embodiment and perform the same operations.
- the information display device I reduces the transmittance of the cell corresponding to the display position of the information to be notified to the user irrespective of the luminous intensity of the external light R2, so that the information can be visually recognized. It was intended to secure the nature. However, in a situation where the luminous intensity of the external light R2 is low, the visibility of the information can be secured without reducing the transmittance of the light control panel 15. On the other hand, for example, in the case of nighttime, it can be assumed that the luminous intensity of the external light R2 is increased only in a part of the compina due to the influence of a street light, a headlight of an automobile, and the like.
- the information display device 13 adjusts the transmittance of the light control panel 15 according to the luminous intensity of the external light R2, thereby improving the visibility of the object corresponding to the information to be notified to the user. Is to try.
- the information display device 13 according to the present embodiment is provided with the external light luminous intensity distribution detecting section 18 together with the respective components shown in FIG.
- the external light luminous intensity distribution detection unit 18 has a camera installed on the incident surface side of the external light R2 with respect to the light control panel 15, that is, on the other side of the light control panel 15 as viewed from the user.
- the luminous intensity of the external light R2 incident on each cell of the light control panel 15, that is, the luminous intensity distribution of the external light R2 incident on the light control panel 15, is detected based on the captured image.
- the specific installation position and imaging range of the camera are arbitrary.
- a camera is set with an angle of view set so that the entire surface of the light control panel 15 can be imaged, and an image of the incident surface side of the external light R2 of the light control panel 15 is taken, and the reflection on the light control panel 15
- the luminous intensity distribution of the external light R2 may be calculated based on the light amount.
- a camera may capture an image in the direction opposite to the dimming panel 15 and calculate the luminous intensity distribution of the external light R2 based on the captured image.
- the captured video and each light on the light control panel 15 are calculated. It is necessary that the cell numbers are associated with each other. Therefore, in the present embodiment, a table in which the position on the image captured by the camera and the cell number of the light control panel 15 are associated is recorded in the recording unit 11, and the external light intensity distribution detection unit 18 Calculates the luminous intensity of the external light R2 in each cell based on this table. Then, the calculation result is output to the transmittance control unit 16.
- the transmittance control unit 16 specifies the cell number based on the coordinates (x, y) group supplied from the display information generation unit 12 and supplies the cell number from the external light intensity distribution detection unit 16.
- the luminous intensity of the external light R2 incident on the cell of the specified number is calculated based on the calculated data, and the transmittance of the light control panel 15 is controlled based on the calculation result.
- the transmittance control unit 16 specifies a method of specifying a cell whose transmittance is to be reduced, and a display information generating unit.
- the processing executed by the projector 13 based on the image data and the coordinate (x, y) group supplied from 12 is the same as that in FIG. 6 described above.
- the method by which the transmittance control unit 16 controls the transmittance is arbitrary. For example, the following method is adopted. It is possible.
- This method sets a light intensity threshold value in advance, and reduces the transmittance of the cell when the light intensity of the external light R2 incident on the cell corresponding to the display area of the object exceeds the threshold value.
- This is a method for driving the light control panel 15.
- This method is a method of controlling the transmittance of the light control panel 15 so that the luminous intensity of the external light R2 transmitted through the light control panel 15 takes a constant value between the cells.
- the relationship between the voltage applied to each electrode of the light control panel 15 and the transmittance is measured in advance, and a voltage corresponding to the transmittance to be set is applied to the electrode corresponding to the cell. It is necessary to control so that.
- the information display device 13 further includes the external light intensity distribution detecting unit 18 that detects the light intensity of the external light R2 incident on the light control panel 15,
- the transmittance control unit 16 is configured to reduce the transmittance of the external light R2 in the dimming panel 15 in accordance with the luminosity detected by the luminous intensity distribution detection unit 18.
- the transmittance of the light control panel 15 is adjusted according to the intensity of the external light R2, so that the transmittance is reduced only when it is necessary to reduce the transmittance, thereby reducing power consumption. It is possible to plan.
- the information display device 13 of the present embodiment employs a configuration using the complier 14 and the projector 13. However, also in the present embodiment, it is possible to adopt a configuration in which the display panel 17 is provided instead of these as in the third modification of the first embodiment.
- a camera is provided in the external light luminous intensity distribution detecting section 18 and the luminous intensity of the external light R2 corresponding to each coordinate value is calculated, so that the information on the complier 14 is calculated. It is configured to calculate the luminous intensity distribution. However, if there is no need to calculate the luminous intensity distribution, a photometer can be used instead, and the transmittance can be controlled uniformly according to the detection result of the photometer.
- a camera is installed on the incident surface side of the external light R2 with respect to the light control panel 15, and based on the luminous intensity of the external light R2 incident on the light control panel 15, the camera is provided to the light control panel 15.
- permeability in this case was employ
- the information display device 14 according to the present embodiment is configured so that the image actually viewed by the user, that is, the external light R1 transmitted through the complier 14 and the light beam R1 projected on the complier 14 are different from each other. The purpose is to control the transmittance of the light control panel 15 based on the image that is combined and visually recognized by the user.
- FIG. 11 shows the configuration of the information display device 14 having a powerful function.
- the same elements as those in FIG. 1 described above are denoted by the same reference numerals.
- the information display device 14 according to the present embodiment is provided with a superimposed video imaging unit 19 and a video analysis unit 20 in addition to the components shown in FIG.
- the superimposed video imaging means 19 has a camera installed at a position visually recognized by the user, and the angle of view of the camera is set so that the entire surface of the complier 14 can be imaged! Data corresponding to the video imaged by this camera is supplied from the superimposed video imaging unit 19 to the video analysis unit 20 and analyzed by the video analysis unit 20.
- the video analysis unit 20 evaluates the visibility of the object in the combiner 14 based on the data supplied from the superimposed video imaging unit 19. Specifically, the video analysis unit 20 executes the following processing.
- the following process is an example of a specific process for evaluating the visibility of the broadcast information, and the process performed in the video analysis unit 20 is not necessarily limited to this.
- the video analysis unit 20 performs edge detection processing based on the data supplied from the superimposed video imaging unit 19, and converts an image corresponding to the data into a line image.
- the video analysis unit 20 calculates a luminance difference between each pixel included in the data and an adjacent pixel.
- the image corresponding to the data is binarized by setting the difference value to “0” if the difference value is equal to or more than a certain value, and “1” if the difference value is equal to or less than the certain value (hereinafter, the image obtained by binarizing the Is referred to as “line image data”).
- the “line image processing” in the claims includes, for example, a process of generating the line image data.
- the video analyzing unit 20 supplies the image data from the display information generating unit 12 to the projector 13. After acquiring the image data to be processed, the edge detection process is performed on the image data to perform line image processing. Then, the video analysis unit 20 compares the two pieces of line image data for each cell including the coordinates for displaying the object, and calculates the degree of inconsistency T corresponding to each cell.
- the degree of inconsistency T is an index indicating the coincidence of line images, and is calculated by the following equation.
- the visibility of the object included in the image is good, the state of the object is good, the coincidence of the two-line images is good, and the calculated value of (Wi ⁇ Zi) is “0”. Is related to the increase in the number of pixels.
- the calculated values of (Wi-Zi) are “1” because the binary values of each pixel do not match. ”, The number of pixels is increased.
- the video analysis unit 20 compares the calculation result of the degree of mismatch T with a predetermined threshold, and transmits the cell corresponding to the degree of mismatch T when the calculation result is larger than the threshold. The rate is to be reduced. In this way, when it is determined whether or not to reduce the transmittance for each cell, the video analysis unit 20 outputs a signal indicating the determination result to the transmittance control unit 16, and transmits the signal indicating the determination result.
- the control unit 16 reduces the transmittance of the cell corresponding to the display area of the object only when the determination result indicated by the signal indicates that the transmittance should be reduced.
- the information display device 14 includes the superimposed video imaging unit 19 that captures an image of the complier 14 in a state where the object is displayed, and the object display in the data corresponding to the captured video.
- the difference between the luminance component corresponding to the pixels included in the display range and the neighboring range and the luminance component of the range in the image data including the object is calculated, and the user information is visually recognized based on the difference value.
- a video analyzing unit 20 that generates information indicating the state.
- the transmittance control unit 16 further includes a visible light in the dimming panel 15 based on the information generated by the video analyzing unit 20. A cell that reduces the transmittance of the cell is specified, and the light transmittance of the cell is reduced.
- FIG. 12 is a block diagram showing a configuration of a navigation device 100 that is useful in this application example.
- a navigation device 100 includes a GPS (Global Positioning System) receiving unit 101, a sensor unit 102, and an interface unit 103 (hereinafter, “interface” is abbreviated as “IZF”.
- VICS Vehicle Information Communicatio
- Data receiving unit 104 HD drive 105, operation unit 106, information display unit 107, audio processing unit 108, speaker 109, system control unit 110, ROM (Read Only Memory) / RAM (Random Access Memory) unit 111 and a data bus 112 for interconnecting these elements.
- ROM Read Only Memory
- RAM Random Access Memory
- this application example is an application of the information display device I to the information display unit 107 of the car navigation device 100.
- a dimming panel 15 and a complier 14 are provided on the entire front glass of a vehicle, and are superimposed on a real scene visually recognized by a user (driver) through a windshield, for example, a route. This is for projecting an object such as an arrow for guidance on the compiler 14.
- GPS receiving section 101 receives a GPS radio wave transmitted by a geodetic satellite, calculates a latitude Z longitude corresponding to the current position of the vehicle based on the received GPS radio wave, and calculates the calculated latitude.
- the Z longitude is output to the IZF unit 103 as GPS data.
- the sensor unit 102 includes various sensors for detecting the traveling speed, acceleration, and azimuth of the vehicle, and outputs traveling data corresponding to the detection result to the IZF unit 103.
- IZF section 103 calculates the current position of the vehicle based on the traveling data input from sensor section 102 and the GPS data input from GPS receiving section 101, and corresponds to the calculation result.
- the vehicle position data is output to the system control unit 110.
- VICS data receiving section 104 receives a broadcast radio wave such as an FM multiplex broadcast, extracts VICS data included in the broadcast radio wave, and outputs it to system control section 110.
- VICS refers to a road traffic information communication system
- VICS data refers to road traffic information such as traffic jams, accidents, and traffic restrictions.
- the HD drive 105 has an HD in which information can be rewritten.
- a three-dimensional schematic diagram (hereinafter referred to as a "driver's view map") that shows the driver's viewpoint is used. )
- a map that three-dimensionally represents the scenery seen by the driver while the vehicle is moving, or a flat map background or polygons, polygons, text for road display, etc.
- Object data such as polygons for displaying objects such as arrows for the inspection, and data for ground inspection.
- the recording unit 11 shown in FIG. 1 is replaced by the HD.
- the operation unit 106 is configured by a remote control device or the like having various keys such as numeric keys and cursor keys, and transmits a control signal corresponding to an input operation of a user (driver or passenger) to the system control unit 110. Output.
- the audio processing unit 108 generates an audio signal under the control of the system control unit 110, and vocalizes the generated audio signal via the speaker 109.
- the navigation device 100 that is powerful in the present embodiment, for example, notifies the user of the traveling direction of the vehicle at the next intersection, or Notify the user of the situation.
- the system control unit 110 mainly includes a CPU (Central Processing Unit) and includes various input / output ports such as a GPS reception port, a key input port, and a display control port. To control general functions. At the time of such control, the system control unit 110 reads out a control program recorded in the ROMZRAM unit 111, executes each process, and temporarily holds the data being processed in the ROMZRAM unit 111.
- CPU Central Processing Unit
- the system control unit 110 performs correction processing such as map matching based on data recorded in the HD, and determines a destination point according to an input operation on the operation unit 106. Then, a route setting process for setting a route from the current vehicle position to the destination point is performed.
- the system control unit 110 controls the voice processing unit 108 to output the guidance voice
- the information display unit 107 is controlled such that an object such as an arrow indicating the vehicle traveling direction is projected on the konnoiner 14 so as to be superimposed on the scenery visually recognized by the user.
- the system control unit 110 searches the map data recorded in the HD of the HD drive 105 based on the vehicle position data supplied from the IZF unit 103, and searches for the driver corresponding to the current vehicle position. Read the map data of the zoom view map. Then, the system control unit 110 determines the road shape that is actually visually recognized by the user based on the map data. And the coordinates ( x , y) at which the road is likely to be visually recognized on the compiner 14, and displayed based on the calculated road shape and the route set in the route setting process. It determines the shape of the arrow. The arrow shape and the coordinates (x, y) group determined in this way are output to the information display unit 107, and the information display unit 107 generates image data according to the information.
- the information display unit 107 has the same configuration as the information display device I shown in FIG. 10 described above, and both the complier 14 and the dimming panel 15 constituting the information display unit 107 Is formed on the entire surface of the windshield. Further, in this application example, the angle of view of the projector 13 is set such that an image is projected on the entire area of the light control panel 15 provided on the entire surface of the windshield of the vehicle.
- the display information generating unit 12 reads out object data recorded in the HD, Based on the object data and information supplied from the system control unit 110, an object to be presented to the user, that is, image data in which an arrow for route guidance is drawn is generated.
- FIG. 5 is a diagram showing a subroutine executed at the time of route guidance.
- the system control unit 110 first obtains the vehicle position data supplied from the IZF unit 103 (Step Sal), and stores the driver's view map data corresponding to the vehicle position data. Read from HD drive 105 (step Sa2). At this time, the system control unit 110 specifies the current position of the vehicle based on the vehicle position data, and reads out the map data corresponding to the current position.
- the system control unit 110 determines the road shape and coordinates (X, y) After calculating the group (step Sa3), an object to be displayed on the combiner 14, that is, an arrow shape is determined based on the calculated road shape (step Sa4). Then, after outputting the determined object shape and coordinates (x, y) group to the information display unit 107 (step Sa5), the system control unit 110 determines whether or not the force for which the route guidance processing has been completed is determined. determining state becomes (step Sa6), whereas the process ends is determined that r y ESJ, repeats the processing of "no" and step again when determining Sal ⁇ step Sa6.
- the display information generation unit 12 of the information display unit 107 reads the object data from the HD drive 105, and based on the object data, It generates image data corresponding to an image to be projected on the compiler 14 (including an arrow for route guidance). At this time, the display information generating unit 12 determines the arrow shape based on the object shape supplied from the system control unit 110, and determines the position where the object is to be displayed based on the coordinate (x, y) group. To identify.
- the navigation apparatus 100 empowering the present embodiment, information such as arrows for route guidance is superimposed on a real scene visually recognized by the driver at the current position of the vehicle. It is possible to display, realizing virtual reality, and thereby ensuring user convenience.
- the transmittance control unit 16 controls the transmittance of the light control panel 15 based on the detection result of the external light intensity distribution detection unit 18.
- the function as a sun visor can be realized. With such a configuration, for example, it can be used as an anti-glare device in an area where light cannot be shielded by a sun visor such as headlights of an oncoming vehicle or sunrise.
- the map data corresponding to the driver's view map is configured to be created so as to match the scenery visually recognized by the driver. It is also possible to create a scene that matches the scenery to be viewed.
- the route guidance arrow is superimposed and displayed on the road visually recognized by the driver.
- the information to be superimposed and displayed is the route guidance arrow.
- the names of buildings and intersections may be displayed so as to be superimposed on actual buildings and the like visually recognized by the driver. Even in this case, it is the same as the above-mentioned first application example, except that the object from which the shape and the like are extracted by the system control unit 110 is other than the road.
- the system control unit 110 employs a configuration in which each time an arrow is displayed, the system control unit 110 calculates the shape and display position of the arrow based on the map data. It is good to keep the results in a table in advance!
- map data corresponding to the driver's view map is generated in advance so as to match the scenery visually recognized by the driver of the vehicle, and the driver's power is also visually recognized using the map data.
- the configuration that determines the shape of the road and the coordinates (x, y) at which the road is displayed on the compiler 14 is employed.
- the driver power without preparing such map data in advance may be configured to determine the shape of the road that will be visually recognized and the coordinates (x, y) at which the road is displayed on the compiler 14. It is possible.
- a camera near the user's viewpoint, recognize the scenery seen by the user based on the image captured by the camera, and display an arrow at the position where the user intersection can be seen. It is.
- a feature point on an image input from a camera is extracted.
- a parameter may be set based on the feature point, and the scene on each video may be recognized based on the parameter (for example, “A Natural Feature -Based ⁇ Object Tracking” published by AIST. Method for Wearable Augmented RealityJ).
- the information display devices I, 12, 13, and 14 can be applied to, for example, a head mount display.
- an external device IZF such as IEEE1394 is provided in the display information generating unit 12 and connected to an external device such as a navigation device via the external device iZF so that data input from the external device can be used.
- the display information generator 12 may generate the image data. Even in this case, by performing the same control as described above, it is possible to reduce the transmittance of the cell corresponding to the display area of the information to be notified to the user.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Optics & Photonics (AREA)
- Mathematical Physics (AREA)
- Automation & Control Theory (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Instrument Panels (AREA)
- Navigation (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006514788A JP4475595B2 (ja) | 2004-06-18 | 2005-06-16 | 情報表示装置、ナビゲーション装置 |
US11/629,902 US7903048B2 (en) | 2004-06-18 | 2005-06-16 | Information display apparatus and navigation apparatus |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004-181821 | 2004-06-18 | ||
JP2004181821 | 2004-06-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005124431A1 true WO2005124431A1 (ja) | 2005-12-29 |
Family
ID=35509844
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/011064 WO2005124431A1 (ja) | 2004-06-18 | 2005-06-16 | 情報表示装置、ナビゲーション装置 |
Country Status (3)
Country | Link |
---|---|
US (1) | US7903048B2 (ja) |
JP (1) | JP4475595B2 (ja) |
WO (1) | WO2005124431A1 (ja) |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008242134A (ja) * | 2007-03-28 | 2008-10-09 | Nikon Corp | 表示装置 |
JP2010521347A (ja) * | 2007-03-05 | 2010-06-24 | ザ・ボーイング・カンパニー | ヘッドアップディスプレイ用の電気的に減光可能なコンバイナ |
JP2012030611A (ja) * | 2010-07-28 | 2012-02-16 | Denso Corp | 車両用表示装置 |
JP2014092787A (ja) * | 2012-11-05 | 2014-05-19 | Exelis Inc | 画像データを選択的に遮るために可変マスクを利用する画像ディスプレイ |
WO2015050203A1 (ja) | 2013-10-03 | 2015-04-09 | 富士フイルム株式会社 | 投映像表示用ハーフミラーおよびその製造方法、ならびに投映像表示システム |
JP2017193190A (ja) * | 2016-04-18 | 2017-10-26 | ソニー株式会社 | 画像表示装置及び画像表示装置、並びに移動体 |
JPWO2017060966A1 (ja) * | 2015-10-06 | 2018-07-05 | マクセル株式会社 | ヘッドアップディスプレイ |
WO2018198868A1 (ja) * | 2017-04-27 | 2018-11-01 | パナソニックIpマネジメント株式会社 | スクリーンシステム、表示システム、スクリーンシステムの制御方法、プログラム、記憶媒体及び表示システムを備える移動体 |
WO2019073688A1 (ja) * | 2017-10-11 | 2019-04-18 | ソニー株式会社 | 画像表示装置 |
JP2019191291A (ja) * | 2018-04-20 | 2019-10-31 | 株式会社ジャパンディスプレイ | 表示装置 |
JP2020006890A (ja) * | 2018-07-11 | 2020-01-16 | アルパイン株式会社 | 視点撮影装置、視点撮影方法および評価方法 |
JP2020031358A (ja) * | 2018-08-23 | 2020-02-27 | 日本信号株式会社 | 撮像装置及び監視システム |
JPWO2020022288A1 (ja) * | 2018-07-27 | 2021-08-12 | 京セラ株式会社 | 表示装置および移動体 |
JP2021536592A (ja) * | 2018-08-31 | 2021-12-27 | マジック リープ, インコーポレイテッドMagic Leap, Inc. | 拡張現実デバイスのための空間的に分解された動的調光 |
CN114072723A (zh) * | 2019-07-05 | 2022-02-18 | 奇跃公司 | 用于减轻透视像素阵列中的伪影的几何形状 |
WO2022091398A1 (ja) * | 2020-11-01 | 2022-05-05 | 正典 伊原 | 透過率制御部を備えた表示装置 |
US11662586B2 (en) | 2020-03-06 | 2023-05-30 | Magic Leap, Inc. | Angularly selective attenuation of light transmission artifacts in wearable displays |
JP2023075061A (ja) * | 2021-11-18 | 2023-05-30 | 東陽實業廠股▲分▼有限公司 | インテリジェントな透明遮光システム |
US11693252B2 (en) | 2019-10-17 | 2023-07-04 | Magic Leap, Inc. | Attenuation of light transmission artifacts in wearable displays |
US12001110B2 (en) * | 2020-06-25 | 2024-06-04 | Magic Leap, Inc. | Tunable attenuation of light transmission artifacts in wearable displays |
US12013537B2 (en) | 2019-01-11 | 2024-06-18 | Magic Leap, Inc. | Time-multiplexed display of virtual content at various depths |
Families Citing this family (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4970302B2 (ja) * | 2008-02-14 | 2012-07-04 | 富士フイルム株式会社 | 画像処理装置,画像処理方法及び撮像装置 |
US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
CN102906623A (zh) | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | 交互式头戴目镜上的本地广告内容 |
US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
US20120249797A1 (en) | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
US9759917B2 (en) | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
US9285589B2 (en) | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
US20150309316A1 (en) | 2011-04-06 | 2015-10-29 | Microsoft Technology Licensing, Llc | Ar glasses with predictive control of external device based on event input |
US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
JP5630141B2 (ja) * | 2010-08-18 | 2014-11-26 | ブラザー工業株式会社 | 画像表示システム、画像表示システムに含まれる携帯型情報処理装置のためのコンピュータプログラム、画像表示システムに含まれるヘッドマウントディスプレイ、及び画像表示方法 |
US20120069046A1 (en) * | 2010-09-22 | 2012-03-22 | Raytheon Company | Systems and methods for displaying computer-generated images on a head mounted device |
WO2012046379A1 (ja) * | 2010-10-04 | 2012-04-12 | パナソニック株式会社 | 透過型表示装置、移動体及び制御装置 |
US9372343B2 (en) * | 2012-01-12 | 2016-06-21 | Htc Corporation | Head-up display, vehicle and controlling method of head-up display |
US8935606B2 (en) | 2012-02-29 | 2015-01-13 | Ebay Inc. | Systems and methods for providing a user interface with grid view |
DE102012204303B4 (de) | 2012-03-19 | 2022-07-14 | Bayerische Motoren Werke Aktiengesellschaft | Helligkeitssteuerung für ein Head-Up-Display |
FR2993676B1 (fr) * | 2012-07-18 | 2015-03-27 | Valeo Etudes Electroniques | Dispositif et procede d'emission d'un faisceaux lumineux destine a former une image, systeme de projection et afficheur utilisant ledit dispositif |
JP6082272B2 (ja) * | 2013-02-25 | 2017-02-15 | 東京エレクトロン株式会社 | 支援情報表示方法、基板処理装置の保守支援方法、支援情報表示制御装置、基板処理システム及びプログラム |
US20140253605A1 (en) * | 2013-03-05 | 2014-09-11 | John N. Border | Controlling brightness of a displayed image |
CA2960492A1 (en) * | 2013-05-10 | 2014-11-13 | Magna International Inc. | Projector utilizing opaque/ transparent projection screen |
FR3011090B1 (fr) * | 2013-09-26 | 2016-12-23 | Valeo Vision | Lunettes a affichage de donnees munies d'un ecran anti-eblouissement |
TWI541543B (zh) | 2013-10-21 | 2016-07-11 | 財團法人工業技術研究院 | 分光模組及應用其之投影裝置 |
CN104777613B (zh) * | 2014-01-10 | 2017-11-28 | 联想(北京)有限公司 | 眼镜显示装置及增强现实显示方法 |
US10033992B1 (en) * | 2014-09-09 | 2018-07-24 | Google Llc | Generating a 3D video of an event using crowd sourced data |
TWI578085B (zh) | 2014-12-24 | 2017-04-11 | 財團法人工業技術研究院 | 投影裝置 |
JPWO2017056953A1 (ja) * | 2015-10-02 | 2018-07-26 | 株式会社リコー | 表示装置 |
EP3395601A4 (en) * | 2015-12-24 | 2019-09-11 | KYOCERA Corporation | HIGH-HEAD VEHICLE DISPLAY SYSTEM |
CN105607260B (zh) * | 2016-03-03 | 2019-06-28 | 北京小鸟看看科技有限公司 | 一种画面显示方法和装置及一种头戴设备 |
US10895741B2 (en) | 2017-10-03 | 2021-01-19 | Industrial Technology Research Institute | Ultra-wide head-up display system and display method thereof |
WO2019209169A1 (en) * | 2018-04-28 | 2019-10-31 | Unitech Mechatronics Pte Ltd | Precise positioning system |
DE102018206656B4 (de) * | 2018-04-30 | 2024-03-21 | Audi Ag | Anzeigevorrichtung für ein Kraftfahrzeug und Kraftfahrzeug |
WO2020149646A1 (en) * | 2019-01-17 | 2020-07-23 | Samsung Electronics Co., Ltd. | Method of acquiring outside luminance using camera sensor and electronic device applying the method |
US10970902B2 (en) * | 2019-03-26 | 2021-04-06 | At&T Intellectual Property I, L.P. | Allocating and extrapolating data for augmented reality for 6G or other next generation network |
CN110286509A (zh) * | 2019-05-08 | 2019-09-27 | 江苏集萃有机光电技术研究所有限公司 | 透明显示装置及透明显示方法 |
US20220020319A1 (en) * | 2019-08-27 | 2022-01-20 | Lg Electronics Inc. | Display apparatus and operation method thereof |
CN111624808A (zh) * | 2020-06-30 | 2020-09-04 | 京东方科技集团股份有限公司 | 透明显示面板及其控制方法和装置、显示装置和显示系统 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01219883A (ja) * | 1988-02-29 | 1989-09-01 | Meitetsuku:Kk | 車載用ナビゲーションシステムの表示装置 |
JPH0876050A (ja) * | 1994-09-05 | 1996-03-22 | Fujitsu Ten Ltd | 車両用表示装置 |
JPH08160340A (ja) * | 1994-12-05 | 1996-06-21 | Canon Inc | 画像観察装置 |
JP2003149730A (ja) * | 2001-11-09 | 2003-05-21 | Seiko Epson Corp | 投射型表示装置 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4839475A (en) | 1988-04-29 | 1989-06-13 | Limitorque Corporation | Modular switch assembly having wiping contacts |
JPH0715789Y2 (ja) * | 1988-07-15 | 1995-04-12 | 矢崎総業株式会社 | 車両用表示装置 |
JPH05104980A (ja) | 1991-10-17 | 1993-04-27 | Fujitsu Ltd | 車両用表示装置 |
US5305012A (en) * | 1992-04-15 | 1994-04-19 | Reveo, Inc. | Intelligent electro-optical system and method for automatic glare reduction |
US7339149B1 (en) * | 1993-02-26 | 2008-03-04 | Donnelly Corporation | Vehicle headlight control using imaging sensor |
US5931888A (en) * | 1994-09-22 | 1999-08-03 | Aisin Aw Co., Ltd. | Navigation system for vehicles with alternative route searching capabilities |
JPH08253059A (ja) | 1995-03-17 | 1996-10-01 | Honda Motor Co Ltd | 車両用運転支援システム |
WO1998023918A1 (fr) * | 1996-11-25 | 1998-06-04 | Toyota Jidosha Kabushiki Kaisha | Dispositif de navigation pour vehicule et preparation de donnees de configuration de la route utilisees avec ce dispositif |
JP4588243B2 (ja) * | 2001-04-12 | 2010-11-24 | 三菱電機株式会社 | ナビゲーション装置、ナビゲーション方法及びナビゲーションプログラム |
AU2003214595A1 (en) * | 2002-03-07 | 2003-09-16 | Yechezkal Evan Spero | Enhanced vision for driving |
JP2003341383A (ja) * | 2002-05-29 | 2003-12-03 | Denso Corp | 車両用表示装置 |
JP4069378B2 (ja) * | 2002-12-18 | 2008-04-02 | アイシン・エィ・ダブリュ株式会社 | ナビゲーション装置並びに該装置用プログラム及び記録媒体 |
-
2005
- 2005-06-16 US US11/629,902 patent/US7903048B2/en not_active Expired - Fee Related
- 2005-06-16 WO PCT/JP2005/011064 patent/WO2005124431A1/ja active Application Filing
- 2005-06-16 JP JP2006514788A patent/JP4475595B2/ja not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01219883A (ja) * | 1988-02-29 | 1989-09-01 | Meitetsuku:Kk | 車載用ナビゲーションシステムの表示装置 |
JPH0876050A (ja) * | 1994-09-05 | 1996-03-22 | Fujitsu Ten Ltd | 車両用表示装置 |
JPH08160340A (ja) * | 1994-12-05 | 1996-06-21 | Canon Inc | 画像観察装置 |
JP2003149730A (ja) * | 2001-11-09 | 2003-05-21 | Seiko Epson Corp | 投射型表示装置 |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2010521347A (ja) * | 2007-03-05 | 2010-06-24 | ザ・ボーイング・カンパニー | ヘッドアップディスプレイ用の電気的に減光可能なコンバイナ |
JP2008242134A (ja) * | 2007-03-28 | 2008-10-09 | Nikon Corp | 表示装置 |
JP2012030611A (ja) * | 2010-07-28 | 2012-02-16 | Denso Corp | 車両用表示装置 |
JP2014092787A (ja) * | 2012-11-05 | 2014-05-19 | Exelis Inc | 画像データを選択的に遮るために可変マスクを利用する画像ディスプレイ |
WO2015050203A1 (ja) | 2013-10-03 | 2015-04-09 | 富士フイルム株式会社 | 投映像表示用ハーフミラーおよびその製造方法、ならびに投映像表示システム |
JPWO2017060966A1 (ja) * | 2015-10-06 | 2018-07-05 | マクセル株式会社 | ヘッドアップディスプレイ |
US11034296B2 (en) | 2016-04-18 | 2021-06-15 | Sony Corporation | Image display device, image display method, and moving object |
JP2017193190A (ja) * | 2016-04-18 | 2017-10-26 | ソニー株式会社 | 画像表示装置及び画像表示装置、並びに移動体 |
WO2018198868A1 (ja) * | 2017-04-27 | 2018-11-01 | パナソニックIpマネジメント株式会社 | スクリーンシステム、表示システム、スクリーンシステムの制御方法、プログラム、記憶媒体及び表示システムを備える移動体 |
JP2018189956A (ja) * | 2017-04-27 | 2018-11-29 | パナソニックIpマネジメント株式会社 | スクリーンシステム、表示システム、スクリーンシステムの制御方法、プログラム、及び表示システムを備える移動体 |
US11187898B2 (en) | 2017-10-11 | 2021-11-30 | Sony Corporation | Image display apparatus |
WO2019073688A1 (ja) * | 2017-10-11 | 2019-04-18 | ソニー株式会社 | 画像表示装置 |
JP2019191291A (ja) * | 2018-04-20 | 2019-10-31 | 株式会社ジャパンディスプレイ | 表示装置 |
JP7027238B2 (ja) | 2018-04-20 | 2022-03-01 | 株式会社ジャパンディスプレイ | 表示装置 |
JP2020006890A (ja) * | 2018-07-11 | 2020-01-16 | アルパイン株式会社 | 視点撮影装置、視点撮影方法および評価方法 |
JP7007062B2 (ja) | 2018-07-11 | 2022-02-10 | アルパイン株式会社 | 視点撮影装置、視点撮影方法および評価方法 |
JPWO2020022288A1 (ja) * | 2018-07-27 | 2021-08-12 | 京セラ株式会社 | 表示装置および移動体 |
JP2022119758A (ja) * | 2018-07-27 | 2022-08-17 | 京セラ株式会社 | 表示装置および移動体 |
JP7332747B2 (ja) | 2018-07-27 | 2023-08-23 | 京セラ株式会社 | 表示装置および移動体 |
US11276360B2 (en) | 2018-07-27 | 2022-03-15 | Kyocera Corporation | Display device and mobile body |
JP2020031358A (ja) * | 2018-08-23 | 2020-02-27 | 日本信号株式会社 | 撮像装置及び監視システム |
JP2021536592A (ja) * | 2018-08-31 | 2021-12-27 | マジック リープ, インコーポレイテッドMagic Leap, Inc. | 拡張現実デバイスのための空間的に分解された動的調光 |
US12073509B2 (en) | 2018-08-31 | 2024-08-27 | Magic Leap, Inc. | Spatially-resolved dynamic dimming for augmented reality device |
US12013537B2 (en) | 2019-01-11 | 2024-06-18 | Magic Leap, Inc. | Time-multiplexed display of virtual content at various depths |
JP2022538583A (ja) * | 2019-07-05 | 2022-09-05 | マジック リープ, インコーポレイテッド | シースルーピクセルアレイ内のアーチファクトを軽減するための幾何学形状 |
CN114072723A (zh) * | 2019-07-05 | 2022-02-18 | 奇跃公司 | 用于减轻透视像素阵列中的伪影的几何形状 |
US20230324744A1 (en) * | 2019-07-05 | 2023-10-12 | Magic Leap, Inc. | Geometries for mitigating artifacts in see-through pixel arrays |
JP7476244B2 (ja) | 2019-07-05 | 2024-04-30 | マジック リープ, インコーポレイテッド | シースルーピクセルアレイ内のアーチファクトを軽減するための幾何学形状 |
US11693252B2 (en) | 2019-10-17 | 2023-07-04 | Magic Leap, Inc. | Attenuation of light transmission artifacts in wearable displays |
US11662586B2 (en) | 2020-03-06 | 2023-05-30 | Magic Leap, Inc. | Angularly selective attenuation of light transmission artifacts in wearable displays |
US12001110B2 (en) * | 2020-06-25 | 2024-06-04 | Magic Leap, Inc. | Tunable attenuation of light transmission artifacts in wearable displays |
WO2022091398A1 (ja) * | 2020-11-01 | 2022-05-05 | 正典 伊原 | 透過率制御部を備えた表示装置 |
JP2023075061A (ja) * | 2021-11-18 | 2023-05-30 | 東陽實業廠股▲分▼有限公司 | インテリジェントな透明遮光システム |
JP7499824B2 (ja) | 2021-11-18 | 2024-06-14 | 東陽實業廠股▲分▼有限公司 | インテリジェントな透明遮光システム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2005124431A1 (ja) | 2008-07-31 |
US20080048932A1 (en) | 2008-02-28 |
US7903048B2 (en) | 2011-03-08 |
JP4475595B2 (ja) | 2010-06-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2005124431A1 (ja) | 情報表示装置、ナビゲーション装置 | |
EP3413118B1 (en) | Information providing method and information providing vehicle therefor | |
US8536995B2 (en) | Information display apparatus and information display method | |
US20170352277A1 (en) | Transparent display apparatus for displaying information of danger element, and method thereof | |
RU2677122C1 (ru) | Устройство отображения транспортного средства | |
US20170309257A1 (en) | Display control device | |
US20120224060A1 (en) | Reducing Driver Distraction Using a Heads-Up Display | |
CN109584596A (zh) | 车辆驾驶提示方法及装置 | |
US20100088015A1 (en) | Navigation System Having Simplified Display Mode and Display Method Thereof | |
JP2005182306A (ja) | 車両用表示装置 | |
US8767066B2 (en) | Information provision device using retro-reflecting article | |
JP2009227018A (ja) | 車両用防眩装置 | |
WO2019097762A1 (ja) | 重畳画像表示装置及びコンピュータプログラム | |
JP2015069111A (ja) | 画像表示装置、スクリーンおよび画像表示方法 | |
JP6232691B2 (ja) | 車両用表示制御装置、車両用表示装置及び車両用表示制御方法 | |
CN101726313A (zh) | 具有简化显示模式的导航系统及其显示方法 | |
JP2021039085A (ja) | 重畳画像表示装置、重畳画像描画方法及びコンピュータプログラム | |
CN114667474A (zh) | 用于车辆的智能玻璃 | |
KR20170135522A (ko) | 차량용 제어장치 및 그것의 제어방법 | |
US20210116710A1 (en) | Vehicular display device | |
JP6780960B2 (ja) | 画像表示装置 | |
JP2006036166A (ja) | 車両用表示装置 | |
JP2005070231A (ja) | 車両における表示方法 | |
JP7400356B2 (ja) | 重畳画像表示装置及びコンピュータプログラム | |
JP2023083117A (ja) | ヘッドアップディスプレイ装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DPEN | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006514788 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11629902 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Country of ref document: DE |
|
122 | Ep: pct application non-entry in european phase | ||
WWP | Wipo information: published in national office |
Ref document number: 11629902 Country of ref document: US |