CN109493321A - A kind of vehicle-mounted HUD visual system parallax calculation method - Google Patents
A kind of vehicle-mounted HUD visual system parallax calculation method Download PDFInfo
- Publication number
- CN109493321A CN109493321A CN201811199655.0A CN201811199655A CN109493321A CN 109493321 A CN109493321 A CN 109493321A CN 201811199655 A CN201811199655 A CN 201811199655A CN 109493321 A CN109493321 A CN 109493321A
- Authority
- CN
- China
- Prior art keywords
- module
- visual field
- parallax
- visual system
- visual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/97—Determining parameters from multiple pictures
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Eye Examination Apparatus (AREA)
Abstract
The invention belongs to optics field of display technology, are related to a kind of visual system disparity calculation algorithm.This algorithm is mainly by interactive interface setting module, calculate grid dividing module, ray tracing direction determining module, disparity computation convergence rate control module, disparity computation precision controlling module, Data induction analysis module, seven part such as Graphical output module composition, wherein ray tracing direction determining module determines ray tracing direction according to trial-and-error method, then parallax convergence rate control module is relied on to determine single calculation convergence rate apart from starting point distance according to target point in the direction, it is calculated by successive ignition, tracing point is gradually drawn close to target point, target point is determined eventually by parallax precision controlling module.The present invention is as a kind of visual system disparity calculation algorithm, the one-touch simulation analysis and Graphical output of entire field parallax within the scope of the entire emergent pupil of visual system can be achieved, disparity computation is high-efficient, computational accuracy is adjustable, calculates in real time for visual system parallax and provides a kind of efficient, high-precision simulation analysis means.
Description
Technical field
The invention belongs to optics field of display technology, are related to a kind of parallax calculation method.Present invention can apply to head-up display/
In virtual reality/augmented reality (VR/AR) system that head-mounted display (HUD/HMD) etc. requires image-forming range.
Background technique
In aircraft high-speed flight, pilot, which needs to bow repeatedly, watches the information such as lower aobvious upper air speed, the aspect shown, is
Reduce pilot reaction time, military aircraft equips head-up display system (HUD), will can collimate for information about and image at the moment
Infinite point, pilot can see the information clearly without bowing while watching distant object, promote battlefield production rate, are same
When see the relevant information of infinite point target and display clearly, the two necessarily is in the field depth of eyes, this is just to HUD/HMD
Display image-forming range, that is, parallax claims.
In vehicle high-speed row driving process, driver need to watch frequently the speed that automobile instrument panel shows, engine speed,
The information of vehicles such as water temperature.It is reduction driver to the reaction time of urgent emergency case, it is aobvious that many manufacturers release vehicle-mounted head-up
Show system, by relevant information Collimation Display at 5m-8m at the moment, this also claims to image-forming range, that is, parallax of HUD.
In addition, larger parallax, observer head will be clearly felt that display picture when mobile if it exists for HUD/HMD system
Bounce, easily causes visual fatigue when watching when long, and when different visual field parallax difference are larger will appear the out-of-flatness of display picture
Phenomenon, i.e., different visual field characters are at different distance, therefore must be controlled to parallax in HUD/HMD system design process
System carries out detailed simulation analysis to parallax after the design is completed.Existing parallax multi-pass is excessively to the indirect of HUD/HMD system disc of confusion
It is calculated, not intuitively, and characteristic point can only be calculated, it is difficult to meet actual demand.
Summary of the invention
Parallax control when the purpose of the present invention is for the design of HUD/HMD system provides means, and complete after the completion of designing for system
The simulation analysis of different visual field parallaxes provides a kind of high efficiency, high-precision simulating analysis within the scope of emergent pupil, grasps comprehensively
The parallax distribution situation of HUD/HMD system entire field within the scope of entire emergent pupil.
To achieve the above object, the technical scheme is that a kind of visual system parallax calculation method, including net is calculated
Lattice division module, ray tracing direction determining module, disparity computation convergence rate control module, disparity computation precision controlling mould
Block, Data induction analysis module, Graphical output module.
The calculating grid dividing module carries out visual field, emergent pupil normalization first, and according to the stepping of Disparity Analysis requirement pair
Visual field, emergent pupil after normalization are divided.
The ray tracing direction determining module is according to the light gradient symbol that delivery altitude changes with visual field, eye position in image planes
Number determine the step direction of visual field or eye position;
The disparity computation convergence rate control module is according to the difference of light delivery altitude and ideal position coordinate in image planes
It is worth and determines step size, makes a little to restrain to ideal position coordinate rapidly wait calculate;
The disparity computation precision controlling module determines exit criteria according to disparity computation required precision;
The Data induction analysis module is to output data according to eye position (x, y) and visual field (θx, θy) change and concluded, it deposits
It is stored in a four-dimensional array;
The Graphical output module calls the data being stored in above-mentioned four-dimensional array according to actual analysis demand, and
Draw trend chart of the parallax with eye position or parallax with visual field;
The invention has the following advantages that 1, disparity computation fast convergence rate of the present invention, certainly by starting point distance objective point distance
It is dynamic to adjust single calculation step-length, farthest reduce the number of iterations, improves computational efficiency;2. disparity computation precision is adjustable, this
Algorithm includes disparity computation precision controlling module, can be according to the change condition of convergence is required, so as to adjust disparity computation precision, highest
Computational accuracy is up to 0.1 ';3 uses of the invention can grasp the visual systems such as HMD, HUD in entire emergent pupil range comprehensively
The parallax distributed intelligence of interior entire field, and a kind of effective means is provided for optical system parallax control.
Detailed description of the invention
Fig. 1 is ray tracing walking direction module flow diagram;
Fig. 2 is disparity computation convergence rate determination module flow chart.
Fig. 3 is visual system numerical computation algorithm flow chart.
Wherein, interactive interface setting module 1, calculating grid dividing module 2, ray tracing direction determining module 3, disparity computation are received
Hold back rate control module 4, disparity computation precision controlling module 5, Data induction analysis module 6 and Graphical output module 7.
Specific embodiment
It is the technological means realizing point-by-point high-precision disparity computation and being taken for the present invention is further explained, below in conjunction with attached drawing grade
Preferred embodiment, to a kind of specific steps of visual system parallax numerical computation method proposed according to the present invention, feature,
Function elaborates.
A kind of the step of visual system parallax numerical computation method progress point-by-point disparity computation, is as shown in figure 3, include interaction
Interface setting module 1 calculates grid dividing module 2, ray tracing direction determining module 3, disparity computation convergence rate control mould
Seven parts such as block 4, disparity computation precision controlling module 5, Data induction analysis module 6, Graphical output module 7.
Interactive interface setting module 1 includes the setting of the options such as visual field, eye box, computational accuracy, and user can be according to actual analysis precision
It needs to be defined.
The calculating grid dividing module 2 carries out visual field, emergent pupil normalization first, and according to the stepping of Disparity Analysis requirement pair
Visual field, emergent pupil after normalization are divided.
Ray tracing direction determining module 3 utilizes trial-and-error method, judges to the convergent ray tracing direction of target point.
First, any visual field (θ at trace design eye positionx, θy) chief ray (reference ray), wherein θx、θyRespectively orientation side
To, pitch orientation field angle, and record image height (x of the chief ray in image planes0, y0)
Then orientation visual field is finely tuned along orientation positive direction, by position of eye point grid interval tracking optical system emergent pupil model one by one
Enclose (θ at interior non-design eye positionx+ ε, θy)、(θx+ 2 ε, θy) field rays, and x direction of the light in image planes is recorded respectively
Image height x1, x2, ε is visual field grid interval;
If | x2-x0|≤|x1-x0|, then along the azimuth direction trace, if | x2-x0| > | x1-x0|, then along the azimuth direction
Opposite direction trace;
Then pitching visual field is finely tuned along pitching positive direction, opsition dependent grid dividing step-length is one by one within the scope of tracking optical system emergent pupil
(θ at non-design eye positionx, θy+ε)、(θx, θy+ 2 ε) field rays, and y direction image height of the light in image planes is recorded respectively
y1, y2;
If | y2-y0|≤|y1-y0|, then along the pitch orientation trace, if | y2-y0| > | y1-y0|, then along the pitch orientation
Opposite direction trace;
According to light, the difference of delivery altitude and ideal position coordinate in image planes determines disparity computation convergence rate control module 4
Step size makes a little to restrain to ideal position coordinate rapidly wait calculate;
X-direction, Y first by chief ray at light at calculating position within the scope of optical system emergent pupil and design eye position in image planes
Direction image height difference is divided into multiple regions, determines single calculation step-length according to image height difference;In the present embodiment, C=0.1 ', B=are taken
1 ', A=10 '.
When | x1-x0|/f > 10', orientation visual field material calculation are 10 ', work as 10'> | x1-x0When |/f >=1', orientation visual field is calculated
Step-length is 1 ', works as 1'> | x1-x0When |/f >=0.1', orientation visual field material calculation is 0.1 ', and wherein f is optical system focal length;
When | y1-y0|/f > 10', pitching visual field material calculation are 10 ', work as 10'> | x1-x0When |/f >=1', pitching visual field is calculated
Step-length is 1 ', works as 1'> | x1-x0When |/f >=0.1', pitching visual field material calculation is 0.1 ';
Disparity computation precision controlling module 5 determines exit criteria according to disparity computation required precision;
Calculate within the scope of emergent pupil whether light and image height difference of the chief ray in image planes at design eye position are full at calculating position first
The sufficient condition of convergence, such asThen meet the condition of convergence, calculate and terminate, wherein τ is the parallaxometer
Calculate precision;
Then it reads azimuth direction and presses 10 ', 1 ', 0.1 ' step size computation number ν1、ν2、ν3, computer azimuth direction parallax Δ α=10
ν1+ν2+0.1ν3;
Then it reads pitch orientation and presses 10 ', 1 ', 0.1 ' step size computation number κ1、κ2、κ3, calculate pitch orientation parallax Δ β=10
κ1+κ2+0.1κ3;
Data induction analysis module 6 is to output data according to eye position (x, y) and visual field (θx, θy) change and concluded, it is stored in
In one four-dimensional array;
Graphical output module 7 carries UGR drawing option using CODEV software and draws different visual fields at Optical System Design eye position
Parallax distribution curve.Parallax information of a certain visual field within the scope of emergent pupil at different positions of eye point is transferred, a certain visual field is drawn
Change parallax distribution curve with position of eye point.
Claims (8)
1. a kind of vehicle-mounted HUD visual system parallax calculation method, which is characterized in that be based on the macro extension function of optical design software
Can, using retrodirected ray tracking method, calculating grid dividing is carried out to visual field, emergent pupil, specific implementation includes that interactive interface is set
Module 1 calculates grid dividing module 2, ray tracing direction determining module 3, disparity computation convergence rate control module 4, parallax
Computational accuracy control module 5, Data induction analysis module 6 and Graphical output module 7, it may be achieved visual system entirely goes out
The simulation calculation of any visual field parallax within the scope of pupil;
Wherein, the disparity computation convergence rate control module 4 is poor by the image height of tracing point and target point in image planesMultiple regions are divided into, different material calculations is respectively corresponded, thus most
Fast speed forces tracing point to be drawn close to target point, improves computational efficiency;
The disparity computation precision controlling module 5 determines exit criteria according to disparity computation required precision.
2. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that it is special
Sign is that the disparity computation convergence rate control module 4 is implemented as follows:
X-direction, Y first by chief ray at light at calculating position within the scope of optical system emergent pupil and design eye position in image planes
Direction image height difference is divided into multiple regions, determines single calculation step-length according to image height difference;
When | x1-x0|/f > A, orientation visual field material calculation be A, when B≤| x1-x0When |/f < A, orientation visual field material calculation is B,
When C≤| x1-x0When |/f < B, orientation visual field material calculation is C;
When | x1-x0|/f > A, pitching visual field material calculation be A, when B≤| x1-x0When |/f < A, pitching visual field material calculation is B,
When C≤| x1-x0When |/f < B, orientation visual field material calculation is C;
Wherein, f is optical system focal length, and C is according to actual requirement value, less than the computational accuracy requirement of calculation method, B > 10C, A
>10B。
3. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that it is special
Sign is that the disparity computation precision controlling module 5 is implemented as follows:
Calculate within the scope of emergent pupil whether light and image height difference of the chief ray in image planes at design eye position are full at calculating position first
The sufficient condition of convergence, such asThen meet the condition of convergence, calculate and terminate, wherein τ is the parallaxometer
Calculate precision;
Then it reads azimuth direction and presses A, B, C step size computation number ν1、ν2、ν3, computer azimuth direction parallax △ α=A ν1+Bν2+C
ν3;
Then it reads pitch orientation and presses A, B, C step size computation number k1、k2、k3, calculate pitch orientation parallax △ α=A k1+B k2+
C k3。
4. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that it is special
Sign is that the calculating grid dividing module 2 carries out visual field, emergent pupil normalization first, and according to Disparity Analysis
Stepping requires to divide the visual field after normalization, emergent pupil.
5. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that described
Ray tracing direction determining module 3 according to light in image planes delivery altitude with visual field, eye position change gradient signs determine
The step direction of visual field or eye position.
6. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 5, which is characterized in that described
Ray tracing direction determining module 3 be implemented as follows:
First, any visual field (θ at trace design eye positionx, θy) chief ray (reference ray), wherein θx、θyRespectively azimuth direction,
Pitch orientation field angle, and record image height (x of the chief ray in image planes0, y0);
Then orientation visual field is finely tuned along orientation positive direction, by position of eye point grid interval tracking optical system emergent pupil model one by one
Enclose (θ at interior non-design eye positionx+ ε, θy)、(θx+ 2 ε, θy) field rays, and x direction of the light in image planes is recorded respectively
Image height x1, x2, ε is visual field grid interval;
If | x2-x0|≤|x1-x0|, then along the azimuth direction trace, if | x2-x0| > | x1-x0|, then along the azimuth direction
Opposite direction trace;
Then pitching visual field is finely tuned along pitching positive direction, opsition dependent grid dividing step-length is one by one within the scope of tracking optical system emergent pupil
(θ at non-design eye positionx, θy+ε)、(θx, θy+ 2 ε) field rays, and y direction image height of the light in image planes is recorded respectively
y1, y2;
If | y2-y0|≤|y1-y0|, then along the pitch orientation trace, if | y2-y0| > | y1-y0|, then along the pitch orientation
Opposite direction trace.
7. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that described
Data induction analysis module 6 to output data according to eye position (x, y) and visual field (θx, θy) change and concluded, it is stored in one
In four-dimensional array.
8. a kind of vehicle-mounted HUD visual system parallax calculation method according to claim 1, which is characterized in that described
Graphical output module 7 data being stored in above-mentioned four-dimensional array are called according to actual analysis demand, and draw parallax with
Eye position or parallax with visual field trend chart.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811199655.0A CN109493321B (en) | 2018-10-16 | 2018-10-16 | Parallax calculation method for vehicle-mounted HUD visual optical system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811199655.0A CN109493321B (en) | 2018-10-16 | 2018-10-16 | Parallax calculation method for vehicle-mounted HUD visual optical system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109493321A true CN109493321A (en) | 2019-03-19 |
CN109493321B CN109493321B (en) | 2021-11-12 |
Family
ID=65690342
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811199655.0A Active CN109493321B (en) | 2018-10-16 | 2018-10-16 | Parallax calculation method for vehicle-mounted HUD visual optical system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109493321B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110349078A (en) * | 2019-05-24 | 2019-10-18 | 杨巧雪 | A kind of AR pattern algorithm shown for HUD normotopia outdoor scene |
CN111323209A (en) * | 2020-03-13 | 2020-06-23 | 江苏泽景汽车电子股份有限公司 | HUD stray light automatic test system and test method |
CN113313656A (en) * | 2020-11-18 | 2021-08-27 | 江苏泽景汽车电子股份有限公司 | Distortion correction method suitable for HUD upper, middle and lower eye boxes |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7126605B1 (en) * | 2001-07-03 | 2006-10-24 | Munshi Aaftab A | Method and apparatus for implementing level of detail with ray tracing |
CN101963703A (en) * | 2009-07-22 | 2011-02-02 | 索尼公司 | Image display device and optical devices |
CN102402005A (en) * | 2011-12-06 | 2012-04-04 | 北京理工大学 | Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces |
CN103530518A (en) * | 2013-10-14 | 2014-01-22 | 中国科学院电工研究所 | Computing method of lighting surface light condensation energy flux density distribution of solar tower type electric generation thermal absorber |
CN105547649A (en) * | 2015-12-05 | 2016-05-04 | 中国航空工业集团公司洛阳电光设备研究所 | Short-wave infrared lens stray radiation detection method |
-
2018
- 2018-10-16 CN CN201811199655.0A patent/CN109493321B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7126605B1 (en) * | 2001-07-03 | 2006-10-24 | Munshi Aaftab A | Method and apparatus for implementing level of detail with ray tracing |
CN101963703A (en) * | 2009-07-22 | 2011-02-02 | 索尼公司 | Image display device and optical devices |
CN102402005A (en) * | 2011-12-06 | 2012-04-04 | 北京理工大学 | Bifocal-surface monocular stereo helmet-mounted display device with free-form surfaces |
CN103530518A (en) * | 2013-10-14 | 2014-01-22 | 中国科学院电工研究所 | Computing method of lighting surface light condensation energy flux density distribution of solar tower type electric generation thermal absorber |
CN105547649A (en) * | 2015-12-05 | 2016-05-04 | 中国航空工业集团公司洛阳电光设备研究所 | Short-wave infrared lens stray radiation detection method |
Non-Patent Citations (3)
Title |
---|
B.VANDEGHINSTE ET AL.: "Characterizing the parallax error in multi-pinhole micro-spect reconstruction", 《2011 IEEE NUCLEAR SCIENCE SYMPOSIUM CONFERENCE RECORD》 * |
SHUJUN XING ET AL.: "Groud autostereoscopic display with new type non-uniform barrier", 《OPTIK》 * |
蔡辉跃等: "基于光线跟踪的虚拟场景立体图像对视差研究", 《南京工程学院学报(自然科学版)》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110349078A (en) * | 2019-05-24 | 2019-10-18 | 杨巧雪 | A kind of AR pattern algorithm shown for HUD normotopia outdoor scene |
CN110349078B (en) * | 2019-05-24 | 2022-07-15 | 深圳市锐思华创技术有限公司 | AR (augmented reality) graphic method for HUD (head Up display) orthostatic live-action display |
CN111323209A (en) * | 2020-03-13 | 2020-06-23 | 江苏泽景汽车电子股份有限公司 | HUD stray light automatic test system and test method |
CN113313656A (en) * | 2020-11-18 | 2021-08-27 | 江苏泽景汽车电子股份有限公司 | Distortion correction method suitable for HUD upper, middle and lower eye boxes |
Also Published As
Publication number | Publication date |
---|---|
CN109493321B (en) | 2021-11-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109493321A (en) | A kind of vehicle-mounted HUD visual system parallax calculation method | |
AU2020204550B2 (en) | Virtual and augmented reality systems and methods having improved diffractive grating structures | |
US11231540B2 (en) | Anisotropically formed diffraction grating device | |
US11307414B2 (en) | Near-eye display system | |
US20210191038A1 (en) | Planar Optical Waveguide Based on Two-Dimensional Optical Grating | |
US7961117B1 (en) | System, module, and method for creating a variable FOV image presented on a HUD combiner unit | |
US9915826B2 (en) | Virtual and augmented reality systems and methods having improved diffractive grating structures | |
US9244281B1 (en) | Display system and method using a detached combiner | |
US11693238B2 (en) | Optical arrangements including Fresnel lens elements | |
AU2021203240A1 (en) | Improved manufacturing for virtual and augmented reality systems and components | |
CN112925100A (en) | Optical system | |
KR20180018766A (en) | Virtual and augmented reality systems and methods | |
CN111016785A (en) | Head-up display system adjusting method based on human eye position | |
CN104280885A (en) | Large-exit-pupil holographic waveguide glasses system | |
US9581819B1 (en) | See-through augmented reality system | |
US10895671B1 (en) | Diffraction grating with a variable refractive index using ion implantation | |
KR102547822B1 (en) | Head up display system and optical element for head up display | |
CN113168012A (en) | Volume display device representing virtual image and method thereof | |
US9638836B1 (en) | Lenses having astigmatism correcting inside reflective surface | |
CN103631009B (en) | Image segmentation type virtual image display device | |
Sisodia et al. | Advanced helmet mounted display (AHMD) | |
US9864193B2 (en) | Compact head-up display having a large exit pupil | |
US20160127718A1 (en) | Method and System for Stereoscopic Simulation of a Performance of a Head-Up Display (HUD) | |
Wood et al. | Head-up display | |
Collinson | Displays and man–machine interaction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |