CN111435199A - Space display scheme - Google Patents
Space display scheme Download PDFInfo
- Publication number
- CN111435199A CN111435199A CN201811600521.5A CN201811600521A CN111435199A CN 111435199 A CN111435199 A CN 111435199A CN 201811600521 A CN201811600521 A CN 201811600521A CN 111435199 A CN111435199 A CN 111435199A
- Authority
- CN
- China
- Prior art keywords
- micro
- space
- angle observation
- dimensional
- viewing angle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
The invention discloses a space display scheme which can be used for near-eye display, does not need complex three-dimensional graph calculation, and enables an observer to directly see a three-dimensional model, and has free visual angle and free far-near depth during observation.
Description
Technical Field
The invention relates to a three-dimensional display scheme, in particular to a space display scheme.
Background
The existing three-dimensional display scheme generates a specific visual angle image through calculation, so that the calculation amount is large, the visual angle is not free, and the display depth is also not free.
Disclosure of Invention
The invention provides a space display scheme which can be used for near-eye display glasses, does not need complex three-dimensional graph calculation, and can freely observe and display a three-dimensional model at multiple visual angles.
The invention is realized by the following technical scheme:
a spatial display scheme includes a micro-space section, a micro-viewing angle observation section, and a large-viewing angle observation section.
The micro-space part displays a three-dimensional object model with a micro scale, and the micro-space model can be generated by a micro-space generation system and can also be an actual luminous three-dimensional object.
The micro-viewing angle observation section is an input section of the near-eye display in the near-eye display eyeglass scheme, and the large-viewing angle observation section is an output section of the near-eye display in the near-eye display eyeglass scheme.
The micro-space generating system generates a micro-space by exchanging time for space, and comprises an image source, a lens group and a micro-space control part.
The principle of the invention is as follows:
the micro-space part is an observable three-dimensional space with micro-scale, the micro-visual angle observation part can observe the micro-space part at a specific visual angle, and the model for observing the micro-space at double visual angles can be realized by arranging the two micro-visual angle observation parts.
The micro-visual angle observation part transmits and expands light to the large-visual angle observation part, and the target space model can be observed in the large-visual angle observation part.
The principle of the micro-space generation system is as follows: the image source forms an imaging surface at a certain distance through the lens group, the micro-space control part adjusts the distance of the imaging surface back and forth at a high speed to form a scanning imaging surface, and the scanning imaging surfaces at different moments in a period of time are combined together to form a three-dimensional micro-space.
The micro-space control part fills the pixels of the three-dimensional model to be displayed into the micro-space formed by scanning at an accurate time point, and because the scanning speed is high, the lighting and closing speeds of the spatial pixels are also high, and the micro-space control part can display the model in the whole three-dimensional micro-space range.
The farther the scan imaging plane is from the micro-viewing angle observation portion, the larger the spatial span of the scan imaging plane is set, and the larger the space occupied by the corresponding single pixel on the plane is, so that a larger range of micro-space can be generated with less scanning time.
When the micro-space is generated by adopting a time-sharing scanning scheme, the pixels at one time point cannot be shielded by the pixels at another time point, so that the observed model is semitransparent.
The invention has the advantages that the three-dimensional model can be observed in multiple visual angles without complex three-dimensional calculation, the displayed three-dimensional model is free in distance and depth, and an observer can freely focus the eye on the near or far position to watch.
Drawings
Fig. 1, 2 and 3 are schematic diagrams of the present invention.
Detailed Description
The following describes exemplary embodiments of the present invention with reference to the accompanying drawings.
In fig. 1, 1 is an image source, 2 is a zoom machine, controlled by a micro-space control part, 3 is a lens group, 4, 5 are two micro-viewing angle observation areas 6, 7 are two waveguide sheets, 8, 9 are display output areas of two waveguides, and the two display output areas are used as a large-viewing angle observation part.
The image source 1 generates an imaging surface through the zoom machine 2 and the lens group 3, the micro-space control part generates a scanning imaging surface by controlling the zoom machine 2 to adjust the distance of the imaging surface at a high speed, and the scanning imaging surfaces in a period of time are combined together to form a three-dimensional micro-space. The micro-space generated by the micro-space generating system outputs light rays at the lens group 3, the two micro-visual angle observation areas 4 and 5 observe the model displayed by the micro-space at the position of double visual angles, the light rays are transmitted to the two large-visual angle observation parts through the two waveguides 6 and 7, and the three-dimensional model can be observed by two eyes at the large-visual angle observation parts. Because the large-visual-angle observation part adopts the transparent waveguide sheet, an observer can see the surrounding actual scene and also can see the virtual model.
The image plane generated by the image source in fig. 2 is a two-dimensional XY plane. X is indicated in FIG. 2 and Y is perpendicular to the page. The direction of variation of the imaging plane in the third dimension, the Z direction, is marked in fig. 2. A, B in FIG. 2 is the two-view viewpoint of the micro-viewing section.
The image source is a two-dimensional image source, the two-dimensional resolution is Xn × Yn, in fig. 2, X1, X2, X3, X4, and X5.. Xn are pixel arrangements in the X direction of an imaging plane at a certain time, and pixel arrangements in the Y direction are perpendicular to a paper surface and are not marked in the figure.
When the micro-space control part adjusts the distance of the imaging surface at a high speed, each distance adjusting cycle can be divided into thousands of time intervals, in fig. 2, Z1, Z2, Z3, Z4, and Z5.
In fig. 2, the farther the scan imaging plane is from the micro-viewing angle observation part, the larger the spatial span of each scan imaging plane is set, and the larger the spatial span occupied by the corresponding single pixel is, because the very small spatial pixel cannot be distinguished when the micro-viewing angle observation part observes the model at the farther distance, and setting such a change in spatial scan span can generate a micro-space of a larger range with a relatively small spatial scan time.
When the space model is displayed, as shown in fig. 3, a double-line rectangle in the drawing is a cross section of an object model to be displayed, the cross section is parallel to the gaze direction, pixel space occupied by the double-line rectangle is marked by a shaded part, and the space model can be displayed by controlling an image source to light or close a pixel at a specific position (positioning X, Y coordinate) at a specific moment (positioning Z coordinate). The shaded portion in fig. 3 marks the set of spatial pixels to be lit during scanning. The viewing angles of the micro-viewing part at different positions can see different scenes, and the specific scenes observed at the two viewing angles are marked by dotted lines in fig. 3.
The above-mentioned embodiments are merely illustrative of the preferred embodiments of the present invention, and do not limit the scope of the present invention, and various modifications and improvements made to the technical solution of the present invention by those skilled in the art without departing from the spirit of the present invention should fall within the protection scope defined by the claims of the present invention.
Claims (4)
1. A space display scheme is characterized by comprising a micro-space part, a micro-visual angle observation part and a large-visual angle observation part, wherein the micro-space part comprises a three-dimensional object model with a micro scale, the three-dimensional object model is generated by a micro-space generation system through a method of exchanging time for space, the micro-visual angle observation part receives light of the micro-space part at the micro scale, and the large-visual angle observation part outputs light of a target scale for human eyes to observe.
2. The spatial display scheme according to claim 1, wherein said micro-viewing angle observation part observes the model of the micro-space at a micro-scale viewing angle, the micro-viewing angle observation part transmits light to the large-viewing angle observation part, the light is expanded to form the large-viewing angle observation light, the three-dimensional space picture observed by the large-viewing angle observation part at a certain viewing angle is not generated by rendering calculation but is directly expanded by the micro-viewing angle observation part to see the scene.
3. The spatial display scheme according to claim 1, wherein the micro-space generating system comprises an image source, a lens set, and a micro-space control portion, wherein the image source forms an imaging surface at a certain distance through the lens set, the micro-space control portion adjusts the imaging surface back and forth at a high speed to form a scanning imaging surface, the scanning imaging surfaces at different times are combined together to form a three-dimensional micro-space in a period of time, the micro-space control portion fills pixels of a three-dimensional model to be displayed to corresponding positions of the three-dimensional micro-space at corresponding times to form a spatial pixel, and since the scanning speed is fast, the lighting and closing speeds of the spatial pixel are fast, and the micro-space control portion can display the three-dimensional model in the whole micro-space range.
4. A spatial display scheme according to claim 1 wherein the further the scan image plane is from the micro-perspective viewing portion, the larger the spatial span of the scan image plane is set, and the larger the three-dimensional space occupied by the corresponding individual pixel on the scan image plane, so that a greater range of micro-spaces can be created with less scan time.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811600521.5A CN111435199A (en) | 2018-12-26 | 2018-12-26 | Space display scheme |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811600521.5A CN111435199A (en) | 2018-12-26 | 2018-12-26 | Space display scheme |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111435199A true CN111435199A (en) | 2020-07-21 |
Family
ID=71579736
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811600521.5A Pending CN111435199A (en) | 2018-12-26 | 2018-12-26 | Space display scheme |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111435199A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102361497A (en) * | 2011-11-15 | 2012-02-22 | 南京大学 | Display method and display system for spatial three-dimensional video |
CN103487938A (en) * | 2013-08-28 | 2014-01-01 | 成都理想境界科技有限公司 | Head mounted display |
CN103995356A (en) * | 2014-05-30 | 2014-08-20 | 北京理工大学 | Light field helmet display device increasing real stereoscopic impression |
CN106537220A (en) * | 2014-03-05 | 2017-03-22 | 亚利桑那大学评议会 | Wearable 3D augmented reality display with variable focus and/or object recognition |
CN106537219A (en) * | 2014-05-30 | 2017-03-22 | 奇跃公司 | Methods and system for creating focal planes in virtual and augmented reality |
US20180035103A1 (en) * | 2016-07-26 | 2018-02-01 | Samsung Electronics Co., Ltd. | See-through type display apparatus |
-
2018
- 2018-12-26 CN CN201811600521.5A patent/CN111435199A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102361497A (en) * | 2011-11-15 | 2012-02-22 | 南京大学 | Display method and display system for spatial three-dimensional video |
CN103487938A (en) * | 2013-08-28 | 2014-01-01 | 成都理想境界科技有限公司 | Head mounted display |
CN106537220A (en) * | 2014-03-05 | 2017-03-22 | 亚利桑那大学评议会 | Wearable 3D augmented reality display with variable focus and/or object recognition |
CN103995356A (en) * | 2014-05-30 | 2014-08-20 | 北京理工大学 | Light field helmet display device increasing real stereoscopic impression |
CN106537219A (en) * | 2014-05-30 | 2017-03-22 | 奇跃公司 | Methods and system for creating focal planes in virtual and augmented reality |
US20180035103A1 (en) * | 2016-07-26 | 2018-02-01 | Samsung Electronics Co., Ltd. | See-through type display apparatus |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105866949B (en) | The binocular AR helmets and depth of field adjusting method of the depth of field can be automatically adjusted | |
JP3492251B2 (en) | Image input device and image display device | |
US20050146788A1 (en) | Software out-of-focus 3D method, system, and apparatus | |
TW201643503A (en) | Virtual image generator | |
CN108628087B (en) | Projection device and space imaging method | |
JPH10239634A (en) | Stereoscopic video display device | |
US9905143B1 (en) | Display apparatus and method of displaying using image renderers and optical combiners | |
CN103913845A (en) | Method for displaying three-dimensional integral images using mask and time division multiplexing | |
JP2014219621A (en) | Display device and display control program | |
CN103605211A (en) | Tablet non-auxiliary stereo display device and method | |
CN106291945B (en) | A kind of display panel and display device | |
WO2011086874A1 (en) | Display device and display method | |
CN110035274A (en) | 3 D displaying method based on grating | |
CN109445003A (en) | A kind of color solid More imaging optical device | |
JPH0396913A (en) | Image display device | |
CN107179608A (en) | A kind of three-dimensional augmented reality display device and method | |
KR102016397B1 (en) | Contents generation method of 3D display system using an OLED panel | |
CN111435199A (en) | Space display scheme | |
EP2408217A2 (en) | Method of virtual 3d image presentation and apparatus for virtual 3d image presentation | |
CN206133120U (en) | Display panel and display device | |
Iwamoto et al. | High resolution binocular video see-through display for interactive work support-development of system and evaluation of depth perception and peg-in-hole tasks | |
JP2018074463A (en) | Image generation device, image display system and program | |
CN110794590B (en) | Virtual reality display system and display method thereof | |
Yao et al. | Uniform luminance light field near eye display using pinhole arrays and gradual virtual aperture | |
JP4995672B2 (en) | Stereoscopic image display device and method of creating stereoscopic image display |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20200721 |
|
WD01 | Invention patent application deemed withdrawn after publication |