KR101882977B1 - Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image - Google Patents
Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image Download PDFInfo
- Publication number
- KR101882977B1 KR101882977B1 KR1020160060144A KR20160060144A KR101882977B1 KR 101882977 B1 KR101882977 B1 KR 101882977B1 KR 1020160060144 A KR1020160060144 A KR 1020160060144A KR 20160060144 A KR20160060144 A KR 20160060144A KR 101882977 B1 KR101882977 B1 KR 101882977B1
- Authority
- KR
- South Korea
- Prior art keywords
- image
- smart terminal
- lens
- fisheye
- lens module
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/56—Accessories
- G03B17/565—Optical accessories, e.g. converters for close-up photography, tele-convertors, wide-angle convertors
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B7/00—Mountings, adjusting means, or light-tight connections, for optical elements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B7/00—Mountings, adjusting means, or light-tight connections, for optical elements
- G02B7/003—Alignment of optical elements
- G02B7/005—Motorised alignment
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/56—Accessories
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Optics & Photonics (AREA)
- Studio Devices (AREA)
Abstract
The 360-degree lens module according to the present embodiment includes: a housing having a light outlet; a pair of fisheye lenses facing each other in the opposite directions; a pair of fisheye lenses And a fixing member formed on the housing and fixing the lens module to the smart terminal.
Description
The present invention relates to a lens module and a 360 degree image forming application capable of forming a 360 degree image.
Virtual reality is emerging. Virtual reality refers to the duplication of real or virtual environments. In other words, virtual reality refers to a technique of duplicating the relationship between a user's actual existence and a user or between a user and a specific object in order to form a mutual relationship. Virtual reality provides users with a sense of artificially created visual, tactile, auditory, and olfactory sense.
Among the virtual reality contents, the visual contents based on the reality are formed using the 360 degree camera. The 360 degree camera forms an external image using a plurality of lenses, performs an image processing process to form an image in a direction provided by the user, and provides the image to a user.
Conventional 360 degree cameras have a high price because optical parts, imaging parts and image processing parts are integrally formed. Therefore, there is a high cost that a consumer has to pay to experience a 360 degree image. In addition, according to the conventional 360-degree camera, when a user uploads a 360-degree image to a web or transmits it to another person, the user removes the memory from the camera, reads the memory from a terminal such as a computer, It is difficult to transmit the image immediately after shooting.
The lens module and the application according to the present embodiment are provided to solve the difficulties of the related art and provide a 360 degree lens module and an application which can be mounted on a portable smart terminal such as a mobile phone and can form an image of 360 degrees economically Is one of the main objects of the present embodiment.
It is also a primary object of the present invention to provide a 360 degree lens module and application that can be mounted on a portable smart terminal to instantly transmit an image stored in a portable smart terminal by forming and storing a 360 degree image.
The 360-degree lens module according to the present embodiment includes: a housing having a light outlet; a pair of fisheye lenses facing each other in the opposite directions; a pair of fisheye lenses And a fixing member formed on the housing and fixing the lens module to the smart terminal.
(A) obtaining a pair of circular electronic images acquired by a lens module coupled to the smart terminal and including a pair of fish-eye lenses and provided to the smart terminal camera; and (b) converting electronic images into equirectangular image data, wherein the step of converting to equivalent square data comprises the steps of: (b1) converting a pair of Calculating coordinates of pixels forming circular electronic images; and (b2) writing pixel values of the calculated coordinates into pixel values of pixels forming an equivalent rectangular image.
According to the 360-degree lens module and the 360-degree image forming application according to the present embodiment, a 360-degree image can be formed using the computation capability of a smart terminal by mounting it on a widely deployed smart terminal, An advantage of being able to form a 360 degree image is provided. Also, since the photographed 360-degree image is stored in the smart terminal, it is advantageous that the smart terminal can be uploaded to the web after being photographed or transmitted.
1 is a diagram schematically showing an optical path of a lens module according to the present embodiment.
2 and 3 are views showing one side surface and the other side surface of the lens module according to the present embodiment.
4 is a diagram schematically showing an optical path of a fish-eye lens.
5 is a view showing a state in which the lens module is mounted on a smart terminal.
6 is a view showing that the light provided by the lens module according to the present embodiment is provided to the camera formed on the smart terminal through the light outlet.
7 is a flowchart schematically showing each step of the application process according to the present embodiment.
8 is a diagram showing an electronic image obtained by photographing an optical image provided by a lens module to a camera.
Fig. 9 is a diagram illustrating a state in which the optical axis of the fisheye lens and the optical axis of the sensor plane are aligned and unaligned.
Fig. 10 (a) is a diagram showing an outline of the equivalent rectangular image data, and Fig. 10 (b) is a diagram showing a spherical object plane when the fish-eye lens is at the center.
11 is a diagram showing a state in which the coordinate axes are rotated.
12 (a) is a fisheye lens image in which distortion correction is not performed, and FIG. 12 (b) is a fisheye lens image in which distortion correction is performed.
13 (a) is an equivalent quadrangle image formed by a fisheye lens image whose distortion is not corrected, and 13 (b) is an equivalent quadrangle image formed by a fisheye lens image whose distortion is corrected.
The description of the present invention is merely an example for structural or functional explanation, and the scope of the present invention should not be construed as being limited by the embodiments described in the text. That is, the embodiments are to be construed as being variously embodied and having various forms, so that the scope of the present invention should be understood to include equivalents capable of realizing technical ideas.
Meanwhile, the meaning of the terms described in the present application should be understood as follows.
The terms "first "," second ", and the like are used to distinguish one element from another and should not be limited by these terms. For example, the first component may be referred to as a second component, and similarly, the second component may also be referred to as a first component.
It is to be understood that when an element is referred to as being "on" or "on" another element, it may be directly on top of the other element, but other elements may be present in between. On the other hand, when an element is referred to as being "in contact" with another element, it should be understood that there are no other elements in between. On the other hand, other expressions that describe the relationship between components, such as "intervening" and "intervening", between "between" and "immediately" or "neighboring" Direct neighbors "should be interpreted similarly.
It should be understood that the singular " include "or" have "are to be construed as including a stated feature, number, step, operation, component, It is to be understood that the combination is intended to specify that it is present and not to preclude the presence or addition of one or more other features, numbers, steps, operations, components, parts or combinations thereof.
Each step may take place differently from the stated order unless explicitly stated in a specific order in the context. That is, each step may occur in the same order as described, may be performed substantially concurrently, or may be performed in reverse order.
The drawings referred to for explaining embodiments of the present disclosure are exaggerated in size, height, thickness, and the like intentionally for convenience of explanation and understanding, and are not enlarged or reduced in proportion. In addition, any of the components shown in the drawings may be intentionally reduced, and other components may be intentionally enlarged.
All terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs, unless otherwise defined. Terms such as those defined in commonly used dictionaries should be interpreted to be consistent with the meanings in the context of the relevant art and can not be interpreted as having ideal or overly formal meaning unless explicitly defined in the present application .
Hereinafter, a lens module and an application according to the present embodiment will be described with reference to the accompanying drawings. FIG. 1 is a view schematically showing an optical path of a lens module according to an embodiment of the present invention, and FIGS. 2 and 3 are views showing one side surface and another side surface of a lens module according to the present embodiment. 1 to 3, the lens module according to the present embodiment includes a housing H having a light outlet E and a pair of
The housing H accommodates the
Fig. 4 schematically shows an optical path of a fish-eye lens f. Referring to FIG. 4, the fish-
The distance f from the focal point of the fish-eye lens to the imaging surface is a value peculiar to the lens and fixed. The position R at which the image of the object is formed at the intersection of the sensor image pickup surface and the optical axis can be obtained irrespective of the type of the fisheye lens as long as the angle of incidence of the light entering the fisheye lens from the object P can be grasped . It can be seen that the position R is a function of the angle of incidence φ, and is expressed as R (φ).
1 to 3, the
The first reflecting
In the lens module according to the present embodiment, a plurality of
When the alignment of the
The alignment marks 300 are formed on the same plane as the
A fixing
6 is a view showing that the light provided by the lens module according to the present embodiment is provided to the camera c formed on the smart terminal s through the light outlet E. The light condensed by the
In one embodiment of the lens module, the light exit E may be provided with a
Hereinafter, a driving process of an application for forming a 360-degree image from a photographed image using the lens module according to the present embodiment will be described with reference to FIGS. 7 to 11. FIG. 7 is a flowchart schematically showing each step of an application driving process according to the present embodiment. Referring to FIGS. 7 and 8, a pair of optical images acquired by the lens module and provided to the camera c of the smart terminal s are acquired to acquire electronic images (SlOO). The optical image provided by the fisheye lens to the camera is circular as shown in Fig. 8, and the application controls the smart terminal (s) to convert the two circular optical images provided by the pair of fisheye lenses included in the lens module into electronic images Data.
9A is a diagram illustrating an example in which the optical axis of a fisheye lens and the optical axis of the sensor plane are aligned. In the state where the optical axis and the imaging plane of the sensor are aligned as shown in FIG. 9A, When the
When an image in which a distortion occurs is obtained, a distortion of the image can be corrected by performing a perspective transformation. (X1 ', y1'), (x2 ', y2'), (x3 ', y3'), (x4 ', y4'), and the coordinates of the alignment marks in the image in which distortion occurs are (x1, y1), (x2, y2), (x3, y3), (x4, y4).
The distortion of the image due to misalignment can be corrected by a perspective transform as described above, and the perspective transformation can be performed by mapping the coordinates to a homography matrix expressed by Equation (1) below.
a, b, c, d, e, f, g, h: constant coefficient after the x,
The coordinate values x ', y' of the pixel where the perspective transformation has been performed are as shown in
In Equation (2), the formula for moving the coordinates of a point having coordinates (x, y) to a new point (x ', y') is eight in all, Y1 ', x2', y2 ', x3', y3 ', x4', y4 ', which are coordinates of the point at which the pair of alignment marks are located. Also, coordinate values x1, y1, x2, y2, x3, y3, x4, and y4 where the four pairs of alignment marks are located in the distorted image can be obtained. Thus, the unknowns are a, b, c, d, e, f, g, h, and they can be obtained by calculating the determinant of Equation 3 below.
The image distortion caused by the misalignment can be corrected by performing the perspective transformation process on the pixels included in the distorted image by obtaining eight unknown values in Equation (3) and calculating Equation (2) from the obtained result.
The distortion of the image is not limited to the case where the optical axis of the fisheye lens and the sensor plane are misaligned as well as the misalignment between the fisheye lens and the reflective member, the camera of the lens module and the smart terminal, And the fisheye lens of the
The application processes the two circular image data to form an equirectangular image. FIG. 10 (a) is a diagram showing an outline of the equivalent rectangular image data, and FIG. 10 (b) is a diagram showing a spherical object plane when the fish-eye lens is at the center. In Fig. 10 (b), the portion indicated by the light green color shows the object plane taken by one fish-eye lens, which corresponds to the pale green portion in the equivalent rectangular image in Fig. 10 (a). Similarly, in Fig. 10 (b), the portion indicated by yellow indicates the object plane taken by another fisheye lens, and corresponds to the yellow portion in the equivalent rectangular image in Fig. 10 (a).
In an equivalent rectangular image shown in FIG. 10 (a), a pixel P can be expressed by coordinates of? And? Indicating longitude and latitude, respectively. According to the general display method, the latitude ranges from + 90 ° to -90 °, but for convenience of explanation, the latitude of the zenith is set to 0 and the latitude of the lowest point (nadir) is set to 180 ° for convenience.
10 (b), the radius of the sphere is assumed to be 1, and the coordinates of the point P are expressed by an orthogonal coordinate system.
Axis to rotate the coordinate system along the y-axis as shown in Fig. 11 to facilitate calculation of the coordinates of the point corresponding to the point P in the circular image stored in the smart terminal. Let X ', Y', and Z 'be the new coordinate axes formed by the rotation, and obtain the coordinates of the P point in X', Y ', and Z'.
The spherical coordinate is calculated from the new coordinate pair obtained in Equation (5), as shown in Equation (6) below.
(acos: inverse function of cos, atan2 (x, y): function to find the angle between X axis and coordinate (x, y) in XY plane)
? 'Obtained from Equation (6) is an incident angle when light is incident on the fisheye lens from the point P of the object plane,?' Is the rotation angle from the X 'axis, and?' Does not change even when passing through the fisheye lens. Therefore, the coordinates of the point P in the circular electronic images stored in the smart terminal s are obtained as shown in Equation (7) below.
(R (?): A position function of an image of an object formed at an intersection of an image pickup surface and an optical axis in a fish-eye lens, see Table 1)
The coordinates of the X ', Y', and Z 'axes are converted into the coordinates of the X, Y, and Z axes again, and the coordinates of the X', Y ' And? 'Are replaced with the results of Equation (6), Equations (8) and (9) can be obtained. From the results of Equations (8) and (9), the coordinates of the image point provided by the fisheye lens to which the point having the coordinates? And? Of the equivalent rectangular image are mapped can be obtained (S200).
In Equation (5), if the z 'coordinate value is negative, it corresponds to an image obtained by another fisheye lens, and the x and y coordinates are calculated as shown in the following Equation 9 (S200).
Equations (8) and (9) are calculated with respect to the coordinates of the equivalent quadrangle to obtain the corresponding coordinates of the image obtained from the fish-eye lens, and then the pixel values of the obtained coordinates are read and used as pixel values of the corresponding coordinates of the equivalent quadrature image, Thereby forming an image (S300).
The application forms a virtual spherical image from the formed equivalent square image and performs a gnomonic projection on the spherical image to display the image on the smart terminal.
Experimental Example
Hereinafter, an experiment will be described with reference to Figs. 12 to 13, in which distortion is corrected using an application according to the present embodiment with respect to an image obtained in a state where the optical axis of the image sensor and the optical axis of the fisheye lens are misaligned.
12 is an image taken using two fisheye lenses in a misaligned state. The part shown in the circle is an alignment mark. Fig. 12 (a) shows the fisheye lens image without distortion correction, and Fig. 12 (b) shows the yellow solid line before distortion on the image. It can be seen from FIG. 12 (b) that the far-field transformation is performed by calculating the homography matrix accurately.
13 (a) is an equivalent quadrangle image formed by a fisheye lens image with no distortion corrected, and 13 (b) is an equivalent quadrangle image formed by a fisheye lens image whose distortion is corrected. It can be seen that in the equivalent quadrangle image formed by the fisheye lens image without distortion correction, a step is generated at the portion where the two images are combined and deviated, but it is confirmed that no step is generated in the equivalent square image formed by the distortion-corrected image .
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is clearly understood that the same is by way of illustration and example only and is not to be taken by way of illustration, It will be appreciated that other embodiments are possible. Accordingly, the true scope of the present invention should be determined by the appended claims.
100a, 100b: Fish eye lens 200: Reflecting member
200a, 200b: first reflecting member, second reflecting member 300: alignment mark
400: Fixing member 500: Correction lens
Claims (15)
A housing having a light outlet formed therein;
A pair of fisheye lenses facing away from each other and located on different planes;
A reflecting member for changing a path of light provided by the pair of fisheye lenses in the same direction and providing the light to the outside of the housing through the light outlet;
A fixing member formed on the housing and fixing the lens module to the smart terminal; And
The fisheye lens and the smart terminal camera are arranged on the surfaces on which the fisheye lenses are disposed to perform a perspective transformation for correcting distortion caused by misalignment of at least one of the fisheye lens and the smart terminal camera, Each of which is arranged at a vertex of a quadrangle and includes an alignment mark displayed in a different form according to a plane on which the respective fisheye lenses are arranged,
Wherein the fixing member is capable of directly fixing the lens module to the smart terminal.
Wherein the reflective member comprises:
A first reflection member for reflecting the light condensed by the fish-eye lenses in a direction perpendicular to the optical axis of the fish-eye lenses,
And a second reflection member that reflects the light reflected by the first reflection member to the outside of the housing through the light exit.
The lens module includes:
Disposed at the light exit
And an adjustment lens for adjusting the light reflected by the second reflection member.
Wherein:
A clip for picking up and fixing the smart terminal, and a slot in which the smart terminal is inserted and fixed.
Wherein the light outlet is formed at a position corresponding to a camera lens formed on the smart terminal.
The smart terminal includes:
Wherein the camera is equipped with a laptop computer, a mobile phone, and a portable tablet.
(a) obtaining a pair of circular electronic images acquired by a lens module coupled to a smart terminal and provided to the smart terminal camera, including a pair of fisheye lenses facing away from each other and facing each other, ,
(b) converting the electronic images into equirectangular image data, the step of converting to equivalent square data comprises:
(b1) calculating coordinates of a pixel forming the pair of circular electronic images corresponding to image coordinates of a pixel forming the equivalent rectangular image;
(b2) writing the computed pixel value of the coordinate as the pixel value of the pixel forming the equivalent square image,
The lens module includes alignment marks which are respectively disposed on the faces on which the fisheye lenses are disposed and are arranged at vertexes of a quadrangle centered on the fisheye lens and are displayed in different shapes according to the plane on which the fisheye lenses are arranged ,
The 360 degree image forming method
(c) performing a perspective transform to correct for distortion caused by misalignment of one or more of the fisheye lens and the smart terminal camera using the alignment indications.
The image forming method includes:
And performing a gnomonic projection on the equivalent rectangular image data to display a portion of the 360 degree image.
The step (b1)
Computing the angular distance from any one of a zenith and a node of a pixel forming the circular electronic images,
And calculating coordinates of a pixel forming the circular electronic images using the obtained angular distance and the fisheye lens characteristic.
Wherein the step of correcting the perspective transition comprises:
Calculating a coefficient of the homography matrix using the coordinates of the plurality of alignment marks in the absence of misalignment and the coordinates of the plurality of alignment marks obtained in each photographed circular electronic image,
And performing perspective correction using the homography matrix in which the coefficients are obtained.
Wherein the step of correcting the perspective transition comprises:
Equation
(X1, y1, x2, y2, x3, y3, x4) are performed using a homography matrix using the coefficients a, b, c, d, e, f, x4 ', y4', x4 ', y4' of the plurality of alignment marks in the absence of misalignment, x4 ', y2' Coordinates of multiple alignment marks).
Wherein the 360 degree image forming method is performed by the smart terminal.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160060144A KR101882977B1 (en) | 2016-05-17 | 2016-05-17 | Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image |
PCT/KR2017/005106 WO2017200289A1 (en) | 2016-05-17 | 2017-05-17 | Lens module capable of constructing 360-degree image, method for constructing 360-degree image, computer program for implementing method, and storage medium for storing computer program |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160060144A KR101882977B1 (en) | 2016-05-17 | 2016-05-17 | Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image |
Publications (2)
Publication Number | Publication Date |
---|---|
KR20170129448A KR20170129448A (en) | 2017-11-27 |
KR101882977B1 true KR101882977B1 (en) | 2018-07-27 |
Family
ID=60326259
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020160060144A KR101882977B1 (en) | 2016-05-17 | 2016-05-17 | Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image |
Country Status (2)
Country | Link |
---|---|
KR (1) | KR101882977B1 (en) |
WO (1) | WO2017200289A1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101839116B1 (en) * | 2017-11-24 | 2018-03-15 | 주식회사 이노벡스 | System and Method for providing circle image processing |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005244861A (en) * | 2004-02-27 | 2005-09-08 | Japan Science & Technology Agency | Imaging apparatus and imaging system parameter correction method |
JP2006011103A (en) * | 2004-06-28 | 2006-01-12 | Yokogawa Electric Corp | Spherical mirror imaging apparatus |
JP2013157810A (en) * | 2012-01-30 | 2013-08-15 | Toshiba Corp | Image sensor system, information processor, information processing method and program |
KR101431378B1 (en) * | 2013-04-08 | 2014-08-19 | 포항공과대학교 산학협력단 | Method for generating omni-directional image and apparatus therefor |
JP2015031901A (en) * | 2013-08-06 | 2015-02-16 | 日立マクセル株式会社 | Auxiliary lens unit |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100596976B1 (en) * | 2004-11-08 | 2006-07-05 | (주)아이디스 | apparatus and method for correcting distorted image and image display system using it |
KR20150124557A (en) * | 2014-04-28 | 2015-11-06 | 주식회사 테크웍스플러스 | Camera accessory for portable electronic device |
KR101583663B1 (en) * | 2014-04-30 | 2016-01-12 | 아진산업(주) | Method for generating calibration indicator of camera for vehicle |
-
2016
- 2016-05-17 KR KR1020160060144A patent/KR101882977B1/en active IP Right Grant
-
2017
- 2017-05-17 WO PCT/KR2017/005106 patent/WO2017200289A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005244861A (en) * | 2004-02-27 | 2005-09-08 | Japan Science & Technology Agency | Imaging apparatus and imaging system parameter correction method |
JP2006011103A (en) * | 2004-06-28 | 2006-01-12 | Yokogawa Electric Corp | Spherical mirror imaging apparatus |
JP2013157810A (en) * | 2012-01-30 | 2013-08-15 | Toshiba Corp | Image sensor system, information processor, information processing method and program |
KR101431378B1 (en) * | 2013-04-08 | 2014-08-19 | 포항공과대학교 산학협력단 | Method for generating omni-directional image and apparatus therefor |
JP2015031901A (en) * | 2013-08-06 | 2015-02-16 | 日立マクセル株式会社 | Auxiliary lens unit |
Also Published As
Publication number | Publication date |
---|---|
KR20170129448A (en) | 2017-11-27 |
WO2017200289A1 (en) | 2017-11-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6859442B2 (en) | Calibration equipment, calibration system, and calibration method | |
US10645284B2 (en) | Image processing device, image processing method, and recording medium storing program | |
CN106600546B (en) | Distortion correction method and system for ultra-wide-angle camera | |
Micusik et al. | Autocalibration & 3D reconstruction with non-central catadioptric cameras | |
US8619248B2 (en) | System and method for calibrating ultra wide-angle lenses | |
US10210622B2 (en) | Image processing device, image processing method, and recording medium storing program | |
US9436973B2 (en) | Coordinate computation device and method, and an image processing device and method | |
US20090041379A1 (en) | Method for providing output image in either cylindrical mode or perspective mode | |
KR20170020777A (en) | Wide field of view array camera for hemispheric and spherical imaging | |
CN103685917A (en) | Image processor, image processing method and program, and imaging system | |
Li | Full-view spherical image camera | |
US11456326B2 (en) | Plenoptic camera for mobile devices | |
US11281087B2 (en) | Imaging device, image processing apparatus, and image processing method | |
US20230152672A1 (en) | Mounting systems for multi-camera imagers | |
US11380016B2 (en) | Fisheye camera calibration system, method and electronic device | |
WO2019229887A1 (en) | Camera apparatus | |
JP2010276433A (en) | Imaging device, image processor, and distance measuring device | |
CN113259642B (en) | Film visual angle adjusting method and system | |
KR101882977B1 (en) | Lens Module for Forming 360 Degree Image and Application for Forming 360 Degree Image | |
TWI705292B (en) | Method of determining assembly quality of camera module | |
Ueno et al. | Compound-Eye Camera Module as Small as 8.5$\times $8.5$\times $6.0 mm for 26 k-Resolution Depth Map and 2-Mpix 2D Imaging | |
Jo et al. | Design of omnidirectional camera lens system with catadioptic system | |
KR101902999B1 (en) | Camera for Forming 360 Degree Image | |
Kuthirummal et al. | Flexible mirror imaging | |
JP7328382B2 (en) | wearable device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A201 | Request for examination | ||
E902 | Notification of reason for refusal | ||
E902 | Notification of reason for refusal | ||
E701 | Decision to grant or registration of patent right | ||
GRNT | Written decision to grant |