WO2013081287A1 - 3d 차량 주변 영상 생성 방법 및 장치 - Google Patents
3d 차량 주변 영상 생성 방법 및 장치 Download PDFInfo
- Publication number
- WO2013081287A1 WO2013081287A1 PCT/KR2012/007781 KR2012007781W WO2013081287A1 WO 2013081287 A1 WO2013081287 A1 WO 2013081287A1 KR 2012007781 W KR2012007781 W KR 2012007781W WO 2013081287 A1 WO2013081287 A1 WO 2013081287A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- vehicle
- camera
- virtual
- virtual camera
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000013507 mapping Methods 0.000 claims abstract description 11
- 230000002093 peripheral effect Effects 0.000 claims description 17
- 238000004891 communication Methods 0.000 claims description 7
- 230000008859 change Effects 0.000 claims description 5
- 239000002131 composite material Substances 0.000 claims description 3
- 230000002194 synthesizing effect Effects 0.000 claims description 3
- 230000008901 benefit Effects 0.000 abstract description 5
- 238000010586 diagram Methods 0.000 description 13
- 238000009434 installation Methods 0.000 description 9
- 230000036544 posture Effects 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 238000012937 correction Methods 0.000 description 5
- 239000000284 extract Substances 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000005452 bending Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 210000000988 bone and bone Anatomy 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000012827 research and development Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/04—Texture mapping
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/27—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view providing all-round vision, e.g. using omnidirectional cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/02—Rear-view mirror arrangements
- B60R1/08—Rear-view mirror arrangements involving special optical features, e.g. avoiding blind spots, e.g. convex mirrors; Side-by-side associations of rear-view and other mirrors
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/28—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with an adjustable field of view
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
- G06T15/205—Image-based rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/06—Topological mapping of higher dimensional structures onto lower dimensional surfaces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/282—Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/60—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective
- B60R2300/602—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective with an adjustable viewpoint
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/60—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective
- B60R2300/602—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective with an adjustable viewpoint
- B60R2300/605—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective with an adjustable viewpoint the adjustment being automatic
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/60—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective
- B60R2300/607—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective from a bird's eye viewpoint
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2215/00—Indexing scheme for image rendering
- G06T2215/16—Using real world measurements to influence rendering
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
Definitions
- the present invention relates to a method and apparatus for generating a surrounding image of a vehicle, and more particularly, to a method and apparatus for generating a surrounding image of a vehicle, which can provide a 3D effect in 3D.
- AVM Around View Monitoring
- the surrounding image display system captures the surrounding environment through cameras installed on the front, rear, left and right sides of the car, and displays the surrounding environment of the car on the screen by correcting the overlapped area to look natural based on the captured image. do. Accordingly, the driver can accurately recognize the surroundings of the vehicle through the displayed surrounding environment and can conveniently park without looking at the side mirror or the rear mirror.
- an image captured by a camera is basically mapped to a 3D space virtual plane using a 3D space model.
- the composite image may be generated from an input image mapped to the 3D space virtual surface according to a predetermined correspondence according to the position of the camera virtual viewpoint determined by the driving state of the vehicle or the user's selection, the direction of the gaze, and the focal length.
- the synthesized image does not look natural.
- a hemisphere model is used as a 3D spatial model
- objects located far from the vehicle are relatively natural, but the ground near the vehicle tends to be unnatural.
- the mapping surface non-ground parts such as people standing around the vehicle or obstacles are stretched out to obtain a proper image because the image is long and the actual camera image of each view does not match well during the matching process. There were these difficult problems.
- an object of the present invention is to provide a 3D vehicle surrounding image generation method and apparatus capable of more naturally and three-dimensionally expressing a vehicle surrounding image including surrounding obstacles.
- a 3D space model in the form of a container in which a bottom surface of the image photographed from a plurality of cameras installed in a vehicle is flat and the radius thereof is widened upward. Mapping to the virtual plane defined by the step, and generating a view image of the viewpoint of the virtual camera using the image mapped to the virtual plane.
- the viewpoint of the virtual camera may be determined by at least one of a driving state of the vehicle and a user selection.
- the size of the bottom surface of the three-dimensional space model may be inversely proportional to the traveling speed of the vehicle.
- the synthesis image generation may be performed by referring to a lookup table in which a corresponding relationship between the image mapped to the virtual plane and the view image of the virtual camera view point is defined in advance.
- the center of the region where the images captured by the plurality of cameras overlap may be changed according to the tilt angle of the virtual camera.
- the plurality of cameras includes a front camera, a right camera, a rear camera, and a left camera respectively installed at the front, right, rear, and left sides of the vehicle, and the position of the virtual camera is left or right with respect to the center of the vehicle.
- the center of the overlapping region may be changed to further include an image photographed by a camera installed at the side where the virtual camera is located according to the degree located at.
- the virtual camera viewpoint may be linked to a steering angle direction or a gear position of the vehicle.
- the method may further include displaying an infant (UI) for receiving a user selection with respect to the viewpoint of the virtual camera, and displaying a change of the viewpoint of the virtual camera according to a user selection input through the infant. can do.
- UI infant
- the method may further include synthesizing building information and road information around a driving position of the vehicle to the view image according to the virtual camera viewpoint.
- a computer readable medium records a program for causing a computer to execute any one of the above methods.
- a vehicle surrounding image display system includes an image input unit configured to receive images captured from a plurality of cameras installed in a vehicle, and a radius of the image inputted to the image input unit is flat and the radius thereof is widened upward. And a peripheral image generating apparatus that maps to a virtual plane defined by a three-dimensional space model in the form of a container, and generates a view image viewed from the viewpoint of the virtual camera using the image mapped to the virtual plane.
- the surrounding image raw growth value may be synthesized with the view image according to the virtual camera viewpoint by building information and road information around the driving position of the vehicle.
- the plurality of cameras may transmit the captured image to the surrounding image generating apparatus through short range wireless communication.
- the surrounding image of the vehicle can be expressed more naturally and three-dimensionally, including surrounding obstacles.
- the user can conveniently adjust the viewpoint of the virtual camera.
- FIG. 1 is a block diagram illustrating a vehicle surrounding image display system including an external parameter estimating apparatus according to an exemplary embodiment of the present invention.
- FIG. 2 is a schematic view for explaining an arrangement of a calibration plate on which an installation position and a triangular pattern of a camera according to an exemplary embodiment of the present invention are displayed.
- FIG. 3 is a schematic diagram provided to explain external parameters of a vehicular installation camera according to an embodiment of the present invention.
- FIG. 4 is a diagram illustrating an example of a front camera input image including a correction plate displaying a triangular pattern according to an exemplary embodiment of the present invention.
- FIG. 5 is a flowchart provided to explain a method of estimating a camera external parameter according to an embodiment of the present invention.
- FIG. 6 is a diagram illustrating vertices extracted from images captured by front, rear, left, and right cameras according to an exemplary embodiment of the present invention.
- FIG. 7 is a flowchart provided to explain a method of estimating a pan angle, an X coordinate, and a Y coordinate among camera external parameters according to an embodiment of the present invention.
- FIG. 8 is a flowchart provided to explain a 3D vehicle surrounding image generating method according to an exemplary embodiment.
- FIG. 9 is a view provided to explain a three-dimensional space model according to an embodiment of the present invention.
- FIG. 10 is a diagram provided to explain an example of mapping an image photographed by an actual camera to a virtual surface of a 3D spatial model according to an exemplary embodiment.
- FIG. 11 is a diagram provided to explain a method of determining a virtual camera viewpoint by user selection according to an embodiment of the present invention.
- FIG. 12 is a diagram provided to explain an area where images captured by a plurality of cameras overlap when generating a view image of a virtual camera.
- FIG. 13 illustrates an example of displaying building and road information around a vehicle in a view image of a virtual camera.
- FIG. 1 is a block diagram illustrating a vehicle surrounding image display system including an external parameter estimating apparatus according to an exemplary embodiment of the present invention.
- a vehicle surrounding image display system includes four cameras 210, 220, 230, and 240, a surrounding image generating apparatus 300, and a display apparatus 400, and includes an external parameter estimating apparatus 100. It may further include.
- the peripheral image generating apparatus 300 and the external parameter estimating apparatus 100 may be implemented in one form.
- the vehicle surrounding image display system displays the surrounding image generated by processing images captured by four cameras 210, 220, 230, and 240 installed on the vehicle, so that the driver can check the surrounding situation of the vehicle. It is a system.
- the vehicle surrounding image display system may provide a vehicle surrounding image as a 3D image viewed from a virtual viewpoint. To this end, it is necessary to know the external parameters for the posture and the position of the four cameras 210, 220, 230, 240 installed in the vehicle.
- the four cameras 210, 220, 230, and 240 may be installed at the front, rear, left, and right sides of the vehicle, respectively, and may include a lens having a large angle of view, such as a wide angle lens or a fisheye lens.
- the cameras 210, 220, 230, and 240 may photograph a 3D object as a 2D image through a lens, and provide the captured image to an external parameter estimating apparatus 100, a peripheral image generating apparatus 300, or the like. have.
- the cameras 210, 220, 230, and 240 are equipped with short-range wireless communication modules such as Wi-Fi, Bluetooth, Zigbee, and UWB, and include a peripheral image generating apparatus 300 and an external parameter estimating apparatus 100. It may be implemented to wirelessly transmit the image.
- the external parameter estimating apparatus 100 extracts vertices having a triangular pattern from images provided from four cameras 210, 220, 230, and 240, and uses the extracted vertices for cameras 210, 220, 230. , An external parameter of 240 may be estimated.
- the external parameter estimation method is described in detail below.
- the peripheral image generating apparatus 300 uses the images provided from the four cameras 210, 220, 230, and 240, and the virtual viewpoint for the virtual camera having a predetermined virtual viewpoint determined according to a vehicle driving situation or a user selection. An image may be generated and output to the display device 400. To this end, the peripheral image generating apparatus 300 may obtain a texture map by mapping the input image provided from the cameras 210, 220, 230, and 240 to the 3D model plane, which is an external parameter estimating apparatus 100. ), A camera-specific projection model obtained based on the camera external parameter obtained in FIG. In the process of mapping the input image to the 3D model virtual plane, a weighted blending technique or the like may be used to make the input image photographed by different cameras look natural. The peripheral image generating apparatus 300 may generate and output a virtual viewpoint image from a texture map by using a projection model of a virtual camera having a corresponding virtual viewpoint when a predetermined virtual viewpoint is determined.
- the display device 400 displays the generated peripheral image on a display module such as a liquid crystal display (LCD) and an organic light emitting diode (OLED).
- a display module such as a liquid crystal display (LCD) and an organic light emitting diode (OLED).
- LCD liquid crystal display
- OLED organic light emitting diode
- an automatic navigation device installed in the vehicle may receive the surrounding image and display it on the screen.
- the peripheral image generating apparatus 300, the external parameter estimating apparatus 100, and the display apparatus 400 may be implemented as a portable communication terminal such as a smartphone or a tablet PC.
- operations such as correcting lens distortion of four cameras 210, 220, 230, and 240 and obtaining a projection model for each camera may be required. This may be performed using a known technique that is generally known. Omit.
- FIG. 2 is a schematic view for explaining an arrangement of a calibration plate on which an installation position and a triangular pattern of a camera according to an exemplary embodiment of the present invention are displayed.
- cameras 210, 220, 230, and 240 may be installed at the front, rear, left, and right sides of the vehicle V, respectively.
- the front camera 210 is installed at the center of the bone net of the vehicle V
- the cameras 230, The 240 may be installed to be positioned at edges or bottoms of both side mirrors of the vehicle V, respectively.
- the camera 220 installed at the rear may be installed at the center of the rear bumper. Since the scale, image quality, etc.
- the heights of the cameras 210 and 220 installed in the front and the rear are the same, and the cameras 230, It is preferable to make the height of 240 same with each other.
- the cameras 230 and 240 installed on the left and right sides are installed so that 170 ° or more can be photographed based on the vertical line in the ground direction.
- the installation positions of the cameras 210, 220, 230, and 240 may be different according to the type of the vehicle, and installation restrictions may occur due to the design of the vehicle.
- a wide-angle camera may be deteriorated in image quality due to a lack of light amount around the lens, and may cause more distortion in the peripheral portion than the central portion of the lens. Also, the image quality of the peripheral part is severely degraded when the viewpoint image is converted by the camera. Therefore, the cameras 210 and 220 installed in the front and rear of the camera lens 210 and 220 installed in the front and rear of the camera lens so that the optical axis is parallel to the horizon, and the cameras 230 and 240 installed in the left and right are perpendicular to the ground. Can be installed as possible.
- the installation height of the cameras 210, 220, 230, and 240 may be determined to be photographed up to about 1.5 m away from the front, rear, left, and right sides of the vehicle V, and the cameras 210, 220, 230, and 240 may be determined. Can be installed to be photographed from about 30 ° to 60 ° from the vertical axis with respect to the ground.
- the installation position of the camera described above has been described with reference to a preferred example, and the camera external parameter estimation apparatus 100 according to the present invention does not necessarily have to correctly install the cameras 210, 220, 230, and 240 at the corresponding position.
- the calibration plates PL1, PL2, PL3, and PL4 marked with a triangular pattern have a certain distance from each corner of the vehicle V so that two calibration plates are included in each image photographed by the cameras 210, 220, 230, and 240.
- the images captured by the installed camera 230 include calibration plates PL1 and PL4, and the images captured by the camera 240 installed on the right include calibration plates PL2 and PL3. At a certain distance from the calibration plates (PL1, PL2, PL3, PL4) can be installed.
- the calibration plates PL1, PL2, PL3, and PL4 need only be installed on the front left, front right, rear right and rear left sides of the vehicle V, respectively, but are not necessarily installed at the correct predetermined position. However, the calibration plates PL1, PL2, PL3 and PL4 should be placed parallel to the ground on which the vehicle V is located.
- the calibration plates (PL1, PL2, PL3, PL4) can be displayed as an equilateral triangle pattern having a constant thickness, as shown in Figure 2, to implement so that the size of the triangle inside the border is 0.4 ⁇ 0.8 of the size of the outer triangle Can be.
- This is for automatically and accurately extracting the triangular pattern displayed on the calibration plates PL1, PL2, PL3, and PL4 from the surrounding similar triangular patterns, but is not necessarily limited thereto.
- the calibration plates PL1, PL2, PL3, and PL4 may be implemented in various forms as long as they can extract vertices corresponding to vertices of an equilateral triangle.
- FIG. 3 is a schematic diagram provided to explain external parameters of a vehicular installation camera according to an embodiment of the present invention.
- the external parameters of the cameras 210, 220, 230, and 240 installed in the vehicle V may include three-dimensional space coordinates (x, y, z) and the cameras 210, 220, 230, and 240. Pan, tilt, roll angles ( ⁇ , ⁇ , ⁇ ).
- the coordinate z may correspond to the height from the ground (G) where the vehicle V is located, and as shown in FIG. 210, the heights of the left camera 230 and the right camera 240 may be z F , z L , and z B , respectively.
- the coordinate (x) and the coordinate (y) may correspond to the position on the virtual plane parallel to the ground (G) where the vehicle (V) is located, as shown in Figure 3 (b) of the vehicle (V)
- the center O may be used as a coordinate reference.
- the pan angle ⁇ may be defined as an angle at which the head directions of the cameras 210, 220, 230, and 240 form the traveling direction of the vehicle V, and the camera 210 as illustrated in FIG. 3 (b).
- the fan angles 220, 230, and 240 may have values of ⁇ F , ⁇ B , ⁇ L , and ⁇ R , respectively.
- the tilt angle ⁇ may be defined as an angle formed with the ground G. As illustrated in FIG. 3A, the tilt angles of the front camera 210 and the rear camera 220 are ⁇ F and ⁇ B , respectively. It can have a value of.
- the roll angle ⁇ may be defined as the rotation angle of the cameras 210, 220, 230, and 240 with respect to the camera head direction axis, and the roll angle of the left camera 230 as illustrated in FIG. 3 (a). May have a value of ⁇ L.
- FIG. 4 is a diagram illustrating an example of a front camera input image including a correction plate displaying a triangular pattern according to an embodiment of the present invention
- FIG. 5 illustrates a method of estimating camera external parameters according to an embodiment of the present invention. Is a flowchart provided.
- images captured and input by the front camera 210 may include correction plates PL1 and PL2 displaying two triangle patterns.
- the triangular pattern displayed on the correction plates PL1 and PL2 is actually an equilateral triangle whose side length is A, the input image is determined by the lens distortion, tilt angle ( ⁇ ), roll angle ( ⁇ ) and height (z) of the front camera 210.
- the sides of the triangular pattern in (a 1 , a 2 , a 3 , a 4 , a 5 , a 6 ) are different.
- the lens distortion in the input image by the front camera 210 is corrected by a known method, and converted into an image of a virtual camera having a virtual viewpoint in a direction (top view) looking down at the ground G from above the vehicle V.
- the triangle patterns PL1 and PL2 have the same length, the lengths a 1 , a 2 , a 3 , a 4 , a 5 , and a 6 are equal to each other.
- the tilt angle ⁇ , roll angle ⁇ , and height z of the front camera 210 may be estimated.
- FIG. 5 is a flowchart provided to explain a method of estimating tilt angle, roll angle, and height among camera external parameters according to an embodiment of the present invention.
- the external parameter estimating apparatus 100 corrects lens distortion with respect to an input image by the front camera 210 (S510). Thereafter, six vertices (P1, P2, P3, P4, and P5), three in each of the triangular patterns displayed on the front left calibration plate PL1 and front right calibration plate PL2 included in the image photographed by the front camera. , P6) is extracted (S520).
- the external parameter estimation apparatus 100 extracts six vertices P1, P2, P3, P4, and P5 extracted while changing the tilt angle ⁇ , roll angle ⁇ , and height z of the front camera 210. , P6) is converted into world coordinates (S530).
- the world coordinate may be an arbitrary reference point or a point O on the ground G where the center of the vehicle V is located as the coordinate reference point.
- the external parameter estimating apparatus 100 uses the front left calibration plate PL1 and the front right calibration plate using the world coordinates of the six vertices P1, P2, P3, P4, P5, and P6 obtained in step S530.
- the length (a 1 , a 2 , a 3 , a 4 , a 5 , a 6 ) of the sides of the triangular pattern indicated by PL2) is obtained (S540).
- step S540 to minimize the magnitude of the difference between the actual length of the side of the triangle pattern (A) and the length of the side of the triangle pattern (a 1 , a 2 , a 3 , a 4 , a 5 , a 6 ) obtained in step S540.
- the tilt angle ⁇ , roll angle ⁇ and height z may be estimated as the tilt angle ⁇ F , roll angle ⁇ F and height z F of the front camera 210 (S550). .
- step S550 the tilt angle ⁇ F , the roll angle ⁇ F , and the height z F that minimize the value, f ( ⁇ , ⁇ , z) obtained by the following equation 1 may be obtained.
- the tilt angle ⁇ , roll angle ⁇ , and height z of the remaining rear camera 220, the right camera 230, and the left camera 240 may be estimated.
- FIG. 6 is a diagram illustrating vertices extracted from images captured by front, rear, left, and right cameras according to an exemplary embodiment of the present invention.
- the vertex P1 F , P2 F , P3 F Are three vertices extracted from an image taken by the front camera 210 with respect to the front left calibration plate PL1, and the vertices P4 F , P5 F , P6 F ) Are three vertices extracted from an image captured by the front camera 210 with respect to the front right side calibration plate PL2.
- vertex (P10 L , P11 L , P12 L ) Are three vertices extracted from the image photographed by the left camera 230 with respect to the rear left calibration plate PL4, and vertices P1.
- L , P2 L , P3 L Are three vertices extracted from an image captured by the left camera 240 with respect to the front left calibration plate PL1.
- FIG. 7 is a flowchart provided to explain a method of estimating a pan angle, an X coordinate, and a Y coordinate among camera external parameters according to an embodiment of the present invention.
- world coordinates may be obtained for vertices extracted after lens distortion correction from images captured by the cameras 210, 220, 230, and 240 (S710).
- the tilt angle ⁇ , roll angle ⁇ and height z of the cameras 210, 220, 230, and 240 use values estimated by the method described above, and the position coordinates x, y) and the fan angle ⁇ are set to a predetermined initial value.
- the pan angle ⁇ of the front camera 210 is set to 0 °
- the left camera 240 is set to 90 °
- the rear camera 220 is set to 180 °
- the right camera 230 is set to 270 °.
- the position coordinates (x, y) may be set to initial values in consideration of the case where the cameras 210, 220, 230, and 240 are correctly installed using the vehicle center O as the coordinate reference point.
- vertices P1 F , P2 F , and P3 F obtained in step S710 and vertices P1 L , P2 L , and P3 L
- the world coordinates of the corresponding points of) will coincide with each other so that there will be no position error.
- an error may occur in a camera mounting process or a driving process, and thus a position error may occur between corresponding points.
- the position error may be defined as the distance between the corresponding points based on the world coordinates.
- the external parameter estimating apparatus 100 obtains a position error between corresponding points while changing the position coordinates (x, y) and the pan angle ( ⁇ ) of the cameras 210, 220, 230, and 240 (S720).
- the position coordinates (x, y) and the pan angle ( ⁇ ) of the cameras 210, 220, 230, and 240 may be estimated as values that minimize the position error between the corresponding points (S730).
- the position coordinates (x, y) and the pan angle ( ⁇ ) are minimized to minimize y R ).
- Di is a distance between corresponding points among vertices extracted from images taken by different cameras.
- D1 means the distance between vertex P1 F and vertex P1 L.
- the relative external parameters of each of the cameras 210, 220, 230, and 240 can be estimated, and knowing the absolute position and the posture of one of the cameras 210, 220, 230, and 240, I can get a posture.
- the absolute positions and postures of the remaining cameras may be naturally obtained.
- the absolute position and posture of the camera mounted on the vehicle may be obtained in various ways.
- FIG. 8 is a flowchart provided to explain a 3D vehicle surrounding image generating method according to an exemplary embodiment.
- the peripheral image generating apparatus 300 may map an image photographed from the cameras 210, 220, 230, and 240 to a virtual plane defined by a 3D space model (S810).
- a model M having a container-shaped virtual surface in which the bottom surface A is flat and the radius of the upper portion A becomes flat can be used as illustrated in FIG. 9.
- the three-dimensional space model (M) has a top surface (A) having a long radius (b 1 ) and a short radius (a 1 ) and an upper surface opened at a height (c) with a radius widening toward the top.
- (A ') may be defined as having a long radius (b 1 + b 2 ) and a short radius (a 1 + a 2 ).
- the bottom and top surfaces may be implemented in a circular shape.
- ⁇ is a parameter related to the size of the bottom surface. If ⁇ is '0', there is no bottom surface, and if ⁇ is '1', the bottom surface has a long radius b 1 and a short radius a 1 .
- the size of the base can be adjusted by adjusting the ⁇ value, and the value can be changed according to the vehicle driving condition, driving speed, etc. as set by the system designer or described below.
- FIG. 10 is a diagram provided to explain an example of mapping an image photographed by an actual camera to a virtual surface of a 3D spatial model according to an exemplary embodiment.
- an object S O standing around a vehicle may be projected onto a three-dimensional curved section C 3D rather than a flat bottom surface. Therefore, when projecting on a 2D plane, the object is mapped to be stretched for a long time, but by projecting on a 3D curved section, there is an advantage of preventing the object from being stretched for a long time.
- the size of the bottom portion of the three-dimensional space model in the present embodiment is easy to adjust only the long and short radius coefficient of the ellipse corresponding to the bottom. Therefore, the mapping virtual plane can be easily adjusted according to how much the peripheral region of interest is set. For example, when the vehicle is parked or slowed down, it becomes more necessary to check the floor around the vehicle. Therefore, it is advantageous to adjust the size of the base of the three-dimensional space model to be larger. It is advantageous to adjust the size of the base of the three-dimensional space model to be smaller since the need for checking becomes large. Therefore, the peripheral image generating apparatus 300 may adjust the size of the bottom surface of the 3D space model in inverse proportion to the traveling speed of the vehicle.
- the three-dimensional space model according to the present invention has a merit that the cut surface is smoothly changed without a sudden bending curve, so that it looks natural without an unnatural bending portion in the view image. And finally, in mapping the input image, it is easy to determine the position of the fiducial point on the mapping virtual plane.
- the fiducial point may be determined according to how much intervals ⁇ and ⁇ in the above equations are to be taken as the fiducial point.
- the peripheral image generating apparatus 300 may generate a view image of the viewpoint of the virtual camera using the image mapped to the virtual surface in operation S820.
- the view image of the viewpoint of the virtual camera may be configured by referring to a predefined lookup table between the image mapped to the virtual plane and the view image of the viewpoint of the virtual camera, or may define a corresponding relationship between the two. Can be made.
- the viewpoint of the virtual camera may be implemented to be determined by at least one of a driving state of the vehicle or a user selection.
- the driving state of the vehicle may be implemented such that the viewpoint of the virtual camera is determined in association with the vehicle's speed, steering angle direction, or gear position.
- the virtual camera viewpoint may be set to have a small tilt angle ⁇ in parallel with the ground or at the front of the vehicle. In this case, the tilt angle ⁇ formed with the ground may be set to be large.
- the virtual camera viewpoint may be set backward.
- the virtual camera viewpoint may be set to the front left side of the vehicle and when the steering angle direction is the right side, the front right side of the vehicle.
- a virtual camera C is displayed in a form of looking down the vehicle V from the top. : UI) screen is provided. Then, the user may adjust the virtual camera pan angle ⁇ by rotating the front of the virtual camera C on the screen, and move the virtual camera C to a desired position (X, Y) by dragging.
- a user interface (UI) screen on which the virtual camera C is displayed may be provided in a form of viewing the vehicle V from the side.
- the user may rotate the front part of the virtual camera C to adjust the tilt angle ⁇ , and the virtual camera C may be dragged to adjust the height Z of the virtual camera C.
- a screen capable of three-dimensionally understanding the virtual camera viewpoint may be provided.
- FIG. 12 is a diagram provided to explain an area where images captured by a plurality of cameras overlap when generating a view image of a virtual camera.
- the front camera (as illustrated in FIG. 12A) may be used.
- the image captured by 210 is mapped to areas 1, 2 and 3, and the image captured by the rear camera 220 is mapped to areas 7, 8 and 9.
- the image captured by the left camera 230 is mapped to areas 1, 4, and 7, and the image captured by the right camera 240 is mapped to areas 3, 6, and 9.
- the first, third, seventh, and ninth regions are overlapping regions photographed by the plurality of cameras.
- the first area is the overlapped area photographed by the front camera 210 and the left camera 230
- the third area is the overlapped area photographed by the front camera 210 and the right camera 240. to be.
- the seventh area is an overlapping area captured by the rear camera 220 and the left camera 230
- the ninth area is an overlapping area captured by the rear camera 220 and the right camera 240. to be.
- area 5 is an area where an image corresponding to the vehicle is displayed.
- the centers L C1 , L C2 , L C3 and L C4 of the overlapping area may be positioned in the diagonal direction of the vehicle.
- the processing for the overlapping area is divided into regions based on the centers of the overlapping areas L C1 , L C3 , L C7 , L C9 , for example, in the case of the overlapping area 3, the area 3-1 is the front camera 210.
- the image 3-2 may be implemented to apply the image captured by the right camera 240.
- the weights are differently applied based on the center L C3 of the overlapping area, and the area 3-1 is processed by giving a higher weight to the image captured by the front camera 210, and the area 3. -2) may be implemented to give a higher weight to the image captured by the right camera 240 to process.
- the remaining overlapping region can also be processed by the same method.
- the center of the overlapping area (L C1 , L C3 ) should be closer to the longitudinal axis direction of the vehicle. Can be.
- the case where the virtual camera viewpoint is set in a direction looking forward from the rear of the vehicle is when the vehicle is driving. In this case, it is necessary for the driver to check the blind spots on the left and right side more naturally and comfortably.
- the image captured by the left and right cameras 230 and 240 is more reflected in the view image.
- the center of the overlapped region to further include an image captured by a camera installed at the side where the virtual camera is located according to the position of the virtual camera viewpoint at the left or right side of the vehicle.
- the center L C1 and L C7 of the overlapping region closer to the longitudinal direction of the vehicle.
- the center L C3 L C9 closer to the longitudinal direction of the vehicle.
- the view image of the virtual camera view generated in step S820 is output to the display device 400 (S830).
- the steps S810 to S830 may be repeated while varying the centers L C1 , L C3 , L C7 , L C9 of the overlapped area and the size of the bottom surface of the three-dimensional space model according to the driving state of the vehicle.
- the view image may be fixed and generated at the viewpoint of the virtual camera set by the user, and may be implemented so as not to be affected by the driving state of the vehicle.
- the surrounding image generating apparatus 300 may be provided with driving position information of the current vehicle in association with a GPS receiving module (not shown).
- the surrounding image generating apparatus 300 may display augmented reality by synthesizing the surrounding buildings and the road information to the view image of the virtual camera view in association with the building and the road information DB around the current driving position of the vehicle.
- FIG. 13 illustrates an example in which building information 1320 and road information 1310 around a vehicle are displayed on a view image of a virtual camera.
- the GPS receiving module may use a GPS receiving device included in the corresponding wireless communication terminal, and may be implemented as a GPS receiving module separately installed in a vehicle. have.
- the building and road information DB around the driving position of the vehicle may be stored in an internal memory of the surrounding image generating apparatus 300 or may be provided in connection with a separate navigation device (not shown).
- the surrounding image generating apparatus 300 may be provided.
- Embodiments of the invention include a computer readable medium containing program instructions for performing various computer-implemented operations.
- This medium records a program for executing the 3D vehicle surrounding image generation method described so far.
- the media may include, alone or in combination with the program instructions, data files, data structures, and the like. Examples of such media include magnetic media such as hard disks, floppy disks and magnetic tape, optical recording media such as CDs and DVDs, floppy disks and program commands such as magnetic-optical media, ROM, RAM and flash memory.
- Hardware devices configured to store and perform such operations.
- the medium may be a transmission medium such as an optical or metal wire, a waveguide, or the like including a carrier wave for transmitting a signal specifying a program command, a data structure, and the like.
- program instructions include not only machine code generated by a compiler, but also high-level language code that can be executed by a computer using an interpreter or the like.
- the present invention can be used in a vehicle surrounding image generation method and apparatus capable of providing a three-dimensional (3D) surrounding image of the vehicle.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mechanical Engineering (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Closed-Circuit Television Systems (AREA)
- Image Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Processing Or Creating Images (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (13)
- 차량에 설치된 복수 개의 카메라로부터 촬영된 영상을 밑면이 평평하고 상부로 갈수록 반경이 넓어지는 용기 형태의 3차원 공간 모델에 의해 정의되는 가상면에 맵핑하는 단계, 그리고상기 가상면에 맵핑된 영상을 이용하여 가상 카메라의 시점의 뷰 영상을 생성하는 단계를 포함하는 3D 차량 주변 영상 생성 방법.
- 제 1 항에서,상기 가상 카메라의 시점은 상기 차량의 주행 상태 및 사용자 선택 중 적어도 하나에 의해 결정되는 3D 차량 주변 영상 생성 방법.
- 제 1 항에서,상기 3차원 공간 모델의 밑면의 크기는 상기 차량의 주행 속도에 반비례하는 3D 차량 주변 영상 생성 방법.
- 제 1 항에서,상기 합성 영상 생성은,상기 가상면에 맵핑된 영상과 상기 가상 카메라 시점의 뷰 영상 사이의 대응 관계가 미리 정의된 룩업 테이블을 참조하여 이루어지는 3D 차량 주변 영상 생성 방법.
- 제 1 항에서,상기 뷰 영상을 생성하기 위해 상기 복수 개의 카메라에 의해 촬영된 영상이 중첩되는 영역의 중심이 상기 가상 카메라의 틸트 각도에 따라 변경되는 3D 차량 주변 영상 생성 방법.
- 제 1 항에서,상기 복수 개의 카메라는 상기 차량의 전방, 우측, 후방, 좌측에 각각 설치되는 전방 카메라, 우측 카메라, 후방 카메라 및 좌측 카메라를 포함하고,상기 가상 카메라의 위치가 상기 차량의 중심을 기준으로 좌측 또는 우측에 위치한 정도에 따라 상기 가상 카메라가 위치한 측에 설치된 카메라에 의해 촬영된 영상을 더 포함하도록 중첩 영역의 중심이 변경되는 3D 차량 주변 영상 생성 방법.
- 제 2 항에서,상기 가상 카메라 시점은 상기 차량의 조향각 방향 또는 기어 위치에 연동되는 3D 차량 주변 영상 생성 방법.
- 제 2 항에서,상기 가상 카메라의 시점에 대한 사용자 선택을 입력받기 위한 유아이(UI)를 표시하는 단계, 그리고상기 유아이를 통해 입력된 사용자 선택에 따라 상기 가상 카메라 시점의 변화를 표시하는 단계를 더 포함하는 3D 차량 주변 영상 생성 방법.
- 제 2 항에 있어서,상기 차량의 주행 위치 주변의 건물 정보 및 도로 정보를 상기 가상 카메라 시점에 따라 상기 뷰 영상에 합성하는 단계를 더 포함하는 3D 차량 주변 영상 생성 방법.
- 차량에 설치된 복수 개의 카메라로부터 촬영된 영상을 입력받고, 상기 입력된 영상을 밑면이 평평하고 상부로 갈수록 반경이 넓어지는 용기 형태의 3차원 공간 모델에 의해 정의되는 가상면에 맵핑하고, 상기 가상면에 맵핑된 영상을 이용하여 가상 카메라의 시점에서 바라본 뷰 영상을 생성하는 주변영상생성장치를 포함하는 차량 주변 영상 표시 시스템.
- 제 10 항에서,상기 3차원 공간 모델의 밑면의 크기는 상기 차량의 주행 속도에 반비례하는 차량 주변 영상 표시 시스템.
- 제 10 항에 있어서,상기 주변영상생성장치는,상기 차량의 주행 위치 주변의 건물 정보 및 도로 정보를 상기 가상 카메라 시점에 따라 상기 뷰 영상에 합성하는 차량 주변 영상 표시 시스템.
- 제 10 항에 있어서,상기 복수 개의 카메라는,상기 촬영된 영상을 상기 주변영상생성장치에 근거리 무선 통신을 통해 전달하는 차량 주변 영상 표시 시스템.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/361,546 US9508189B2 (en) | 2011-11-30 | 2012-09-27 | Method and apparatus for creating 3D image of vehicle surroundings |
CN201280068147.3A CN104093600B (zh) | 2011-11-30 | 2012-09-27 | 用于生成车辆环境的三维图像的方法和装置 |
JP2014544648A JP6095684B2 (ja) | 2011-11-30 | 2012-09-27 | 3d車両周辺映像生成方法および装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2011-0127209 | 2011-11-30 | ||
KR1020110127209A KR101265711B1 (ko) | 2011-11-30 | 2011-11-30 | 3d 차량 주변 영상 생성 방법 및 장치 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013081287A1 true WO2013081287A1 (ko) | 2013-06-06 |
Family
ID=48535702
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2012/007781 WO2013081287A1 (ko) | 2011-11-30 | 2012-09-27 | 3d 차량 주변 영상 생성 방법 및 장치 |
Country Status (5)
Country | Link |
---|---|
US (1) | US9508189B2 (ko) |
JP (1) | JP6095684B2 (ko) |
KR (1) | KR101265711B1 (ko) |
CN (1) | CN104093600B (ko) |
WO (1) | WO2013081287A1 (ko) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2513703A (en) * | 2013-02-28 | 2014-11-05 | Bosch Gmbh Robert | Method and apparatus for three-dimensional imaging of at least a partial region of a vehicle environment |
US20160119539A1 (en) * | 2014-10-28 | 2016-04-28 | Nissan North America, Inc. | Vehicle object monitoring system |
CN110503597A (zh) * | 2019-07-09 | 2019-11-26 | 浙江浩腾电子科技股份有限公司 | 一种基于2d结构模型的车辆视点规范化方法 |
JP2020017960A (ja) * | 2019-08-05 | 2020-01-30 | 株式会社デンソーテン | 画像表示装置、画像表示システム、画像表示方法、及び、プログラム |
JP2021067977A (ja) * | 2019-10-17 | 2021-04-30 | 株式会社東海理化電機製作所 | 画像処理装置、画像処理プログラム |
US11697428B2 (en) | 2020-06-29 | 2023-07-11 | Apollo Intelligent Driving Technology (Beijing) Co., Ltd. | Method and apparatus for 3D modeling |
Families Citing this family (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2447672B (en) | 2007-03-21 | 2011-12-14 | Ford Global Tech Llc | Vehicle manoeuvring aids |
US9969428B2 (en) | 2011-04-19 | 2018-05-15 | Ford Global Technologies, Llc | Trailer backup assist system with waypoint selection |
US9926008B2 (en) | 2011-04-19 | 2018-03-27 | Ford Global Technologies, Llc | Trailer backup assist system with waypoint selection |
US9854209B2 (en) | 2011-04-19 | 2017-12-26 | Ford Global Technologies, Llc | Display system utilizing vehicle and trailer dynamics |
JP5904093B2 (ja) * | 2012-10-30 | 2016-04-13 | 株式会社デンソー | 車載画像生成装置 |
KR101504335B1 (ko) * | 2013-07-29 | 2015-03-23 | 인하대학교 산학협력단 | 어라운드뷰 제공 장치, 방법 및 차량 |
CN104691415B (zh) * | 2013-12-07 | 2017-02-01 | 惠州市德赛西威汽车电子股份有限公司 | 全景泊车辅助装置 |
KR101543159B1 (ko) | 2014-05-02 | 2015-08-10 | 현대자동차주식회사 | 카메라를 이용한 영상 조정 시스템 및 방법 |
CN106464847B (zh) * | 2014-06-20 | 2019-06-25 | 歌乐株式会社 | 影像合成系统和用于其的影像合成装置与影像合成方法 |
JP2016046747A (ja) * | 2014-08-26 | 2016-04-04 | 富士通テン株式会社 | 画像処理装置、画像処理方法、及び、画像表示システム |
CN104385987B (zh) * | 2014-11-14 | 2017-01-11 | 东风汽车有限公司 | 一种汽车监控方法及系统 |
US20160159281A1 (en) * | 2014-12-04 | 2016-06-09 | Hyundai Mobis Co., Ltd. | Vehicle and control method thereof |
US20160176340A1 (en) * | 2014-12-17 | 2016-06-23 | Continental Automotive Systems, Inc. | Perspective shifting parking camera system |
DE102015202863A1 (de) * | 2015-02-17 | 2016-08-18 | Conti Temic Microelectronic Gmbh | Verfahren und Vorrichtung zum verzerrungsfreien Anzeigen einer Fahrzeugumgebung eines Fahrzeuges |
KR101684270B1 (ko) * | 2015-03-13 | 2016-12-20 | 주식회사 레드로버 | 표면윤곽 변화를 고려한 동적 프로젝션 매핑 방법 |
CN104960473B (zh) * | 2015-07-03 | 2017-11-14 | 上海寅喆计算机科技有限公司 | 视角切换方法及装置 |
JP2017030530A (ja) * | 2015-07-31 | 2017-02-09 | スズキ株式会社 | 映像表示システム |
KR102336778B1 (ko) * | 2015-07-31 | 2021-12-08 | 현대오토에버 주식회사 | 스카이라인을 이용한 차량의 측위 방법 |
US9896130B2 (en) | 2015-09-11 | 2018-02-20 | Ford Global Technologies, Llc | Guidance system for a vehicle reversing a trailer along an intended backing path |
JP6575445B2 (ja) * | 2015-09-30 | 2019-09-18 | アイシン精機株式会社 | 車両用画像処理装置 |
US9836060B2 (en) | 2015-10-28 | 2017-12-05 | Ford Global Technologies, Llc | Trailer backup assist system with target management |
US10102610B2 (en) * | 2016-04-05 | 2018-10-16 | Qualcomm Incorporated | Dual fisheye images stitching for spherical video |
US10112646B2 (en) | 2016-05-05 | 2018-10-30 | Ford Global Technologies, Llc | Turn recovery human machine interface for trailer backup assist |
US20170374287A1 (en) * | 2016-06-23 | 2017-12-28 | Werner Lang | System for Visually Depicting Fields of View of a Commercial Vehicle |
DE102016217488A1 (de) * | 2016-09-14 | 2018-03-15 | Robert Bosch Gmbh | Verfahren zum Bereitstellen einer Rückspiegelansicht einer Fahrzeugumgebung eines Fahrzeugs |
JP6419128B2 (ja) * | 2016-10-28 | 2018-11-07 | キヤノン株式会社 | 画像処理装置、画像処理システム、画像処理方法及びプログラム |
US20180152628A1 (en) * | 2016-11-30 | 2018-05-31 | Waymo Llc | Camera peek into turn |
US10970915B2 (en) * | 2017-01-06 | 2021-04-06 | Canon Kabushiki Kaisha | Virtual viewpoint setting apparatus that sets a virtual viewpoint according to a determined common image capturing area of a plurality of image capturing apparatuses, and related setting method and storage medium |
CN106791325B (zh) * | 2017-01-06 | 2022-08-23 | 上海寅家电子科技股份有限公司 | 全景环视系统的色差测光区自动识别系统及自动识别方法 |
US10268201B2 (en) * | 2017-02-28 | 2019-04-23 | Mitsubishi Electric Research Laboratories, Inc. | Vehicle automated parking system and method |
US11049219B2 (en) | 2017-06-06 | 2021-06-29 | Gopro, Inc. | Methods and apparatus for multi-encoder processing of high resolution content |
JP6962036B2 (ja) * | 2017-07-07 | 2021-11-05 | 株式会社アイシン | 周辺監視装置 |
EP3675485B1 (en) * | 2017-08-25 | 2021-03-24 | Nissan Motor Co., Ltd. | Surrounding vehicle display method and surrounding vehicle display apparatus |
US10684625B2 (en) * | 2017-08-30 | 2020-06-16 | Robert Bosch Gmbh | Automated parking for virtual parking spot |
JP7013751B2 (ja) * | 2017-09-15 | 2022-02-01 | 株式会社アイシン | 画像処理装置 |
US10580299B2 (en) * | 2017-10-13 | 2020-03-03 | Waymo Llc | Lane change notification |
US11225193B2 (en) | 2017-10-26 | 2022-01-18 | Harman International Industries, Incorporated | Surround view system and method thereof |
US20180095475A1 (en) * | 2017-11-22 | 2018-04-05 | GM Global Technology Operations LLC | Systems and methods for visual position estimation in autonomous vehicles |
JP2019151304A (ja) * | 2018-03-06 | 2019-09-12 | アイシン精機株式会社 | 周辺監視装置 |
KR102077494B1 (ko) * | 2018-05-11 | 2020-02-14 | (주)넥셀 | 어라운드뷰 제공 방법 및 장치 |
GB2573792B (en) * | 2018-05-17 | 2022-11-09 | Denso Corp | Surround monitoring system for vehicles |
JP7208356B2 (ja) * | 2018-09-26 | 2023-01-18 | コーヒレント・ロジックス・インコーポレーテッド | 任意の世界ビューの生成 |
US10715714B2 (en) | 2018-10-17 | 2020-07-14 | Verizon Patent And Licensing, Inc. | Machine learning-based device placement and configuration service |
KR102212229B1 (ko) * | 2018-11-07 | 2021-02-05 | 삼성전자주식회사 | 이동체에 포함된 카메라 시스템 및 그 제어방법. |
GB2581191B (en) * | 2019-02-08 | 2022-02-16 | Jaguar Land Rover Ltd | Image system for a vehicle |
DE102019207415A1 (de) * | 2019-05-21 | 2020-11-26 | Conti Temic Microelectronic Gmbh | Verfahren zur Erzeugung eines Bildes einer Fahrzeugumgebung und Vorrichtung zur Erzeugung eines Bildes einer Fahrzeugumgebung |
KR102235951B1 (ko) * | 2019-06-20 | 2021-04-05 | 주식회사 아이닉스 | 차량용 영상 생성 장치 및 방법 |
US11109067B2 (en) | 2019-06-26 | 2021-08-31 | Gopro, Inc. | Methods and apparatus for maximizing codec bandwidth in video applications |
US11228781B2 (en) | 2019-06-26 | 2022-01-18 | Gopro, Inc. | Methods and apparatus for maximizing codec bandwidth in video applications |
WO2021070226A1 (ja) * | 2019-10-07 | 2021-04-15 | 三菱電機株式会社 | 仮想カメラ制御装置、仮想カメラ制御方法、及び仮想カメラ制御プログラム |
EP4042676A4 (en) * | 2019-10-07 | 2022-11-23 | Gentex Corporation | 3D DISPLAY SYSTEM FOR CAMERA MONITORING SYSTEM |
US11481863B2 (en) | 2019-10-23 | 2022-10-25 | Gopro, Inc. | Methods and apparatus for hardware accelerated image processing for spherical projections |
JP7521345B2 (ja) * | 2020-09-15 | 2024-07-24 | マツダ株式会社 | 車両用表示装置 |
JP2022101979A (ja) * | 2020-12-25 | 2022-07-07 | 株式会社デンソー | 画像生成装置、画像生成方法 |
KR102473407B1 (ko) * | 2020-12-28 | 2022-12-05 | 삼성전기주식회사 | 틸트 카메라를 이용한 차량의 svm 시스템 |
KR102513762B1 (ko) * | 2021-03-02 | 2023-03-27 | 주식회사 켐트로닉스 | 사이드미러 가상 영상 생성 장치 및 방법 |
CN113345074A (zh) * | 2021-06-07 | 2021-09-03 | 苏州易航远智智能科技有限公司 | 一种车载3d环视影像显示方法 |
US11691566B2 (en) * | 2021-11-17 | 2023-07-04 | GM Global Technology Operations LLC | Intelligent vehicle systems and control logic for surround view augmentation with object model recognition |
WO2024057060A1 (en) * | 2022-09-13 | 2024-03-21 | Vinai Artificial Intelligence Application And Research Joint Stock Company | Surround view monitoring system and method |
DE102022214210A1 (de) | 2022-12-21 | 2024-06-27 | Robert Bosch Gesellschaft mit beschränkter Haftung | Verfahren zur Darstellung einer erweiterten Ansicht der Umgebung eines Fahrzeugs, Computerprogramm, Rechenvorrichtung und Fahrzeug |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20010112433A (ko) * | 1999-04-16 | 2001-12-20 | 마츠시타 덴끼 산교 가부시키가이샤 | 화상처리장치 및 감시시스템 |
KR20060130425A (ko) * | 2005-06-14 | 2006-12-19 | 엘지전자 주식회사 | 건물정보 표시방법 |
KR20100005971A (ko) * | 2008-07-08 | 2010-01-18 | 현대자동차주식회사 | 차량 전방위 감시 시스템 |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6498620B2 (en) * | 1993-02-26 | 2002-12-24 | Donnelly Corporation | Vision system for a vehicle including an image capture device and a display system having a long focal length |
US6891563B2 (en) * | 1996-05-22 | 2005-05-10 | Donnelly Corporation | Vehicular vision system |
JP2001527372A (ja) * | 1997-12-31 | 2001-12-25 | ジェンテクス・コーポレーション | 車両視覚システム |
EP2259220A3 (en) * | 1998-07-31 | 2012-09-26 | Panasonic Corporation | Method and apparatus for displaying image |
JP3298851B2 (ja) * | 1999-08-18 | 2002-07-08 | 松下電器産業株式会社 | 多機能車載カメラシステムと多機能車載カメラの画像表示方法 |
KR100866450B1 (ko) * | 2001-10-15 | 2008-10-31 | 파나소닉 주식회사 | 차량 주위 감시 장치 및 그 조정 방법 |
JP3871614B2 (ja) * | 2002-06-12 | 2007-01-24 | 松下電器産業株式会社 | 運転支援装置 |
JP2006287826A (ja) * | 2005-04-05 | 2006-10-19 | Nissan Motor Co Ltd | 車両用画像生成装置および方法 |
KR100674805B1 (ko) | 2005-06-14 | 2007-01-29 | 엘지전자 주식회사 | 카메라 영상과 지도 데이터 간에 건물 매칭 방법 |
US7728869B2 (en) | 2005-06-14 | 2010-06-01 | Lg Electronics Inc. | Matching camera-photographed image with map data in portable terminal and travel route guidance method |
US7728879B2 (en) * | 2006-08-21 | 2010-06-01 | Sanyo Electric Co., Ltd. | Image processor and visual field support device |
US8199975B2 (en) * | 2006-12-12 | 2012-06-12 | Cognex Corporation | System and method for side vision detection of obstacles for vehicles |
JP5369465B2 (ja) * | 2008-03-25 | 2013-12-18 | 富士通株式会社 | 車両用画像処理装置、車両用画像処理方法、車両用画像処理プログラム |
TWI332454B (en) * | 2008-09-10 | 2010-11-01 | Univ Nat Chiao Tung | Intelligent vehicle traffic safety supply system |
JP5182042B2 (ja) | 2008-11-28 | 2013-04-10 | 富士通株式会社 | 画像処理装置、画像処理方法及びコンピュータプログラム |
JP2010185761A (ja) * | 2009-02-12 | 2010-08-26 | Toyota Motor Corp | ナビゲーションシステム、道路地図表示方法 |
JP5439890B2 (ja) * | 2009-03-25 | 2014-03-12 | 富士通株式会社 | 画像処理方法、画像処理装置及びプログラム |
JP5271154B2 (ja) | 2009-05-29 | 2013-08-21 | 富士通テン株式会社 | 画像生成装置及び画像表示システム |
JP2010274813A (ja) * | 2009-05-29 | 2010-12-09 | Fujitsu Ten Ltd | 画像生成装置及び画像表示システム |
US8502860B2 (en) | 2009-09-29 | 2013-08-06 | Toyota Motor Engineering & Manufacturing North America (Tema) | Electronic control system, electronic control unit and associated methodology of adapting 3D panoramic views of vehicle surroundings by predicting driver intent |
JP5548002B2 (ja) * | 2010-03-25 | 2014-07-16 | 富士通テン株式会社 | 画像生成装置、画像表示システム及び画像生成方法 |
DE102010042063B4 (de) * | 2010-10-06 | 2021-10-28 | Robert Bosch Gmbh | Verfahren und Vorrichtung zum Bestimmen von aufbereiteten Bilddaten über ein Umfeld eines Fahrzeugs |
-
2011
- 2011-11-30 KR KR1020110127209A patent/KR101265711B1/ko active IP Right Grant
-
2012
- 2012-09-27 CN CN201280068147.3A patent/CN104093600B/zh active Active
- 2012-09-27 JP JP2014544648A patent/JP6095684B2/ja active Active
- 2012-09-27 WO PCT/KR2012/007781 patent/WO2013081287A1/ko active Application Filing
- 2012-09-27 US US14/361,546 patent/US9508189B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20010112433A (ko) * | 1999-04-16 | 2001-12-20 | 마츠시타 덴끼 산교 가부시키가이샤 | 화상처리장치 및 감시시스템 |
KR20060130425A (ko) * | 2005-06-14 | 2006-12-19 | 엘지전자 주식회사 | 건물정보 표시방법 |
KR20100005971A (ko) * | 2008-07-08 | 2010-01-18 | 현대자동차주식회사 | 차량 전방위 감시 시스템 |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2513703A (en) * | 2013-02-28 | 2014-11-05 | Bosch Gmbh Robert | Method and apparatus for three-dimensional imaging of at least a partial region of a vehicle environment |
GB2513703B (en) * | 2013-02-28 | 2019-11-06 | Bosch Gmbh Robert | Method and apparatus for three-dimensional imaging of at least a partial region of a vehicle environment |
US20160119539A1 (en) * | 2014-10-28 | 2016-04-28 | Nissan North America, Inc. | Vehicle object monitoring system |
US9880253B2 (en) * | 2014-10-28 | 2018-01-30 | Nissan North America, Inc. | Vehicle object monitoring system |
CN110503597A (zh) * | 2019-07-09 | 2019-11-26 | 浙江浩腾电子科技股份有限公司 | 一种基于2d结构模型的车辆视点规范化方法 |
CN110503597B (zh) * | 2019-07-09 | 2023-05-16 | 浙江浩腾电子科技股份有限公司 | 一种基于2d结构模型的车辆视点规范化方法 |
JP2020017960A (ja) * | 2019-08-05 | 2020-01-30 | 株式会社デンソーテン | 画像表示装置、画像表示システム、画像表示方法、及び、プログラム |
JP2021067977A (ja) * | 2019-10-17 | 2021-04-30 | 株式会社東海理化電機製作所 | 画像処理装置、画像処理プログラム |
JP7442029B2 (ja) | 2019-10-17 | 2024-03-04 | 株式会社東海理化電機製作所 | 画像処理装置、画像処理プログラム |
US11697428B2 (en) | 2020-06-29 | 2023-07-11 | Apollo Intelligent Driving Technology (Beijing) Co., Ltd. | Method and apparatus for 3D modeling |
EP3822850B1 (en) * | 2020-06-29 | 2023-09-27 | Beijing Baidu Netcom Science And Technology Co., Ltd. | Method and apparatus for 3d modeling |
Also Published As
Publication number | Publication date |
---|---|
JP2015507386A (ja) | 2015-03-05 |
JP6095684B2 (ja) | 2017-03-15 |
US9508189B2 (en) | 2016-11-29 |
US20140347450A1 (en) | 2014-11-27 |
KR101265711B1 (ko) | 2013-05-20 |
CN104093600B (zh) | 2016-12-07 |
CN104093600A (zh) | 2014-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2013081287A1 (ko) | 3d 차량 주변 영상 생성 방법 및 장치 | |
CN110381255B (zh) | 应用360全景环视技术的车载视频监控系统及方法 | |
WO2012176945A1 (ko) | 차량 주변 시각화를 위한 3차원 영상 합성장치 및 그 방법 | |
CN104506761B (zh) | 一种360度全景立体摄像机 | |
CN111739101B (zh) | 一种消除车辆a柱盲区的装置和方法 | |
US20130141547A1 (en) | Image processing apparatus and computer-readable recording medium | |
KR20040111329A (ko) | 운전 지원 장치 | |
US11380111B2 (en) | Image colorization for vehicular camera images | |
CN113362228A (zh) | 基于改进畸变校正和标记拼接的环视图像拼接方法及系统 | |
CN108153502B (zh) | 基于透明屏幕的手持式增强现实显示方法及装置 | |
KR102057021B1 (ko) | 패널 변환 | |
EP2939211B1 (en) | Method and system for generating a surround view | |
US20160037154A1 (en) | Image processing system and method | |
US10748264B2 (en) | Image processing apparatus and image processing method | |
CN111435540A (zh) | 一种车载环视系统的环视图拼接方法 | |
WO2015056826A1 (ko) | 카메라의 영상 처리 장치 및 방법 | |
CN114881863B (zh) | 一种图像拼接方法、电子设备及计算机可读存储介质 | |
CN206287929U (zh) | 一种3d全景泊车装置及系统 | |
CN106878698B (zh) | 基于光路采集的移动裸眼三维虚拟现实的方法和系统 | |
JP2004265396A (ja) | 映像生成システム及び映像生成方法 | |
KR101265710B1 (ko) | 차량 설치 카메라 외부 파라미터 추정 방법 및 장치 | |
CN112802109B (zh) | 一种汽车鸟瞰全景图生成方法 | |
CN113989112A (zh) | 一种环视图像融合拼接方法、装置、电子设备及存储介质 | |
KR102079332B1 (ko) | 3d 차량 주변 영상 생성 방법 및 시스템 | |
KR101351911B1 (ko) | 카메라의 영상 처리 장치 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12852879 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14361546 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2014544648 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 12852879 Country of ref document: EP Kind code of ref document: A1 |