WO2007111241A1 - 道路画像解析装置及び道路画像解析方法 - Google Patents
道路画像解析装置及び道路画像解析方法 Download PDFInfo
- Publication number
- WO2007111241A1 WO2007111241A1 PCT/JP2007/055986 JP2007055986W WO2007111241A1 WO 2007111241 A1 WO2007111241 A1 WO 2007111241A1 JP 2007055986 W JP2007055986 W JP 2007055986W WO 2007111241 A1 WO2007111241 A1 WO 2007111241A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- continuous component
- component
- road
- continuous
- image analysis
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 13
- 239000000284 extract Substances 0.000 claims abstract description 9
- 238000000605 extraction Methods 0.000 claims description 24
- 238000010191 image analysis Methods 0.000 claims description 23
- 238000003384 imaging method Methods 0.000 claims description 21
- 238000003703 image analysis method Methods 0.000 claims description 14
- 238000004364 calculation method Methods 0.000 claims description 11
- 238000007781 pre-processing Methods 0.000 abstract description 5
- 238000010586 diagram Methods 0.000 description 14
- 238000004458 analytical method Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 2
- 230000004304 visual acuity Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/09—Arrangements for giving variable traffic instructions
- G08G1/0962—Arrangements for giving variable traffic instructions having an indicator mounted inside the vehicle, e.g. giving voice messages
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/02—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to ambient conditions
- B60W40/06—Road conditions
- B60W40/072—Curvature of the road
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/02—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to ambient conditions
- B60W40/06—Road conditions
- B60W40/076—Slope angle of the road
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/457—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by analysing connectivity, e.g. edge linking, connected component analysis or slices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/48—Extraction of image or video features by mapping characteristic values of the pattern into a parameter space, e.g. Hough transformation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/588—Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
- G06V20/647—Three-dimensional objects by matching two-dimensional images to three-dimensional objects
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/403—Image sensing, e.g. optical camera
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2552/00—Input parameters relating to infrastructure
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
- G06T2207/30256—Lane; Road marking
Definitions
- the present invention relates to a road image analysis apparatus and a road image analysis method for analyzing road images and extracting road markings such as a road center line, a road boundary line, and a road outer line, guardrails, and the like.
- Japanese Patent Laid-Open No. 6-266828 discloses a technique for detecting side walls such as card rails by calculating three-dimensional position data from a stereo image acquired by a stereo camera.
- both the road marking and the guardrail are linear images, and thus it is difficult to distinguish them. was there.
- the present invention has been made in view of the above-described conventional problems, and its purpose is to provide road marking and guidance. It is an object of the present invention to provide a road image analysis apparatus and a road image analysis method capable of clearly and rapidly distinguishing from a single rail and acquiring accurate position information.
- the present invention provides a road image analysis apparatus, two imaging means for photographing the same position from different directions, and image data acquired by one of the imaging means.
- a continuous component extracting means for extracting a continuous component from the above, a three-dimensional position calculating means for calculating three-dimensional position information of the continuous component by combining the continuous component and image data acquired by the other of the imaging means, and the 3 Identification means for identifying the continuous component based on the three-dimensional position information calculated by the three-dimensional position calculation means.
- the continuous component extracting means extracts a linear component from the image data acquired by one of the imaging means, and based on the position, inclination, and relationship between the linear components, the continuous component is described above. It is characterized by extracting components.
- the three-dimensional position calculation means calculates the three-dimensional position information by performing a matching process between the end points of the continuous components and the image data acquired by the other of the imaging means.
- the identifying means identifies the continuous component based on the height information of the continuous component included in the three-dimensional position information.
- the continuous component extracting means, the three-dimensional position calculating means, and the identifying means are characterized in that a sub-area set in a direction orthogonal to the traveling direction of the road is used as a processing unit.
- the present invention is a road image analysis method, the step of capturing the same position from two different directions, the step of extracting a continuous component from the captured one image data, and the continuous component Combining the other captured image data, calculating three-dimensional position information of the continuous component, and identifying the continuous component based on the calculated three-dimensional position information. To do.
- the step of extracting the continuous component a linear component is extracted from the one of the captured image data, and the continuous component is calculated based on the position, inclination, and relationship between the linear components. It is characterized by extracting.
- the step of calculating the three-dimensional position information of the continuous component includes calculating the three-dimensional position information by performing matching processing between the end point of the continuous component and the other image data that has been captured.
- the step of identifying the continuous component is characterized in that the continuous component is identified based on height information of the continuous component included in the three-dimensional position information.
- the step of extracting the continuous component, the step of calculating the three-dimensional position information of the continuous component, and the step of identifying the continuous component are subareas set in a direction orthogonal to the road traveling direction. Is executed as a processing unit.
- FIG. 1 is a block diagram of a configuration example of a road image analysis device according to the present invention.
- FIG. 2 is a functional block diagram showing each function of the road image analysis device according to the present invention.
- FIG. 3 is an explanatory diagram of a shooting area by an imaging unit.
- FIG. 4 is a flowchart of an operation example of the road image analysis device according to the present invention.
- FIG. 5 is a diagram showing an example of a main image in which a sub area is set.
- FIG. 6 is an explanatory diagram of edge extraction processing.
- FIG. 7 is a diagram showing an example of extracted linear components.
- FIG. 8 is a diagram showing an example of a result of continuous component extraction processing.
- FIG. 9 is an explanatory diagram of matching processing.
- FIG. 10 is a diagram showing a display example of a road image analysis result.
- FIG. 1 shows a block diagram of a configuration example of a road image analysis apparatus according to the present invention.
- This road image analysis apparatus can be realized by, for example, a personal computer.
- the road image analysis apparatus includes a first camera 10, a second camera 12, interfaces 14 and 16, a processor 18, a storage unit 20, a display unit 22, and an operation unit 24.
- the first camera 10 is a stereo camera provided as a set of two to obtain a stereo image.
- the second camera 12 is the other of the stereo cameras, for example, a right camera.
- the first camera 10 and the second camera 12 take a color image or a monochrome image from the same position in different directions, and constitute an imaging means of the present invention.
- the interfaces 14 and 16 receive the image data acquired by the first camera 10 and the second camera 12, respectively, and transfer them to the processor 18 and the like.
- the processor 18 is constituted by a CPU (Central Processing Unit) and the like, and controls these operations while exchanging data with the interfaces 14 and 16, the storage unit 20, the display unit 22, and the operation unit 24.
- CPU Central Processing Unit
- the storage unit 20 includes a working RAM (random access memory) for the processor 18, a ROM (read-only memory) in which a program executed by the processor 18 is stored, a flash memory as a nonvolatile memory, and a magnetic storage.
- a computer-readable storage medium composed of a medium or the like.
- road markings and guardrail position information which are analysis results by the present apparatus, may be stored.
- the display unit 22 is configured by a liquid crystal display (LCD), for example, and displays an analysis result of the road image analysis device, a user interface for a user to input an operation instruction to the road image analysis device, and the like.
- LCD liquid crystal display
- the operation unit 24 includes a pointing device such as a mouse and an input device such as a keyboard, and is used for inputting data by the user.
- FIG. 2 is a functional block diagram showing the functions of the road image analysis device.
- the road image analysis apparatus includes an imaging unit 26, a preprocessing unit 28, an edge extraction unit 30, a straight line extraction unit 32, a straight line component analysis unit 34, a matching processing unit 36, a map coordinate information calculation unit 38, and an identification unit 40. And a position information acquisition unit 42.
- the imaging unit 26 is realized by the first camera 10, the second camera 12, and the interfaces 14 and 16, and the image data acquired by the second camera 12 using the image data acquired by the first camera 10 as main image data. Are output as auxiliary image data.
- the pre-processing unit 28 is realized by the processor 18, facilitates edge extraction processing by image enhancement processing or the like on the main image data, and also in a plurality of sub-areas that are basic units of image analysis. A process of dividing the image is executed.
- the sub-area is road driving Set to be orthogonal to the direction.
- the edge extraction unit 30 is realized by the processor 18, and performs extraction of edge components of bright and dark patterns and dark and bright patterns in each subarea.
- the edge component is an outline of a columnar structure such as a road marking, a guardrail, a utility pole, etc. included in the main image data.
- a first-order horizontal differential filter can be used for the edge extraction.
- the straight line extraction unit 32 is realized by the processor 18, and extracts a straight line component from the edge component extracted by the edge extraction unit 30. This linear component extraction is performed using, for example, Hough transform.
- the straight line component analysis unit 34 is realized by the processor 18 and extracts continuous components in the main image data based on the position and inclination of the straight line components extracted by the straight line extraction unit 32, the relationship between the straight line components, and the like.
- the continuous component is a linear component that repeatedly exists in a certain direction in the image or is continuous in a certain direction.
- the former is a road boundary line drawn repeatedly at a predetermined length on a road or a utility pole provided at a predetermined length at the end of the road.
- the road outer line or guardrail drawn as a continuous line on the road is an example of the former. Etc. is an example of the latter.
- the continuity of such linear components is determined by the linear component analysis unit 34 based on the positional relationship of the linear components within each subarea and between each subarea.
- linear components of pole-like structures such as utility poles can be removed by their inclination and excluded from the target of linear analysis.
- the relationship between the straight line components is, for example, a distance between the straight line components, and is used to extract a road marking drawn on the road. That is, since the width between the road markings and the width between the road markings is preliminarily divided, it is possible to determine that the combination of straight line components that are clearly larger or smaller than the above width and that make up the distance is not a straight line component of the road markings. it can.
- edge extraction unit 30, the line extraction unit 32, and the line component analysis unit 34 described above constitute the continuous component extraction unit of the present invention.
- the matching processing unit 36 is realized by the processor 18, and performs matching processing between the vertices of the continuous components of the main image data and the auxiliary image data! ⁇ Based on information such as parallax between corresponding points and camera position Triangulation calculates relative 3D position information for each continuous line segment Put out.
- the map coordinate information calculation unit 38 is realized by the processor 18, and is based on the coordinate information acquired by the position information acquisition unit 42 and the relative three-dimensional position information obtained by the matching processing unit 36. Coordinate information is calculated.
- the matching processing unit 36 constitutes a three-dimensional position calculation unit of the present invention.
- the identification unit 40 is realized by the processor 18, and identifies whether the continuous component is a force guard rail that is a road marking based on the height information of each continuous component included in the three-dimensional position information.
- the position information acquisition unit 42 is realized by a GPS (Global Positioning System), a gyro, or the like, obtains an imaging position and orientation by the imaging unit 26, determines coordinates of corresponding image data, and serves as coordinate information. Output to the map coordinate information calculation unit 38.
- GPS Global Positioning System
- gyro gyro
- FIG. 3 is an explanatory diagram of a shooting area by the imaging unit 26 mounted on the vehicle.
- the arrangement direction of the first camera 10 and the second camera 12 of the imaging unit 26 is represented by X
- the photographing direction (the traveling direction of the vehicle) is represented by Z.
- the shooting area by the left camera as the first camera 10 is indicated by ⁇
- the shooting area by the right camera as the second camera is indicated by / 3. It is also possible to attach the camera in the direction opposite to the traveling direction of the vehicle.
- the image data in the shooting area a is main image data
- the image data in the shooting area ⁇ is auxiliary image data.
- the matching processing by the matching processing unit 36 is performed on the image data of the object existing in the overlapping area ⁇ where the imaging area ex and the imaging area ⁇ overlap.
- FIG. 4 shows a flow of an example of the operation of the road image analysis apparatus that is useful for the present embodiment described in FIGS. 1 and 2.
- the preprocessing unit 28 acquires main image data from the imaging unit 26 (Sl).
- the image data captured by the first camera 10 is the main image data.
- the image data captured by the second camera 12 may be the main image data.
- the pre-processing unit 28 performs the above-described initialization process on the acquired main image data, and sets a plurality of sub-areas (S2).
- FIG. 5 shows a main image in which a sub area is set.
- a plurality of sub-areas 44 are set in the main image; each is shown by being separated by a
- the edge extraction unit 30 extracts edge components for each of the subareas 44 (S3).
- This edge extraction processing is performed by recognizing brightness and color light and dark patterns and dark and light light patterns using a primary horizontal differential filter in each sub-area of the initialized main image data. Since the pillar structure such as road markings, guardrails, and utility poles in the main image shown in Fig. 5 has high brightness against the background of roads, the boundary between light and dark can be recognized as an edge. .
- FIG. 6 is an explanatory diagram of edge extraction processing.
- each edge component extracted by the edge extraction unit 30 is indicated by a one-dot chain line.
- the straight line extraction unit 32 analyzes the extracted edge component and extracts a straight line component.
- FIG. 7 an example of the extracted straight line component is indicated by a solid line.
- the extracted linear components are road markings, guardrails, utility poles, or other columnar structures.
- the linear component analysis unit 34 extracts a continuous component to be analyzed based on the position and inclination of the linear component, the relationship between the straight lines, and the like (S5).
- this continuous component it is preferable to exclude columnar structures such as utility poles depending on the slope of the linear component and the position on the road. As a result, subsequent processing can be narrowed down to road markings and guardrails, and processing can be simplified.
- FIG. 8 shows an example of the result of the continuous component extraction process.
- a linear component that is repeatedly present in a certain direction or continuously in a certain direction in the image is extracted.
- the matching processing unit 36 performs matching processing between the vertices of the continuous component and the auxiliary image data (S6).
- FIGS. 9A and 9B are explanatory diagrams of the matching process.
- FIG. 9 (a) shows the continuous components extracted by the linear component analysis unit 34
- FIG. 9 (b) shows the matching result with the auxiliary image.
- the matching processing unit 36 performs matching between the end points of the continuous components shown in FIG. 9 (a) and the auxiliary image data by, for example, the area correlation method, and based on information such as the parallax between the corresponding points, the camera position, and the posture. Calculate relative 3D position information of each continuous component by triangulation To do.
- Fig. 9 (a) 12 examples of end points (al to al2) are shown. The number of end points is not limited to the above 12 points, and may be larger.
- the corresponding point search is performed only at the end points instead of all the points of the continuous component, so that the matching processing can be performed with high accuracy and high speed.
- the map coordinate information calculation unit 38 calculates coordinate information on the map based on the coordinate information acquired by the position information acquisition unit 42 and the relative three-dimensional position information obtained by the matching processing unit 36. Further, the identification unit 40 identifies whether the continuous component is a force guard rail that is a road marking based on the height information of each continuous component included in the coordinate information on the map (S7). This process can be executed by setting a predetermined threshold value for the height information, using a continuous component having a height higher than the threshold value as a guardrail and using a low continuous component as a road marking. Note that the identification processing may be performed based on relative coordinate information, and thereafter the coordinate information force acquired by the position information acquisition unit 42 may also calculate the coordinate information on the map.
- FIG. 10 shows an example of this display.
- the coordinate values of X, ⁇ , and ⁇ are shown for each attribute of the road marking and guardrail.
- the coordinate information may be stored in the storage unit 20 as a database.
- the linear component analysis unit 34 narrows down the analysis target to the road marking and guard rail, and the matching processing unit 36 and the identification unit 40 identify the road marking and the guard rail. .
- road markings and guardrails can be clearly distinguished by a combination of simple processes, and accurate position information can be acquired.
- road markings and guardrails can be clearly and quickly distinguished, and accurate position information can be acquired.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Mathematical Physics (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Automation & Control Theory (AREA)
- Traffic Control Systems (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Closed-Circuit Television Systems (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2007800111981A CN101410872B (zh) | 2006-03-28 | 2007-03-23 | 道路图像解析装置及道路图像解析方法 |
US12/225,583 US8126209B2 (en) | 2006-03-28 | 2007-03-23 | Road image analyzing apparatus and road image analyzing method |
EP07739427.8A EP2000979A4 (en) | 2006-03-28 | 2007-03-23 | DEVICE AND METHOD FOR VIDEO ROAD IMAGE ANALYSIS |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006-089496 | 2006-03-28 | ||
JP2006089496A JP4632987B2 (ja) | 2006-03-28 | 2006-03-28 | 道路画像解析装置及び道路画像解析方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007111241A1 true WO2007111241A1 (ja) | 2007-10-04 |
Family
ID=38541159
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2007/055986 WO2007111241A1 (ja) | 2006-03-28 | 2007-03-23 | 道路画像解析装置及び道路画像解析方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US8126209B2 (ja) |
EP (1) | EP2000979A4 (ja) |
JP (1) | JP4632987B2 (ja) |
CN (1) | CN101410872B (ja) |
WO (1) | WO2007111241A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101901343A (zh) * | 2010-07-20 | 2010-12-01 | 同济大学 | 基于立体约束的遥感影像道路提取方法 |
CN104021693A (zh) * | 2014-06-25 | 2014-09-03 | 耿直 | 基于雷达探测的车行道标示方法及车行道标示系统 |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110205365A1 (en) * | 2008-10-28 | 2011-08-25 | Pasco Corporation | Road measurement device and method for measuring road |
JP5298991B2 (ja) * | 2009-03-18 | 2013-09-25 | 富士通株式会社 | 車両認識装置及びプログラム |
WO2012014627A1 (ja) | 2010-07-29 | 2012-02-02 | 本田技研工業株式会社 | 車両周辺監視装置 |
US9058247B2 (en) * | 2010-09-08 | 2015-06-16 | Toyota Jidosha Kabushiki Kaisha | Risk potential calculation apparatus |
JP5746996B2 (ja) * | 2012-03-21 | 2015-07-08 | 日立オートモティブシステムズ株式会社 | 道路環境認識装置 |
US8897543B1 (en) | 2012-05-18 | 2014-11-25 | Google Inc. | Bundle adjustment based on image capture intervals |
US8965107B1 (en) | 2012-05-18 | 2015-02-24 | Google Inc. | Feature reduction based on local densities for bundle adjustment of images |
JP6168794B2 (ja) * | 2012-05-31 | 2017-07-26 | キヤノン株式会社 | 情報処理方法および装置、プログラム。 |
WO2014096398A1 (en) | 2012-12-21 | 2014-06-26 | Institute Of Technology Blanchardstown | System and method for multiline retroreflection measurement of road markings |
DE102013224791A1 (de) * | 2013-12-03 | 2015-06-03 | Continental Teves Ag & Co. Ohg | Verfahren zur Erkennung von wenigstens einer Fahrspurmarkierung einer einem Fahrzeug vorausliegenden Fahrspur |
JP6936098B2 (ja) * | 2017-09-29 | 2021-09-15 | トヨタ自動車株式会社 | 対象物推定装置 |
WO2021159397A1 (zh) * | 2020-02-13 | 2021-08-19 | 华为技术有限公司 | 车辆可行驶区域的检测方法以及检测装置 |
WO2021241482A1 (ja) * | 2020-05-25 | 2021-12-02 | 国立大学法人静岡大学 | 経路検出装置 |
KR102520189B1 (ko) * | 2021-03-02 | 2023-04-10 | 네이버랩스 주식회사 | 무인 비행체 또는 항공기에 의해 촬영된 항공 영상에 기반하여 hd 맵을 생성하는 방법 및 시스템 |
CN114046749B (zh) * | 2021-10-26 | 2022-07-08 | 刘红霞 | 预制混凝土构件点状凹坑结合面粗糙度检测方法及系统 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01242916A (ja) | 1988-03-24 | 1989-09-27 | Aisin Seiki Co Ltd | 車上距離検出装置 |
JPH04253286A (ja) * | 1991-01-29 | 1992-09-09 | Nissan Motor Co Ltd | 路面パラメータの推定方法および推定装置 |
JPH06266828A (ja) | 1993-03-12 | 1994-09-22 | Fuji Heavy Ind Ltd | 車輌用車外監視装置 |
JPH0778234A (ja) * | 1993-06-30 | 1995-03-20 | Nissan Motor Co Ltd | 走行路検出装置 |
JP2003281552A (ja) * | 2002-03-25 | 2003-10-03 | Toshiba Corp | 画像処理装置及びその方法 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3522317B2 (ja) * | 1993-12-27 | 2004-04-26 | 富士重工業株式会社 | 車輌用走行案内装置 |
JP3437671B2 (ja) * | 1995-03-17 | 2003-08-18 | 株式会社東芝 | ランドマーク認識装置およびランドマーク認識方法 |
JP3337197B2 (ja) * | 1997-04-04 | 2002-10-21 | 富士重工業株式会社 | 車外監視装置 |
JP3516856B2 (ja) * | 1998-01-30 | 2004-04-05 | 富士重工業株式会社 | 車外監視装置 |
KR100302724B1 (ko) * | 1999-03-12 | 2001-09-22 | 이계안 | 차선이탈 경보장치의 도로 모델링 방법 |
US6963661B1 (en) * | 1999-09-09 | 2005-11-08 | Kabushiki Kaisha Toshiba | Obstacle detection system and method therefor |
JP3352655B2 (ja) * | 1999-09-22 | 2002-12-03 | 富士重工業株式会社 | 車線認識装置 |
JP2001101415A (ja) * | 1999-09-29 | 2001-04-13 | Fujitsu Ten Ltd | 画像認識装置および画像処理装置 |
EP1504276B1 (en) * | 2002-05-03 | 2012-08-08 | Donnelly Corporation | Object detection system for vehicle |
JP4377665B2 (ja) * | 2003-12-01 | 2009-12-02 | 本田技研工業株式会社 | 位置検出用マーク、並びに、マーク検出装置、その方法及びそのプログラム |
-
2006
- 2006-03-28 JP JP2006089496A patent/JP4632987B2/ja active Active
-
2007
- 2007-03-23 EP EP07739427.8A patent/EP2000979A4/en not_active Ceased
- 2007-03-23 WO PCT/JP2007/055986 patent/WO2007111241A1/ja active Application Filing
- 2007-03-23 CN CN2007800111981A patent/CN101410872B/zh not_active Expired - Fee Related
- 2007-03-23 US US12/225,583 patent/US8126209B2/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH01242916A (ja) | 1988-03-24 | 1989-09-27 | Aisin Seiki Co Ltd | 車上距離検出装置 |
JPH04253286A (ja) * | 1991-01-29 | 1992-09-09 | Nissan Motor Co Ltd | 路面パラメータの推定方法および推定装置 |
JPH06266828A (ja) | 1993-03-12 | 1994-09-22 | Fuji Heavy Ind Ltd | 車輌用車外監視装置 |
JPH0778234A (ja) * | 1993-06-30 | 1995-03-20 | Nissan Motor Co Ltd | 走行路検出装置 |
JP2003281552A (ja) * | 2002-03-25 | 2003-10-03 | Toshiba Corp | 画像処理装置及びその方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2000979A4 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101901343A (zh) * | 2010-07-20 | 2010-12-01 | 同济大学 | 基于立体约束的遥感影像道路提取方法 |
CN104021693A (zh) * | 2014-06-25 | 2014-09-03 | 耿直 | 基于雷达探测的车行道标示方法及车行道标示系统 |
Also Published As
Publication number | Publication date |
---|---|
EP2000979A2 (en) | 2008-12-10 |
JP2007265038A (ja) | 2007-10-11 |
US20090274362A1 (en) | 2009-11-05 |
US8126209B2 (en) | 2012-02-28 |
CN101410872B (zh) | 2012-06-20 |
CN101410872A (zh) | 2009-04-15 |
JP4632987B2 (ja) | 2011-02-16 |
EP2000979A9 (en) | 2009-03-04 |
EP2000979A4 (en) | 2013-10-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2007111241A1 (ja) | 道路画像解析装置及び道路画像解析方法 | |
US11320833B2 (en) | Data processing method, apparatus and terminal | |
CN105512646B (zh) | 一种数据处理方法、装置及终端 | |
US9415723B2 (en) | Image integration unit and image integration method | |
US11049376B2 (en) | Setting assistance device, setting assistance method, and program recording medium | |
JP2012220489A (ja) | 道案内情報のウィンドシールド表示装置及びその方法 | |
JP6239186B2 (ja) | 表示制御装置及び表示制御方法及び表示制御プログラム | |
KR101261409B1 (ko) | 영상 내 노면표시 인식시스템 | |
JP2003329411A (ja) | カメラ校正装置 | |
JP2003123197A (ja) | 道路標示等認識装置 | |
US20180247139A1 (en) | Landmark Recognition Device and Landmark Recognition Method | |
JP2008299458A (ja) | 車両監視装置および車両監視方法 | |
JP2011170599A (ja) | 屋外構造物計測装置及び屋外構造物計測方法 | |
CN112798811A (zh) | 速度测量方法、装置和设备 | |
Ghilardi et al. | A new approach for automatic detection of tactile paving surfaces in sidewalks | |
KR20130089068A (ko) | 카메라를 이용한 차량 위치 보정 장치 및 방법 | |
JP4762026B2 (ja) | 道路標識データベース構築装置 | |
TWI682361B (zh) | 路面影像重建與載具定位之方法與系統 | |
JP2007127478A (ja) | 追跡対象物速度検出装置および追跡対象物速度検出方法 | |
JP5120627B2 (ja) | 画像処理装置および画像処理プログラム | |
JP3994954B2 (ja) | 物体検出装置及び物体検出方法 | |
CN111860084A (zh) | 图像特征的匹配、定位方法及装置、定位系统 | |
Fusco et al. | Self-localization at street intersections | |
JP6218514B2 (ja) | 歩行者検出装置及び交通制御システム | |
US20210027481A1 (en) | System, method, and computer-readable medium for managing position of target |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07739427 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12225583 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200780011198.1 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2007739427 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007739427 Country of ref document: EP |