EP3924932A1 - Modulares inpainting verfahren - Google Patents

Modulares inpainting verfahren

Info

Publication number
EP3924932A1
EP3924932A1 EP20707560.7A EP20707560A EP3924932A1 EP 3924932 A1 EP3924932 A1 EP 3924932A1 EP 20707560 A EP20707560 A EP 20707560A EP 3924932 A1 EP3924932 A1 EP 3924932A1
Authority
EP
European Patent Office
Prior art keywords
image
incomplete
areas
area
faulty
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP20707560.7A
Other languages
German (de)
English (en)
French (fr)
Inventor
Ekaterina Grünwedel
Charlotte GLOGER
Andreas Panakos
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Aumovio Autonomous Mobility Germany GmbH
Original Assignee
Conti Temic Microelectronic GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Conti Temic Microelectronic GmbH filed Critical Conti Temic Microelectronic GmbH
Publication of EP3924932A1 publication Critical patent/EP3924932A1/de
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/77Retouching; Inpainting; Scratch removal
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R1/00Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/20Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/22Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
    • B60R1/28Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with an adjustable field of view
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/60Image enhancement or restoration using machine learning, e.g. neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/586Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of parking space
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle
    • G06T2207/30264Parking

Definitions

  • the present invention relates to a method in which a scenery is recorded as at least one raw image by at least one optical detection means, which is attached in particular to a means of transport, and in which image data of the scenery in the subsequently rendered Reder image is incomplete in at least one area and / or with errors.
  • the invention also relates to an image processing system, a use of the same and a means of locomotion.
  • Means of transport, and among these in particular land vehicles, are provided with an increasing number of Assistenzsyste men.
  • These also include camera devices that are intended to give the occupants and, among them, preferably the respective driver, a visual impression of one or more scenes outside a passenger compartment, for example to support, facilitate or monitor a parking process.
  • an image processing system according to claim 12, its use according to claim 13 and a means of locomotion according to claim 14 represent solutions to the mentioned problem.
  • the solution therefore consists initially in a modular method, in which in a first step an identification of missing parts in the rendered image based on the visibility restrictions. In a next step, masks are created from the missing parts of the rendered image, and the rendered image is only to be reconstructed in these masks.
  • the correction image is optimized by post-processing to form an optimization image and this is displayed instead of the respective correction image.
  • Possible artifacts can be smoothed and the appearance can be made more pleasant to the eye after the image reconstruction, for which the rendered image can be post-processed, for example to increase the sharpness, reduce the contrast and / or harmonize colors.
  • the rendered images, reconstructed correction images or optimized optimization images are displayed to the viewer in real time or with a negligible delay to allow the user to react promptly to the scenes of the images shown to him.
  • Preferably carries the repetition frequency of the displayable images at least 5fps.
  • the visibility restrictions are determined at least using a three-dimensional model of the respective vehicle as well as using an accommodation of the optical detection means.
  • Different scenarios can also be described in a suitable manner using known visibility restrictions. Concretely related to a scenario with a motor vehicle, those parts of the rendered scene that do not have any image data can be identified using a 3D model of the motor vehicle and the camera housing.
  • data on visibility restrictions, geometric models of environments (and patterns) and (in particular) scenery data that have already been generated can be stored in at least one database in an advantageous variant of the method.
  • parts of the respective render image that are not to be reconstructed for the mask or masks can then be protected or hidden from further processing in a simple manner from the start.
  • the image data to be reconstructed can be generated with the aid of a machine learning approach.
  • the image data can be reconstructed with the aid of an artificial neural network, which accesses the at least one database and is trained on the basis of its data.
  • the incomplete and / or faulty image data can be reconstructed using edge-based methods. Thereby, edges or object transitions are searched for in the rendered images. Processing with algorithms often does not yet provide closed edges; these have to be joined using additional processes to enclose objects.
  • the edge-based method can preferably be a level set method, in particular a fast marching method.
  • the former represents a numerical method for approximating geometric objects and their movement, with curves and surfaces advantageously being able to be calculated on a spatially fixed coordinate system without having to use parameterizations of the objects in question.
  • a special method for solving boundary value problems numerically is the fast marching method, which solves boundary value problems of the eikonal equation, here the development of a closed surface as a function of time with a certain speed.
  • the edge-based method can use a diffusion approach that is used for dimension reduction or feature extraction and is later propagated to predict the information in areas of incomplete and / or faulty mapping.
  • a machine learning approach can be used in a suitable manner, for example a Markov Random Field (MRF) method, whereby the MRFs can be used to segment digital images or classified areas and assume an interaction or mutual influencing of elements of a field.
  • MRF Markov Random Field
  • an image processing system that executes the above method in one of its variants, whose use in a parking assistance system is a
  • a means of locomotion which, for example, carries out a parking process at low speed, and a means of locomotion itself, in particular a land vehicle, is provided with such a system for the task at hand.
  • Fig. La, lb schematic perspective views of a rear outer area on a motor vehicle, recorded with an optical detection means, shown as a render image (Fig. La) and as a correction image (Fig. Lb), which was generated according to the inventive method; and
  • FIG. 2a, 2b are schematic perspective views of an outside area on a motor vehicle, recorded with an optical detection means, shown as a render image (FIG. 2a) and as a correction image (FIG. 2b) that was generated according to the method according to the invention.
  • La and lb show schematic perspective views of a rear scenery in the outside area of a motor vehicle, which is detected with an optical detection means.
  • the rendered image of the scene shown in Fig. La one recognizes an essentially rectangular area which is not completely mapped, since there image data of the scene are missing due to the housing of the optical detection means designed as a camera, not shown.
  • the missing image data can be reconstructed by means of the method according to the invention by means of digital impainting.
  • the aim is to create a coherent image that is based on the entirety of the image itself and then gives the user a better visual experience when looking at it.
  • Said area of missing image data 10 can be recognized by an edge 20 which separates this area from that area 30 known image data, which represents the introductory, identifying step.
  • the area 30 of known image data is provided with a mask, i.e. masks are generated which, with masked areas 30 that are not to be processed, enclose the area or areas 10 of incomplete and / or faulty imaging, so that although these areas 10, but not the areas 30 of known image data that are correctly captured and rendered, are reconstructed.
  • Contour lines of the rendered image that touch the edge 20 of the mask are continued along their imaginary extension into the non-masked area 10 of the image, as well as the structure of an area around the mask edge 20. Different sections are defined in the unmasked area by the contour lines, which are filled with the respective color of the edge assigned to them, after which the respective area is optionally textured.
  • FIGS. 1 a and 1 b it can be seen that the edges 40a, 40b, 40c are shown correctly in the correction image, while the continuation of the edge 40d shows a negligible discontinuity due to the darkness of the upper area.
  • FIG. 2a and 2b which in turn show a render image (FIG. 2a) and a correction image (FIG. 2b) of another, here a lateral scene on a motor vehicle, seen from above, show that those are missing in the render image Image data of the again rectangular area 10 in the correction image, on the one hand, the depicted structure 50, but also the shadow area 60 facing away from a light source, not shown, can be accurately reproduced by the reconstruction according to the invention.
  • the invention described above relates to a method for processing images, wherein a scene is recorded as at least one raw image by at least one optical detection means, which is attached in particular to egg nem means of transport, and image data of the scenery in the subsequently rendered render image at least incomplete and / or in at least one area be mapped with errors.
  • the process comprises the following steps:
  • the visual experience of a user of a system provided with the optical detection means is improved in a suitable manner, since a spot-free, continuous representation of the scenery in the correction image is made available.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Mechanical Engineering (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)
  • Closed-Circuit Television Systems (AREA)
EP20707560.7A 2019-02-11 2020-01-30 Modulares inpainting verfahren Pending EP3924932A1 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
DE102019201702.9A DE102019201702A1 (de) 2019-02-11 2019-02-11 Modulares inpainting Verfahren
PCT/DE2020/200008 WO2020164671A1 (de) 2019-02-11 2020-01-30 Modulares inpainting verfahren

Publications (1)

Publication Number Publication Date
EP3924932A1 true EP3924932A1 (de) 2021-12-22

Family

ID=69723762

Family Applications (1)

Application Number Title Priority Date Filing Date
EP20707560.7A Pending EP3924932A1 (de) 2019-02-11 2020-01-30 Modulares inpainting verfahren

Country Status (6)

Country Link
US (1) US11961215B2 (enExample)
EP (1) EP3924932A1 (enExample)
JP (2) JP2022517849A (enExample)
CN (1) CN113330480A (enExample)
DE (1) DE102019201702A1 (enExample)
WO (1) WO2020164671A1 (enExample)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12165406B2 (en) * 2021-12-20 2024-12-10 Here Global B.V. Method, apparatus, and computer program product for identifying and correcting intersection lane geometry in map data
US12013255B2 (en) 2021-12-20 2024-06-18 Here Global B.V. Method, apparatus, and computer program product for correcting lane geometry in map data
CN115272639B (zh) * 2022-09-19 2022-12-23 武汉天际航信息科技股份有限公司 修复图像中车辆区域的方法、装置和计算机程序产品
CN118351288B (zh) * 2024-04-16 2024-11-22 北京积加科技有限公司 图像区域标记修正方法、装置、设备和计算机可读介质

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005301517A (ja) 2004-04-08 2005-10-27 Toyota Motor Corp 画像処理装置および画像処理方法
JP2006332785A (ja) * 2005-05-23 2006-12-07 Univ Of Tokyo 画像補完装置及び画像補完方法並びにプログラム
US8971612B2 (en) * 2011-12-15 2015-03-03 Microsoft Corporation Learning image processing tasks from scene reconstructions
CN103310415A (zh) 2013-03-15 2013-09-18 清华大学 基于人脸的图像缺损补绘方法及系统
US9886636B2 (en) 2013-05-23 2018-02-06 GM Global Technology Operations LLC Enhanced top-down view generation in a front curb viewing system
US10380239B2 (en) 2013-12-03 2019-08-13 Sharethrough Inc. Dynamic native advertisment insertion
US9542609B2 (en) * 2014-02-04 2017-01-10 Xerox Corporation Automatic training of a parked vehicle detector for large deployment
US11017311B2 (en) * 2014-06-30 2021-05-25 Hewlett Packard Enterprise Development Lp Dataset augmentation based on occlusion and inpainting
EP3224767A1 (en) 2014-11-26 2017-10-04 Curious Al OY Neural network structure and a method thereto
CN104537663B (zh) 2014-12-26 2018-01-02 广东中科遥感技术有限公司 一种图像抖动的快速校正方法
WO2016179303A1 (en) * 2015-05-04 2016-11-10 Kamama, Inc. System and method of vehicle sensor management
US9902322B2 (en) * 2015-10-30 2018-02-27 Bendix Commercial Vehicle Systems Llc Filling in surround view areas blocked by mirrors or other vehicle parts
CN106897655A (zh) * 2015-12-18 2017-06-27 富士通株式会社 停车位的检测装置、方法以及图像处理设备
DE102016220651A1 (de) 2016-10-20 2018-04-26 Conti Temic Microelectronic Gmbh Verfahren und Vorrichtung zur Erzeugung einer Fahrzeugumgebungsansicht bei einem Fahrzeug
US10290085B2 (en) * 2016-12-14 2019-05-14 Adobe Inc. Image hole filling that accounts for global structure and local texture
JP2018107573A (ja) * 2016-12-26 2018-07-05 株式会社東海理化電機製作所 車両用視認装置
US10140690B2 (en) 2017-02-03 2018-11-27 Harman International Industries, Incorporated System and method for image presentation by a vehicle driver assist module
JP6837880B2 (ja) * 2017-03-15 2021-03-03 株式会社東芝 画像処理装置、画像処理システム、画像処理方法、およびプログラム
CN108171663B (zh) * 2017-12-22 2021-05-25 哈尔滨工业大学 基于特征图最近邻替换的卷积神经网络的图像填充系统
CN109242791B (zh) 2018-08-22 2022-07-26 东北农业大学 一种针对破损植物叶片的批量修复方法

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
BESCOS BERTA ET AL: "Empty Cities: Image Inpainting for a Dynamic-Object-Invariant Space", ARXIV.ORG, 20 September 2018 (2018-09-20), XP093276240, Retrieved from the Internet <URL:https://arxiv.org/abs/1809.10239v1> [retrieved on 20250512], DOI: 10.48550/arXiv.1809.10239v1 *

Also Published As

Publication number Publication date
JP2024041895A (ja) 2024-03-27
JP2022517849A (ja) 2022-03-10
US20220156894A1 (en) 2022-05-19
DE102019201702A1 (de) 2020-08-13
US11961215B2 (en) 2024-04-16
WO2020164671A1 (de) 2020-08-20
CN113330480A (zh) 2021-08-31

Similar Documents

Publication Publication Date Title
EP3924932A1 (de) Modulares inpainting verfahren
WO2022128014A1 (de) Korrektur von bildern eines rundumsichtkamerasystems bei regen, lichteinfall und verschmutzung
DE102019131971A1 (de) Ein Bildverarbeitungsmodul
WO2019170202A1 (de) Surroundview-system mit angepasster projektionsfläche
DE102015209391A1 (de) Verfahren und Vorrichtung zum Erzeugen einer Maskierungsvorschrift sowie zum Maskieren einer Bildinformation einer Kamera
EP3292535B1 (de) Verfahren zum erzeugen eines gesamtbildes einer fahrzeugumgebung eines fahrzeuges und entsprechende vorrichtung
DE112018005391T5 (de) Anzeigeverarbeitungsvorrichtung
DE102018100909A1 (de) Verfahren zum Rekonstruieren von Bildern einer Szene, die durch ein multifokales Kamerasystem aufgenommen werden
EP3571091B1 (de) Verfahren und vorrichtung zum darstellen einer umgebung eines fahrzeuges
WO2014009406A1 (de) VERFAHREN UND VORRICHTUNG ZUM BERECHNEN EINER VERÄNDERUNG EINES ABBILDUNGSMAßSTABS EINES OBJEKTS
DE102011082881A1 (de) Darstellung der Umgebung eines Kraftfahrzeugs in einer bestimmten Ansicht unter Verwendung räumlicher Information
DE102018130229B4 (de) Verfahren und Vorrichtung zur Objektextraktion aus eine dreidimensionale Szene darstellenden Szenenbilddaten
DE102017206175A1 (de) Surround-View-System für ein Fahrzeug
DE102022214330A1 (de) Verfahren zur Erzeugung mindestens einer Ground Truth aus der Vogelperspektive
DE102020122908A1 (de) Verfahren zum Anzeigen einer Umgebung eines Fahrzeuges auf einer Anzeigeeinrichtung, Verarbeitungseinheit und Fahrzeug
DE102018008791A1 (de) Verfahren zu einer dreidimensionalen Rekonstruktion einer Szene
DE102021214952A1 (de) Verfahren zur Anzeige einer virtuellen Ansicht einer Umgebung eines Fahrzeugs, Computerprogramm, Steuergerät und Fahrzeug
DE102019008081A1 (de) Verfahren zu einer kamerabasierten dreidimensionalen Rekonstruktion
DE102023201500A1 (de) Verfahren zum Erzeugen von Trainingsdaten zum Trainieren eines Algorithmus des maschinellen Lernens
DE102022206782A1 (de) Verfahren zum Erzeugen einer Ansicht mit einem Kamerasystem sowie Kamerasystem
DE102021212970A1 (de) Verfahren und Vorrichtung zum Erzeugen einer Surround-Ansicht, und Kraftfahrzeug
DE102024109099A1 (de) Verfahren zum Rendern einer Rückprojektion, Renderingsystem, Computerprogrammprodukt und computerlesbares Speichermedium
DE102024203351A1 (de) Verfahren und Vorrichtung zur Verarbeitung von Bilddaten
DE102024109113A1 (de) Verfahren zum Rendern einer Rückprojektion, Renderingsystem, Computerprogrammprodukt und computerlesbares Speichermedium
DE102022105267A1 (de) Kombination von Kamerainformationen, die durch wenigstens ein Kamerabild gegeben sind, und weiteren Informationen, die durch Umfeldsensordaten gegeben sind

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: UNKNOWN

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20210913

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: CONTINENTAL AUTONOMOUS MOBILITY GERMANY GMBH

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20231211

RAP3 Party data changed (applicant data changed or rights of an application transferred)

Owner name: AUMOVIO AUTONOMOUS MOBILITY GERMANY GMBH