WO2020043362A1 - Système et procédé d'assistance d'un usinage de surfaces - Google Patents
Système et procédé d'assistance d'un usinage de surfaces Download PDFInfo
- Publication number
- WO2020043362A1 WO2020043362A1 PCT/EP2019/067615 EP2019067615W WO2020043362A1 WO 2020043362 A1 WO2020043362 A1 WO 2020043362A1 EP 2019067615 W EP2019067615 W EP 2019067615W WO 2020043362 A1 WO2020043362 A1 WO 2020043362A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- glasses
- workpiece
- defect
- scanner
- graphics processor
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 22
- 238000004381 surface treatment Methods 0.000 title abstract description 3
- 230000007547 defect Effects 0.000 claims abstract description 81
- 239000011521 glass Substances 0.000 claims abstract description 74
- 239000003973 paint Substances 0.000 claims description 32
- 238000001514 detection method Methods 0.000 claims description 9
- 238000012545 processing Methods 0.000 claims description 9
- 238000011156 evaluation Methods 0.000 claims description 4
- 238000010422 painting Methods 0.000 description 6
- 238000013459 approach Methods 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000012937 correction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 239000004922 lacquer Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62D—MOTOR VEHICLES; TRAILERS
- B62D65/00—Designing, manufacturing, e.g. assembling, facilitating disassembly, or structurally modifying motor vehicles or trailers, not otherwise provided for
- B62D65/005—Inspection and final control devices
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8803—Visual inspection
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B19/00—Programme-control systems
- G05B19/02—Programme-control systems electric
- G05B19/418—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM]
- G05B19/41875—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM] characterised by quality surveillance of production
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B62—LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
- B62D—MOTOR VEHICLES; TRAILERS
- B62D65/00—Designing, manufacturing, e.g. assembling, facilitating disassembly, or structurally modifying motor vehicles or trailers, not otherwise provided for
- B62D65/02—Joining sub-units or components to, or positioning sub-units or components with respect to, body shell or other sub-units or components
- B62D65/18—Transportation, conveyor or haulage systems specially adapted for motor vehicle or trailer assembly lines
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
- G01N2021/8854—Grading and classifying of flaws
- G01N2021/888—Marking defects
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
- G01N2021/8887—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges based on image processing techniques
- G01N2021/8893—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges based on image processing techniques providing a video image and a processed signal for helping visual decision
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/32—Operator till task planning
- G05B2219/32014—Augmented reality assists operator in maintenance, repair, programming, assembly, use of head mounted display with 2-D 3-D display and voice feedback, voice and gesture command
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/32—Operator till task planning
- G05B2219/32217—Finish defect surfaces on workpiece
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/02—Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]
Definitions
- the present invention relates to a system and a method for supporting a surface treatment, in particular the repair of
- WO 01/96829 A1 describes a method for testing a product
- An object of the present invention is to provide a system and a method for supporting surface processing on a workpiece, which enable a quick and reliable detection and correction of errors.
- the task is solved on the one hand by a system for supporting surface processing on a workpiece, which
- a graphics processor which is set up to display, in an image displayed in the VR glasses, an indication of a defect on or adjacent to a line of sight extending between the eye of a user of the VR glasses and the defect.
- the detection of errors is no longer restricted to the direction of view of the wearer of the VR glasses, and there is in particular no need for the wearer of the VR glasses to circle the workpiece in order to enable error detection from different directions of view.
- the scanner can have several cameras, which are aligned from different directions to a detection region in which the workpiece can be placed. Furthermore, drive means can be provided in order to move the workpiece and the cameras relative to one another. In particular, a variant in which the workpiece is moved past stationary cameras of the scanner is particularly suitable for integration in an assembly line production.
- the cameras can be mounted on a portal, under which the workpiece is transported, and thus capture it from various directions that may be difficult for an employee to access.
- the graphics processor can be set up to display an indication of a defect even if the defect is on a side of the workpiece facing away from the VR glasses. This ensures that a user of the VR glasses, when he approaches the workpiece, knows immediately whether there are defects to be repaired and does not have to go around the workpiece in a time-consuming manner.
- the graphics processor should furthermore be set up to use different forms of information, depending on whether the defect is located on a side of the workpiece facing the VR glasses or on a side facing away from the VR glasses.
- a defect is remedied, a continued indication of the defect can irritate the user of the VR glasses.
- means are preferably also provided for detecting a gesture by the user of the VR glasses in order to
- the gesture to be detected can be a pointing at the defect, a look at a predetermined object in the vicinity of the workpiece or a processing of the defect by the user of the VR glasses.
- an image evaluation unit can be connected to the camera to assess whether a defect is fixed or not, and the indication of the defect only to be deleted if the defect is assessed as having been remedied.
- the graphics processor is also preferably set up to outline the
- Play workpiece in the image displayed in the VR glasses for example, to make it easier for the user of the VR glasses among several workpieces in his environment to identify the workpiece to be machined.
- This outline should coincide with the outline of the workpiece visible from the user's perspective.
- the indication of a defect can be displayed exactly on the line of sight
- Such an indication can include, for example, a circle extending around the line of sight and the defect or an arrow pointing at it.
- the graphics processor can also be set up to insert details of the workpiece in alphanumeric form into the image displayed in the VR glasses.
- This information can be technical information about the workpiece, such as a type designation or a workpiece-specific identifier, information about the defect or also general information on handling the workpiece and in particular how to remedy the defect.
- the object is further achieved by a method for supporting an
- Graphic processor for receiving image data is connected, relative to the workpiece,
- the defects to which the method is applied are paint defects and the workpiece is a vehicle body.
- Fig. 1 is a schematic representation of the system according to the invention.
- Fig. 2 shows an example of the image that a user of the VR glasses in
- Figure 1 presents the system shown; and.
- Fig. 3 shows an alternative, displayed in the VR glasses.
- Fig. 1 shows schematically a system according to the invention, which is used in motor vehicle production for the detection of paint defects on vehicle bodies 1, which are conveyed on a conveyor belt or by other suitable drive means 2 from a painting system, not shown, known per se.
- Vehicle bodies 1 are transported by the drive means 2 through a scanner 3, which comprises a portal 5 equipped with various cameras 4.
- the cameras 4 are aligned with an interior of the portal 5 in order to detect the outer surfaces of the bodies 1 transported through.
- An image evaluation unit 6 connected to the cameras 4 analyzes the images supplied by the cameras and generates a data record for each body 1 that passes through the portal 3, which contains an individual identifier of the body 1 and, if applicable, locations on the body 1 at which a Painting error has been recognized, and may contain information about the nature of the painting error.
- the data records thus obtained are stored in a database 7.
- a process computer 8 accesses the data records in the database 7. By further communicating with a controller of the drive means 2 and recording or specifying their speed, the process computer 8 is able at any time to calculate at which location along the assembly line a particular body 1 is located.
- the process computer 8 is also coupled to a plurality of cameras 9, around which
- One or more of these cameras 9 are fixed in the vicinity of the assembly line; a further camera 9 can be built into the VR glasses 10 and be movable therewith in order to monitor the field of view of the employee 11 wearing the VR glasses 10.
- the process computer 8 is able, for example by triangulation, to determine the location of the Employee 11 (and, albeit with limited accuracy, his direction of view).
- a comparison of the image supplied by the camera of the VR glasses 10 with those of the fixed cameras 9 enables the determination of the viewing direction of the employee 11 or the orientation of the VR glasses 10 worn by him with high accuracy.
- Fig. 2 shows schematically the image that the employee 1 1 presents through the VR glasses 10.
- the VR glasses 10 comprise one in a manner known per se
- Display screen the transparency and color of which can be controlled locally, so that the user can perceive the real environment, here a sequence of the bodies 1 on the assembly line, through the transparent parts of the screen of the VR glasses 10, and at the same time graphic displays displayed on the screen Can perceive elements.
- a graphics processor 12 of the process computer 8 is capable of calculating the outline of each of the bodies 1, as presented from the perspective of the employee 11, and one of them
- the process computer 8 uses the information in the database 7 to select the one that is most advanced on the assembly line, by one
- the frame 13 may be a line extending around the body 1 in question; in the case shown here, the selected body 1 is additionally emphasized in relation to the others in that the transparency of the screen of the VR glasses 10 outside the frame 13 is reduced, so that the bodies not selected are darker and / or with a reduced contrast,
- the graphics processor 12 calculates where, from the point of view of the employee 1, 1 paint defects on the body 1 must be visible, and places an indication 14 in VR glasses 10 of each such defect on one Line of sight 15, which runs from the eye of the employee 11 to this paint defect. 2, these notices 14 have the form of crosses, the point of intersection of which corresponds to the position of the paint defect on the body 1, ie lies on the line of sight that the eye of the user 11 through the screen of the VR glasses 10 connects with the paint defect 16. Note 14 thus hinders the view of the paint defect, but this is irrelevant as long as
- the notice 14 on the screen of the VR glasses 10 can maintain its size or grow less than inversely proportional to the distance, so that with increasing
- the information 14 can cause perspective distortion
- the graphics processor 12 replaces the note 14 with a note 14 “of another shape, which is not immediately in the (to the plane 3 vertical line of sight 15 between the eye of the employee 11 and the one here Lacquer flaw 16, shown schematically as a cross-hatched area, lies in order not to obstruct the view of the latter during the repair.
- Process computer 8 capable of processing each one of the employees 1 1
- the process computer s can also recognize when the employee 11 turns away from the paint defect after successful repair, and then delete the paint defect in question from the database 7 or note it there as remedied.
- the paint defect is not deleted or noted as corrected before the employee 11 has finished repairing it, e.g. Because the employee 1 1 temporarily turns his head in a different direction, it can be provided that the paint defect is only deleted or noted as having been corrected if the
- a predetermined object 17 appears in the field of view of the camera of the VR glasses 10.
- the object 17 can in principle be selected as desired; For an intuitive use of the VR glasses 10, it is expedient if the object 17 has a meaning that is linked to the successful repair, e.g. one with OK
- labeled sign which is mounted adjacent to the assembly line at a position where there is a low probability that the employee 1 1 looks at it involuntarily, e.g. clearly above eye level, or a correspondingly labeled card that the employee 1 1 carries along with repair tools and holds in front of the eye or in front of the camera of the VR glasses 10 when he is using the
- the process computer 8 monitors the progress of the repair on the basis of the images supplied by the camera of the VR glasses 10, decides whether the paint defect has been remedied or not and, if necessary, adjusts the database 7 accordingly.
- Information 18 on the type and identity of the body can be displayed adjacent to this by the graphics processor 12 in the image displayed in the VR glasses 10, preferably outside the frame 13, so as not to obscure the body 1 to be processed
- At least the information on the identity of the body 1 can also be read on the body itself, so that by comparing the information from
- the graphic processor 12 displayed and the readable information of the employees 11 can verify that the indications 14 displayed in the VR glasses 10 regarding paint defects actually relate to the body 1 which is currently in front of them.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Manufacturing & Machinery (AREA)
- Chemical & Material Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Immunology (AREA)
- Automation & Control Theory (AREA)
- Quality & Reliability (AREA)
- Analytical Chemistry (AREA)
- Biochemistry (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Pathology (AREA)
- Combustion & Propulsion (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
L'invention concerne un système d'assistance d'un usinage de surfaces sur une pièce (1). Le système comprend : un scanner (3) servant à identifier et à localiser des défauts sur la surface d'une pièce (1) ; un instrument d'édition mobile par rapport au scanner (3) sous la forme d'un casque de réalité virtuelle (11) servant à éditer les positions, sur lesquelles le scanner (3) identifie un défaut ; des moyens (6, 9) servant à détecter la position et l'orientation du casque de réalité virtuelle (10) par rapport à la pièce (1) ; et un processeur graphique (12), qui est mis au point pour afficher, sur une image s'affichant dans le casque de réalité virtuelle (10), une information concernant le défaut sur ou à proximité d'une ligne de vision (15) s'étendant entre l'œil d'un utilisateur (11) du casque de réalité virtuelle (10) et le défaut.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP19736646.1A EP3844580A1 (fr) | 2018-08-29 | 2019-07-01 | Système et procédé d'assistance d'un usinage de surfaces |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102018006834.0A DE102018006834A1 (de) | 2018-08-29 | 2018-08-29 | System und Verfahren zur Unterstützung einer Oberflächenbearbeitung |
DE102018006834.0 | 2018-08-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020043362A1 true WO2020043362A1 (fr) | 2020-03-05 |
Family
ID=67184998
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2019/067615 WO2020043362A1 (fr) | 2018-08-29 | 2019-07-01 | Système et procédé d'assistance d'un usinage de surfaces |
Country Status (4)
Country | Link |
---|---|
EP (1) | EP3844580A1 (fr) |
DE (1) | DE102018006834A1 (fr) |
MA (1) | MA53471A (fr) |
WO (1) | WO2020043362A1 (fr) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115589534A (zh) * | 2022-09-09 | 2023-01-10 | 广州市斯睿特智能科技有限公司 | 跟随式车辆检测项图片采集装置及方法 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2001096829A1 (fr) | 2000-06-13 | 2001-12-20 | Volkswagen Aktiengesellschaft | Procede de controle qualitatif d'un produit et systeme de mise en oeuvre dudit procede |
US20090141964A1 (en) * | 2007-11-29 | 2009-06-04 | Takashi Magara | Appearance inspection apparatus, appearance inspection system, and appearance inspection appearance |
US20180154962A1 (en) * | 2016-12-06 | 2018-06-07 | Hyundai Motor Company | Vehicle body flaw elimination method using virtual mark and system thereof |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE19730885A1 (de) * | 1997-07-18 | 1999-01-21 | Audi Ag | Verfahren zur automatischen Erkennung von Oberflächenfehlern an Rohkarosserien und Vorrichtung zur Durchführung des Verfahrens |
-
2018
- 2018-08-29 DE DE102018006834.0A patent/DE102018006834A1/de not_active Ceased
-
2019
- 2019-07-01 WO PCT/EP2019/067615 patent/WO2020043362A1/fr unknown
- 2019-07-01 EP EP19736646.1A patent/EP3844580A1/fr not_active Withdrawn
- 2019-07-01 MA MA053471A patent/MA53471A/fr unknown
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2001096829A1 (fr) | 2000-06-13 | 2001-12-20 | Volkswagen Aktiengesellschaft | Procede de controle qualitatif d'un produit et systeme de mise en oeuvre dudit procede |
US20090141964A1 (en) * | 2007-11-29 | 2009-06-04 | Takashi Magara | Appearance inspection apparatus, appearance inspection system, and appearance inspection appearance |
US20180154962A1 (en) * | 2016-12-06 | 2018-06-07 | Hyundai Motor Company | Vehicle body flaw elimination method using virtual mark and system thereof |
Non-Patent Citations (1)
Title |
---|
AZREEN ZURIATHY ET AL: "Applications of Augmented Reality For Inspection and Maintenance Process in Automotive Industry", 24 February 2018 (2018-02-24), pages i,ii,413 - 421, XP055631167, Retrieved from the Internet <URL:https://www.researchgate.net/publication/325310004_Applications_of_Augmented_Reality_For_Inspection_and_Maintenance_Process_in_Automotive_Industry> [retrieved on 20191011], DOI: 10.4314/jfas.v10i3s.35 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115589534A (zh) * | 2022-09-09 | 2023-01-10 | 广州市斯睿特智能科技有限公司 | 跟随式车辆检测项图片采集装置及方法 |
CN115589534B (zh) * | 2022-09-09 | 2023-09-08 | 广州市斯睿特智能科技有限公司 | 跟随式车辆检测项图片采集装置及方法 |
Also Published As
Publication number | Publication date |
---|---|
EP3844580A1 (fr) | 2021-07-07 |
DE102018006834A1 (de) | 2020-03-05 |
MA53471A (fr) | 2021-12-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3507033B1 (fr) | Machine à cintrer munie d'un dispositif d'enregistrement d'images de la zone de travail et procédé pour améliorer la sécurité d'utilisation d'une machine à cintrer | |
DE102017128543B4 (de) | Störbereich-einstellvorrichtung für einen mobilen roboter | |
EP3507035B1 (fr) | Machine à cintrer dotée d'un dispositif de détection d'image de zone de travail et procédé de représentation d'une zone de travail | |
DE102014213518A1 (de) | Verfahren, Bearbeitungsmaschine und Computerprogrammprodukt zum bildbasierten Platzieren von Werkstückbearbeitungsvorgängen | |
DE102016107397B4 (de) | Bahnanzeigevorrichtung zur darstellung der bahn einer werkzeugachse | |
EP3316181B1 (fr) | Procédé de détection d'objets dans un entrepôt et chariot élévateur doté d'un dispositif de détection d'objets dans un entrepôt | |
EP2166510A1 (fr) | Procédé de détermination de la position et de l'orientation d'une caméra installée dans un véhicule | |
DE102017223160A1 (de) | Verfahren zum Erkennen von zumindest einem an einem Kraftfahrzeug anliegenden Objekt sowie Steuervorrichtung und Kraftfahrzeug | |
DE102018006684B4 (de) | Werkzeugmaschine und Verfahren zum Steuern einer axialen Bewegung | |
EP3867027A1 (fr) | Procédé de visualisation d'informations de processus lors de la production de composants en tôle | |
WO2020043362A1 (fr) | Système et procédé d'assistance d'un usinage de surfaces | |
EP2581731B1 (fr) | Dispositif et procédé de détection d'anomalies sur des instruments | |
DE102017208102A1 (de) | Verfahren zur vollautomatischen Korrektur von Beschichtungsfehlern | |
EP3659113B1 (fr) | Système de reconnaissance, procédé de travail et procédé d'apprentissage pour générer un modèle 3d avec des données de référence | |
DE102014214711B4 (de) | Verfahren zum Betrieb eines Assistenzsystems eines Kraftfahrzeugs sowie Assistenzsystem | |
EP3507034B1 (fr) | Machine à cintrer munie d'un dispositif d'affichage de données de cintrage | |
EP4121247B1 (fr) | Procédé et dispositif pour déterminer un état réel de barres de support d'un support de pièce, et machine-outil équipée d'un tel dispositif | |
DE102021117714A1 (de) | Automatische Nahterkennung für einen Schweißprozess | |
DE102017222818B4 (de) | Verfahren zur automatischen Detektion von Fehlstellen und Abarbeitung von diesen Fehlstellen in einer Werkstückoberfläche | |
EP3644280A1 (fr) | Procédé d'orientation d'une caméra en fonction de ses paramètres extrinsèques | |
EP3986662B1 (fr) | Procédé de contrôle et de post-traitement d'une pièce présentant un contour intérieur fermé découpé par laser | |
EP3467544B1 (fr) | Procédé pour un poste de travail de montage, poste de travail de montage, dispositif de mise en uvre du procédé, programme informatique et support lisible par ordinateur | |
DE102019209934B4 (de) | Elektronisches Gerät und Verfahren zur Visualisierung von Oberflächendefekten eines Werkstücks | |
EP2762260A1 (fr) | Dispositif et procédé de surveillance par caméra du soudage et la fermeture de fissures | |
EP2497065B1 (fr) | Procédé pour déterminer l'orientation de la partie supérieure d'une pile de charges isolées |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19736646 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2019736646 Country of ref document: EP Effective date: 20210329 |