WO2005031647A1 - Verfahren und vorrichtung zur berührungslosen optischen 3d-lagebestimmung eines objekts - Google Patents
Verfahren und vorrichtung zur berührungslosen optischen 3d-lagebestimmung eines objekts Download PDFInfo
- Publication number
- WO2005031647A1 WO2005031647A1 PCT/DE2004/002132 DE2004002132W WO2005031647A1 WO 2005031647 A1 WO2005031647 A1 WO 2005031647A1 DE 2004002132 W DE2004002132 W DE 2004002132W WO 2005031647 A1 WO2005031647 A1 WO 2005031647A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- geometric features
- image
- determination
- coordinate system
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
Definitions
- the present invention relates to a method and a device for non-contact optical 3D image determination of an object, the SD position determination comprising the determination of the 3D position and the 3D orientation of the object, and wherein geometric features on the object are known.
- the 3D position determination can be used to calibrate a robot.
- the geometric features are arranged on the robot arm in a known manner, or a calibration target is mounted on the robot arm.
- the robot arm is moved into the field of view of a fixed camera, and then the 3D position determination of the robot arm is carried out in order to check or re-calibrate its position and orientation.
- Another application is the 3D position determination of components in the installed state. By determining the position of a reference part, it can be checked, for example, whether the position of built-in parts matches the position of the reference part, in order to identify components that are badly assembled.
- the known devices for 3D position determination are nowadays exclusively multi-camera systems which cause considerable costs due to the use of two or more cameras.
- 2D methods for gripping components are known to date, ie position and orientation in one plane (3 degrees of freedom) and additionally height information determined by the size of the component (1 additional degree of freedom).
- this known method however, a maximum of 2 1 / 2D information about the position of the component is obtained, so that, for example, correct gripping is not possible if the component is skewed.
- the calibration of a robot is usually carried out manually, since corresponding aids for setting up all 6 degrees of freedom are missing.
- the calibration is therefore extremely time-consuming, which results in long maintenance intervals.
- manual calibration procedures are generally relatively inaccurate.
- the determination of the position of a component after its installation can also only be carried out today with multi-camera systems.
- sensory solutions for individual measurement variables are known which solve the actual overall task via a multi-sensor evaluation.
- Such multiple sensory measurements are also time-consuming and often not very helpful for an overall assessment of the 3D position.
- the present invention is based on the object of designing and developing a method and a device for the contactless optical 3D position determination of an object in such a way that complete SD information about the object to be examined can be determined with simple means and with high measuring speed and measuring accuracy is.
- the complete 3D position determination is carried out in a particularly advantageous manner by means of a single image recording by the camera.
- the real imaging process of the camera can be modeled using a mathematical model.
- properties of the optics used focal length, distortions, etc.
- perspective transformations and digitization and discretization of the light intensity when recording images in the camera and / or in one Calculator are taken into account.
- the known 3D position of the geometric features and the corresponding two-dimensional image information can be assigned on the basis of the model.
- a (usually non-linear) system of equations can be created, in which the parameters of the mapping process are included as unknown quantities.
- two equations are obtained from assigning a 3D point on the object to its 2D image position.
- the camera is held stationary.
- the orientation of the camera can also be predefined.
- points, straight lines, angles, circles, elliptical contours and / or conic sections can be used as geometric features that are provided at known locations on the object. It is only important in this context that the contours and shapes are generally mathematically describable and are known on the component side and can be observed and evaluated well in the camera image.
- the geometric objects can be used both for calibration and for determining the position.
- the geometric objects can be used, for example, to determine the internal camera parameters.
- the calibration can correct optical distortions and, in particular, the 3D positional relationship between the camera coordinate system and the object coordinate system can be determined.
- a separate calibration target can also be used for calibration.
- the 3D position of the object with respect to the camera coordinate system can be determined.
- the 3D position of the object can be determined with respect to any other fixed coordinate system that can be defined as desired, for example with respect to a world coordinate system that can be determined by an additional setup step.
- the 3D position of the object can even be determined with respect to a dynamic coordinate system as part of motion measurements or comparison measurements with reference master parts.
- one or more additional cameras are provided with which the object is recorded, preferably from different recording angles. Several fixed single cameras or one or more moving cameras can be used for the additional recordings. With the additional recordings, the calculated SD positions can be evaluated using an additional quality criterion and corrections made if necessary. The result is a further improved accuracy of the 3D position determination.
- additional supporting sensors could be provided, with the aid of which errors in the 3D position determination can be compensated and / or corrected.
- these are, for example, temperature sensors for compensating for fluctuations in the ambient temperature, so that the temperature-dependent expansion of the object to be examined can be taken into account when calculating the 3D position.
- a trace and / or camber measurement can be carried out on motor vehicles based on geometric features on the rim. It can be provided that striking locations on the rim are automatically offered as geometric features. The striking points can be, for example, the valve or the hub cover. In a preferred embodiment, the prominent positions offered can be accepted interactively by the user or rejected. Knowing the geometric shape on the rim enables you to control the wheelbase in all its degrees of freedom. On the one hand, this eliminates the need for a projected pattern on the rim and, on the other hand, there is the possibility of determining a positive or negative camber with the shock absorber relaxed.
- a device for non-contact optical 3D position determination is characterized by a camera for creating an image of the object, the 3D position of the Object is predictable.
- the device according to the invention is preferably used to carry out a method according to one of claims 1 to 20, so that reference is made to the previous part of the description in order to avoid repetitions.
- the device advantageously comprises an industrial PC which is equipped with suitable image processing software.
- Fig. A schematic representation of an embodiment of an inventive device for non-contact optical 3D position determination of an object.
- the figure shows schematically a device for contactless optical SD position determination of an object 1, the 3D position of the object 1 being determined with respect to a spatially fixed coordinate system with the axes x, y and z.
- a picture of the object 1 is taken by a camera 2 positioned above the object 1.
- the position and orientation of the camera 2 are predefined in relation to the fixed coordinate system xyz.
- light sources 3 arranged around the camera 2 ensure sufficient illumination of the object 1.
- the camera data are transmitted to an industrial PC 4 with a monitor 5.
- There the 2D image information is assigned to the known 3D position of the geometric features present on the object 1 (not shown).
- An over-determined system of equations is created for the parameters of the mapping process, and non-linear optimization methods are used to solve them.
- the exact 3D position of the object 1 can be determined and the coordinates can be transmitted to a robot 6. Knowing the position and orientation of the object 1, the robot 6 can - after appropriate calibration - grasp or process the object 1.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Automatic Assembly (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006527269A JP2007533963A (ja) | 2003-09-26 | 2004-09-23 | 物体の3d位置の非接触式光学的測定方法及び測定装置 |
EP04786848A EP1665162A1 (de) | 2003-09-26 | 2004-09-23 | Verfahren und vorrichtung zur berührungslosen optischen 3d-l agebestimmung eines objekts |
US11/389,381 US8064686B2 (en) | 2003-09-26 | 2006-03-24 | Method and device for the contactless optical determination of the 3D position of an object |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE10345112.9 | 2003-09-26 | ||
DE10345112 | 2003-09-26 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/389,381 Continuation US8064686B2 (en) | 2003-09-26 | 2006-03-24 | Method and device for the contactless optical determination of the 3D position of an object |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005031647A1 true WO2005031647A1 (de) | 2005-04-07 |
Family
ID=34384326
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/DE2004/002132 WO2005031647A1 (de) | 2003-09-26 | 2004-09-23 | Verfahren und vorrichtung zur berührungslosen optischen 3d-lagebestimmung eines objekts |
Country Status (6)
Country | Link |
---|---|
US (1) | US8064686B2 (de) |
EP (1) | EP1665162A1 (de) |
JP (1) | JP2007533963A (de) |
CN (1) | CN1856804A (de) |
DE (1) | DE102004046584A1 (de) |
WO (1) | WO2005031647A1 (de) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7359817B2 (en) * | 2004-07-28 | 2008-04-15 | Fanuc Ltd | Method of and device for re-calibrating three-dimensional visual sensor in robot system |
CN109443321A (zh) * | 2018-10-30 | 2019-03-08 | 中国人民解放军国防科技大学 | 一种监测大型结构形变的串并联像机网络测量方法 |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4914039B2 (ja) * | 2005-07-27 | 2012-04-11 | キヤノン株式会社 | 情報処理方法および装置 |
JP5509645B2 (ja) * | 2009-03-25 | 2014-06-04 | 富士ゼロックス株式会社 | 位置・姿勢認識方法、部品把持方法、部品配置方法、部品組立方法、位置・姿勢認識装置、部品把持装置、部品配置装置、および部品組立装置 |
CN102022979A (zh) * | 2009-09-21 | 2011-04-20 | 鸿富锦精密工业(深圳)有限公司 | 三维光学感测系统 |
CN102834626B (zh) | 2010-04-28 | 2015-06-24 | 宝马股份公司 | 构件连接装置和用于连接构件的方法 |
DE102010041356A1 (de) | 2010-09-24 | 2012-03-29 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren zum Verbinden von Bauteilen |
DE102010042803B4 (de) | 2010-10-22 | 2020-09-24 | Bayerische Motoren Werke Aktiengesellschaft | Bauteilverbindung |
DE102011080483B4 (de) | 2011-08-05 | 2015-07-09 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren zur Herstellung eines Bauteils oder eines aus mehreren Bauteilen bestehenden Bauteilverbunds |
JP5494597B2 (ja) * | 2011-09-16 | 2014-05-14 | 株式会社安川電機 | ロボットシステム |
CN104175085B (zh) * | 2013-05-21 | 2016-11-09 | 重庆长安汽车股份有限公司 | 冲压生产顶杆孔快速定位装置及方法 |
US9747680B2 (en) | 2013-11-27 | 2017-08-29 | Industrial Technology Research Institute | Inspection apparatus, method, and computer program product for machine vision inspection |
JP2018040760A (ja) * | 2016-09-09 | 2018-03-15 | 株式会社鈴木エンタープライズ | 構造物寸法計測システム及び計測方法 |
WO2019039996A1 (en) * | 2017-08-25 | 2019-02-28 | Maker Trading Pte Ltd | ARTIFICIAL VISION SYSTEM AND METHOD FOR IDENTIFYING LOCATIONS OF TARGET ELEMENTS |
US10871384B2 (en) | 2019-02-26 | 2020-12-22 | Thomas P. Moyer | Apparatus and methods utilizing emissive patterns to determine positional information |
CN113432585A (zh) * | 2021-06-29 | 2021-09-24 | 沈阳工学院 | 一种基于机器视觉技术的非接触式轮毂位置精确测定方法 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4630225A (en) * | 1982-09-29 | 1986-12-16 | Omron Tateisi Electronic Co. | System for detecting orientation of articles |
US4942539A (en) * | 1988-12-21 | 1990-07-17 | Gmf Robotics Corporation | Method and system for automatically determining the position and orientation of an object in 3-D space |
US5388059A (en) * | 1992-12-30 | 1995-02-07 | University Of Maryland | Computer vision system for accurate monitoring of object pose |
EP1043689A2 (de) * | 1999-04-08 | 2000-10-11 | Fanuc Ltd | Bildverarbeitungsgerät |
US6300974B1 (en) * | 1997-02-28 | 2001-10-09 | Commissariat A L'energie Atomique | Process and device for locating an object in space |
WO2003030738A1 (en) * | 2001-10-09 | 2003-04-17 | Integra Medical Imaging (Aust) Pty Ltd | 2-d and 3-d pose estimation of articles from 2-d images |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS63133002A (ja) * | 1986-11-25 | 1988-06-04 | Nec Corp | 三次元座標測定装置 |
JPH076769B2 (ja) * | 1988-07-28 | 1995-01-30 | 工業技術院長 | 物体計測装置 |
DE4113992A1 (de) * | 1991-04-29 | 1992-11-05 | Ameling Walter | Verfahren zur automatischen dreidimensionalen ueberwachung von gefahrenraeumen |
US5380978A (en) * | 1991-07-12 | 1995-01-10 | Pryor; Timothy R. | Method and apparatus for assembly of car bodies and other 3-dimensional objects |
US6081273A (en) * | 1996-01-31 | 2000-06-27 | Michigan State University | Method and system for building three-dimensional object models |
US6886231B2 (en) * | 1999-06-25 | 2005-05-03 | Burke E. Porter Machinery Company | Robotic apparatus and method for mounting a valve stem on a wheel rim |
DE19934864A1 (de) * | 1999-07-24 | 2001-02-08 | Bosch Gmbh Robert | Vorrichtung zum Bestimmen der Rad- und/oder Achsgeometrie von Kraftfahrzeugen |
US6771808B1 (en) * | 2000-12-15 | 2004-08-03 | Cognex Corporation | System and method for registering patterns transformed in six degrees of freedom using machine vision |
JP4366023B2 (ja) * | 2001-03-16 | 2009-11-18 | インターナショナル・ビジネス・マシーンズ・コーポレーション | ビデオ・イメージの部分イメージ領域抽出方法、部分イメージ領域抽出システム、部分イメージ領域抽出のためのプログラム、抽出されたビデオ・イメージの配信方法およびコンテンツ作成方法 |
-
2004
- 2004-09-23 JP JP2006527269A patent/JP2007533963A/ja active Pending
- 2004-09-23 DE DE102004046584A patent/DE102004046584A1/de not_active Withdrawn
- 2004-09-23 EP EP04786848A patent/EP1665162A1/de not_active Ceased
- 2004-09-23 WO PCT/DE2004/002132 patent/WO2005031647A1/de active Search and Examination
- 2004-09-23 CN CNA2004800277822A patent/CN1856804A/zh active Pending
-
2006
- 2006-03-24 US US11/389,381 patent/US8064686B2/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4630225A (en) * | 1982-09-29 | 1986-12-16 | Omron Tateisi Electronic Co. | System for detecting orientation of articles |
US4942539A (en) * | 1988-12-21 | 1990-07-17 | Gmf Robotics Corporation | Method and system for automatically determining the position and orientation of an object in 3-D space |
US5388059A (en) * | 1992-12-30 | 1995-02-07 | University Of Maryland | Computer vision system for accurate monitoring of object pose |
US6300974B1 (en) * | 1997-02-28 | 2001-10-09 | Commissariat A L'energie Atomique | Process and device for locating an object in space |
EP1043689A2 (de) * | 1999-04-08 | 2000-10-11 | Fanuc Ltd | Bildverarbeitungsgerät |
WO2003030738A1 (en) * | 2001-10-09 | 2003-04-17 | Integra Medical Imaging (Aust) Pty Ltd | 2-d and 3-d pose estimation of articles from 2-d images |
Non-Patent Citations (6)
Title |
---|
DHOME M ET AL: "DETERMINATION OF THE ATTITUDE OF 3-D OBJECTS FROM A SINGLE PERSPECTIVE VIEW", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, IEEE INC. NEW YORK, US, vol. 11, no. 12, 1 December 1989 (1989-12-01), pages 1265 - 1278, XP000084832, ISSN: 0162-8828 * |
KITE D H ET AL: "DETERMINING THE 3D POSITION AND ORIENTATION OF A ROBOT CAMERA USING 2D MONOCULAR VISION", PATTERN RECOGNITION, ELSEVIER, KIDLINGTON, GB, vol. 23, no. 8, January 1990 (1990-01-01), pages 819 - 831, XP000107907, ISSN: 0031-3203 * |
SALVI J ET AL: "A comparative review of camera calibrating methods with accuracy evaluation", PATTERN RECOGNITION, ELSEVIER, KIDLINGTON, GB, vol. 35, no. 7, July 2002 (2002-07-01), pages 1617 - 1635, XP004345158, ISSN: 0031-3203 * |
See also references of EP1665162A1 * |
WLCZEK P ET AL: "Pose estimation of three-dimensional objects from single camera images", DIGITAL SIGNAL PROCESSING USA, vol. 5, no. 3, July 1995 (1995-07-01), pages 176 - 183, XP002318297, ISSN: 1051-2004 * |
XU WENLI ET AL INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS: "Pose estimation problem in computer vision", PROCEEDINGS OF THE REGION TEN CONFERENCE (TENCON). BEIJING, OCT. 19 - 21, 1993, BEIJING, IAP, CN, vol. VOL. 3, 19 October 1993 (1993-10-19), pages 1138 - 1141, XP010113719, ISBN: 0-7803-1233-3 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7359817B2 (en) * | 2004-07-28 | 2008-04-15 | Fanuc Ltd | Method of and device for re-calibrating three-dimensional visual sensor in robot system |
CN109443321A (zh) * | 2018-10-30 | 2019-03-08 | 中国人民解放军国防科技大学 | 一种监测大型结构形变的串并联像机网络测量方法 |
CN109443321B (zh) * | 2018-10-30 | 2021-01-08 | 中国人民解放军国防科技大学 | 一种监测大型结构形变的串并联像机网络测量方法 |
Also Published As
Publication number | Publication date |
---|---|
US8064686B2 (en) | 2011-11-22 |
EP1665162A1 (de) | 2006-06-07 |
JP2007533963A (ja) | 2007-11-22 |
CN1856804A (zh) | 2006-11-01 |
US20070009149A1 (en) | 2007-01-11 |
DE102004046584A1 (de) | 2005-05-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2515101B1 (de) | Verfahren zur Planung einer Inspektionsbahn und zur Bestimmung von zu inspizierenden Bereichen | |
DE102013021917B4 (de) | Robotersystemanzeigevorrichtung | |
DE102010042540B4 (de) | Verfahren und Vorrichtung zum Kalibrieren einer Abstandsbestimmungsvorrichtung eines optischen Systems | |
EP1376051B1 (de) | Kalibrierung eines Bildsensorsystems an einem Kraftfahrzeug mit Kalibrierobjekt und Lagebezugssensor | |
EP2603767B1 (de) | Verfahren zum kalibrieren eines messsystems und vorrichtung zum durchführen des verfahrens | |
EP1711777B2 (de) | Verfahren zur bestimmung der lage und der relativverschiebung eines objekts im raum | |
WO2005031647A1 (de) | Verfahren und vorrichtung zur berührungslosen optischen 3d-lagebestimmung eines objekts | |
EP2489977A2 (de) | Vorrichtung und Verfahren zur Bestimmung der 3-D-Koordinaten eines Objekts und zum Kalibrieren eines Industrieroboters | |
DE102017207696A1 (de) | Kalibrierungsvorrichtung und Kalibrierungsverfahren | |
DE102008041523A1 (de) | Verfahren zur dreidimensionalen Messung und Vorrichtung zur dreidimensionalen Messung | |
EP1497613A1 (de) | Verfahren und vorrichtung zur bestimmung der räumlichen koordinaten eines gegenstandes | |
EP1910999B1 (de) | Verfahren und anordnung zur bestimmung der relativen lage eines ersten objektes bezüglich eines zweiten objektes, sowie ein entsprechendes computerprogramm und ein entsprechendes computerlesbares speichermedium | |
DE102015109557B4 (de) | Verfahren zur Einstellung einer Prüf- oder Messposition eines berührungslos wirkenden Sensors | |
DE102019102927B4 (de) | Verfahren und Vorrichtung zum Bestimmen von dimensionalen und/oder geometrischen Eigenschaften eines Messobjekts | |
DE102021209178A1 (de) | Verfahren und Vorrichtung zum Bestimmen von Relativposen und zur Kalibrierung bei einem Koordinatenmessgerät oder Roboter | |
DE10016963C2 (de) | Verfahren zur Bestimmung der Position eines Werkstücks im 3D-Raum | |
EP3707569B1 (de) | Kalibrierung eines stationären kamerasystems zur positionserfassung eines mobilen roboters | |
EP1468792A2 (de) | Verfahren zum Kalibrieren eines Roboters | |
DE102006005990B4 (de) | Werkstückvermessung für 3-D Lageerkennung in mehreren Multi-Roboter-Stationen | |
DE102007058293A1 (de) | Kalibriervorrichtung und Verfahren zum Abgleichen eines Roboterkoordinatensystems | |
DE102022104880B4 (de) | Verfahren zur Kalibrierung eines portablen Referenzsensorsystems, portables Referenzsensorsystem und Verwendung des portablen Referenzsensorsystems | |
DE102007032471A1 (de) | Verfahren und Vorrichtung zur Bestimmung der Lage eines Kamerasystems relativ zu einem Objekt | |
DE102010042821B4 (de) | Verfahren und Vorrichtung zur Bestimmung einer Basisbreite eines Stereo-Erfassungssystems | |
DE102010004233B3 (de) | Verfahren zur Bestimmung der Lage eines Kamerasystems bezüglich eines Objekts | |
DE102016109919A1 (de) | Vorrichtung zur Vermessung von Objekten |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200480027782.2 Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
REEP | Request for entry into the european phase |
Ref document number: 2004786848 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2004786848 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2006527269 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11389381 Country of ref document: US |
|
DPEN | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101) | ||
WWP | Wipo information: published in national office |
Ref document number: 2004786848 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 11389381 Country of ref document: US |