EP4069440A1 - Verfahren und vorrichtung zur erkennung von umgefallenen und/oder beschädigten behältern in einem behältermassenstrom - Google Patents
Verfahren und vorrichtung zur erkennung von umgefallenen und/oder beschädigten behältern in einem behältermassenstromInfo
- Publication number
- EP4069440A1 EP4069440A1 EP20808055.6A EP20808055A EP4069440A1 EP 4069440 A1 EP4069440 A1 EP 4069440A1 EP 20808055 A EP20808055 A EP 20808055A EP 4069440 A1 EP4069440 A1 EP 4069440A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- containers
- container
- fallen
- images
- mass flow
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 238000013528 artificial neural network Methods 0.000 claims abstract description 56
- 238000012545 processing Methods 0.000 claims abstract description 39
- 238000012549 training Methods 0.000 claims description 38
- 238000001514 detection method Methods 0.000 claims description 6
- 230000007613 environmental effect Effects 0.000 claims description 6
- 238000003860 storage Methods 0.000 claims description 5
- 230000004807 localization Effects 0.000 claims description 3
- 230000032258 transport Effects 0.000 description 17
- 235000013361 beverage Nutrition 0.000 description 13
- 238000011156 evaluation Methods 0.000 description 11
- 241000196324 Embryophyta Species 0.000 description 4
- 238000005520 cutting process Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 235000013305 food Nutrition 0.000 description 2
- 230000001537 neural effect Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000011144 upstream manufacturing Methods 0.000 description 2
- 240000000111 Saccharum officinarum Species 0.000 description 1
- 235000007201 Saccharum officinarum Nutrition 0.000 description 1
- 241000209140 Triticum Species 0.000 description 1
- 235000021307 Triticum Nutrition 0.000 description 1
- 240000008042 Zea mays Species 0.000 description 1
- 235000005824 Zea mays ssp. parviglumis Nutrition 0.000 description 1
- 235000002017 Zea mays subsp mays Nutrition 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 235000005822 corn Nutrition 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005429 filling process Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000005693 optoelectronics Effects 0.000 description 1
- 238000012858 packaging process Methods 0.000 description 1
- 239000000825 pharmaceutical preparation Substances 0.000 description 1
- 229940127557 pharmaceutical product Drugs 0.000 description 1
- 239000002994 raw material Substances 0.000 description 1
- 238000009420 retrofitting Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000001931 thermography Methods 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B07—SEPARATING SOLIDS FROM SOLIDS; SORTING
- B07C—POSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
- B07C5/00—Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
- B07C5/34—Sorting according to other particular properties
- B07C5/3404—Sorting according to other particular properties according to properties of containers or receptacles, e.g. rigidity, leaks, fill-level
- B07C5/3408—Sorting according to other particular properties according to properties of containers or receptacles, e.g. rigidity, leaks, fill-level for bottles, jars or other glassware
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B07—SEPARATING SOLIDS FROM SOLIDS; SORTING
- B07C—POSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
- B07C2501/00—Sorting according to a characteristic or feature of the articles or material to be sorted
- B07C2501/0063—Using robots
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10144—Varying exposure
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
Definitions
- the deep neural network can be trained with a training data set with images of standing and fallen and / or damaged containers, so that the deep neural network uses the training data set to develop a model to match the standing, fallen and / or damaged containers of the container mass flow from one another distinguish.
- the deep neural network can be trained with a large number of different cases, so that the evaluation is largely independent of the container type and / or environmental influences.
- the training data set can include images of containers of different sizes, orientations or positions.
- the images of the training data set can be recorded with the at least one camera. It is conceivable that this takes place in a test system or directly on site at an operator of the beverage processing system. It is also conceivable that the manufacturer of the beverage processing system creates a database with images of standing and fallen and / or damaged containers in order to then use them with the training data set.
- the images of the training data set can be automatically duplicated in order to create additional images with additional combinations of standing and fallen and / or damaged containers. As a result, the effort involved in creating the training data set can be reduced considerably. It is conceivable that image sections with a standing or fallen and / or damaged container are created during the reproduction. The image sections can come from an original data set. It is conceivable that the image sections are rotated and / or enlarged individually during the reproduction. It is also conceivable that at least one exposure parameter is changed in the image sections during the reproduction. The image sections can then be reassembled to form the images of the training data set. As a result, a very large number of different images of the training data set can be provided using a few original images.
- the exposure parameter can mean a brightness and / or a contrast of an image section.
- the device offers the advantage of active accident and personal protection, since fallen and / or damaged containers do not have to be manually removed from the container mass flow by the operator. This is all the more true since the containers in the mass container flow are subjected to dynamic pressure among one another and intervention by the operating personnel to remove a container as a result of the sudden relief of the container flow harbors the risk of accidents such as crushing and cutting.
- the device for identifying the fallen and / or damaged container in the container mass flow can be arranged in a beverage processing system. It is conceivable that at least one container treatment machine is arranged upstream and / or downstream of the conveyor. In other words, the conveyor can connect two container handling machines to one another.
- FIG. 3 shows an exemplary embodiment of a method according to the invention for recognizing fallen containers as a flow chart
- FIG. 4 shows an exemplary embodiment of a section of the method from FIG. 3 for training the deep neural network
- the camera 6 is arranged on the conveyor 5, which detects the standing containers 2 and the containers 3 that have fallen over from above at an angle.
- the arrangement of the camera 6 is only shown here by way of example. It is also conceivable that there are several cameras that look obliquely from above in the same direction or in opposite directions. An arrangement is also conceivable directly from above, perpendicular to a transport surface of the conveyor 5.
- the neural network 71 is designed to recognize and localize the fallen containers 3. On the basis of the evaluation, the fallen containers 3 can then be removed from the conveyor 5 with a switch (not shown here) or by means of a gripping arm.
- FIG. 2 two exemplary images 11 and I2 of the image data stream output by camera 6 from FIG. 1 are shown.
- the training data set can be obtained from a set larger than 1000, preferably larger than 5000 and particularly preferably larger than 10000 images.
- step 120 the containers 2 of the container mass flow are transported standing on the conveyor 5. It can occasionally happen that one of the containers 2 falls over and then lies on the conveyor 5 as a fallen container 3.
- step 111 images of different container types and / or different lighting conditions are acquired. It is conceivable, for example, that this is done on a test system or that images of the container mass flow M of various beverage processing systems are collected in a database.
- step 112 the images are scaled to a standard size. This enables them to be evaluated uniformly.
- the fallen and standing containers 2, 3 are marked and classified. This can be done manually, semi-automatically or automatically. For example, this can be done manually by an operator on a screen or with a particularly computationally intensive image processing algorithm.
- the marking can be, for example, a surrounding box and the classification can be a container type or a container size.
- step 114 the images are automatically duplicated in order to create further images with additional combinations of standing and fallen containers 2, 3.
- image sections are first created with one standing or one overturned container 2, 3, which are then rotated and / or enlarged individually for reproduction. It is also conceivable that the exposure parameters of the image sections are changed during the reproduction. The image sections can then be put together in the most varied of combinations as further images, from which the training data set is then created in step 115.
- step 116 features are automatically extracted by means of the deep neural network 71.
- a multi-stage filtering process for the training data set is used. It is conceivable that edge filters or the like can be used to extract the outer boundary of each individual container 2, 3.
- the extraction of features here can very generally mean a method for recognizing and / or localizing distinguishing features of the overturned containers 3 compared to the standing containers 2 in the images of the training data set.
- this can also be done manually by an operator.
- the extracted features can include a container closure, a contour of a standing or fallen container 2, 3, a container label and / or the light reflections.
- the extracted features can each include a feature classification, a 2D and / or 3D coordinate.
- the deep neural network 71 is trained with the training data set.
- images of the training data set with the extracted features and the associated markings and classifications of the fallen and standing containers 2, 3 are iteratively given to the deep neural network 71. From this, the deep neural network 71 develops a model in step 118 with which the fallen and standing containers 2, 3 can be recognized.
- the model can then be verified by means of the training data set without specifying the markings and classifications. A comparison is made as to whether the deep neural network 71 actually recognizes the previously specified markings and classifications in the training data set. Likewise, further images with fallen and standing containers 2, 3 can be used for this purpose, which the deep neural network 71 was not trained.
- the substep 140 of the method 100 from FIG. 3 for evaluating the image data stream with the deep neural network 71 is shown in more detail as a flowchart.
- step 142 The features are then extracted in step 142. This takes place in the same way as described in step 116 with reference to FIG.
- the deep neural network then recognizes the orientation and the location of the respective container 2, 3 in step 143 and indicates a probability as to whether this container 2, 3 is transported lying or standing on the conveyor 5.
- This information is then visualized in step 144 and output on a screen in accordance with FIG. 2. In this way, an operator can check whether the recognition is proceeding properly.
- a signal is output in step 145 in order to remove it from the conveyor 5, for example with a switch or a gripper arm.
- the container mass flow M is recorded as an image data stream with the at least one camera 6 and the image data stream is evaluated by the image processing unit 7 with the deep neural network 71, the images of the image data stream can be evaluated on the basis of previously learned empirical values from the deep neural network 71, around the standing and to classify fallen container 2, 3 respectively. Because it is possible to train the deep neural network 71 with images of the most varied of container types and / or environmental conditions, it is no longer necessary to adapt the evaluation of the image data stream in the specific application. Consequently, the method according to the invention is particularly flexible and easy to use.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Quality & Reliability (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Automation & Control Theory (AREA)
- Image Analysis (AREA)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102019132830.6A DE102019132830A1 (de) | 2019-12-03 | 2019-12-03 | Verfahren und Vorrichtung zur Erkennung von umgefallenen und/oder beschädigten Behältern in einem Behältermassenstrom |
PCT/EP2020/082172 WO2021110392A1 (de) | 2019-12-03 | 2020-11-16 | Verfahren und vorrichtung zur erkennung von umgefallenen und/oder beschädigten behältern in einem behältermassenstrom |
Publications (1)
Publication Number | Publication Date |
---|---|
EP4069440A1 true EP4069440A1 (de) | 2022-10-12 |
Family
ID=73455690
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20808055.6A Pending EP4069440A1 (de) | 2019-12-03 | 2020-11-16 | Verfahren und vorrichtung zur erkennung von umgefallenen und/oder beschädigten behältern in einem behältermassenstrom |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230005127A1 (zh) |
EP (1) | EP4069440A1 (zh) |
CN (1) | CN114761145A (zh) |
DE (1) | DE102019132830A1 (zh) |
WO (1) | WO2021110392A1 (zh) |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE4200546A1 (de) * | 1992-01-11 | 1993-07-15 | Alfill Getraenketechnik | Verfahren und vorrichtung zum behandeln von flaschen |
DE20110686U1 (de) | 2001-06-27 | 2002-08-01 | Krones Ag | Vorrichtung zum Erkennen liegender Gefäße |
DE102007014802A1 (de) | 2007-03-28 | 2008-10-09 | Khs Ag | Verfahren zur Überwachung, Steuerung und Optimierung von Abfüllanlagen für Lebensmittel, insbesondere für Getränkeflaschen |
DE102009043976B4 (de) | 2009-09-10 | 2021-07-29 | Krones Aktiengesellschaft | Fördereinrichtung und Verfahren zu deren Steuerung |
DE102013207139A1 (de) * | 2013-04-19 | 2014-10-23 | Krones Ag | Verfahren zur Überwachung und Steuerung einer Abfüllanlage und Vorrichtung zur Durchführung des Verfahrens |
BR112016028533A2 (pt) | 2014-06-06 | 2017-08-22 | Gebo Cermex Canada Inc | dispositivo de intervenção para linha de transporte de produtos e processo de intervenção em uma linha de transporte de produtos |
CN106000904B (zh) * | 2016-05-26 | 2018-04-10 | 北京新长征天高智机科技有限公司 | 一种生活垃圾自动分拣系统 |
DE102016124400A1 (de) * | 2016-12-14 | 2018-06-14 | Krones Ag | Verfahren und Vorrichtung zum Erfassen von Störungen beim Objekttransport |
JP6595555B2 (ja) * | 2017-10-23 | 2019-10-23 | ファナック株式会社 | 仕分けシステム |
CN111712769B (zh) * | 2018-03-06 | 2023-08-01 | 欧姆龙株式会社 | 用于设定照明条件的方法、装置、系统以及存储介质 |
DE102018105301B4 (de) * | 2018-03-08 | 2021-03-18 | Sick Ag | Kamera und Verfahren zur Erfassung von Bilddaten |
JP7076747B2 (ja) * | 2018-05-09 | 2022-05-30 | リョーエイ株式会社 | 分類器の学習支援システム、学習データの収集方法、検査システム |
CN110154272B (zh) * | 2019-05-17 | 2021-04-13 | 佛山市玖州智能装备技术有限公司 | 人工智能废品塑料瓶分拣方法 |
CN110321944A (zh) * | 2019-06-26 | 2019-10-11 | 华中科技大学 | 一种基于接触网画质评估的深度神经网络模型的构建方法 |
-
2019
- 2019-12-03 DE DE102019132830.6A patent/DE102019132830A1/de active Pending
-
2020
- 2020-11-16 WO PCT/EP2020/082172 patent/WO2021110392A1/de unknown
- 2020-11-16 EP EP20808055.6A patent/EP4069440A1/de active Pending
- 2020-11-16 CN CN202080083581.3A patent/CN114761145A/zh active Pending
- 2020-11-16 US US17/756,744 patent/US20230005127A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
WO2021110392A1 (de) | 2021-06-10 |
CN114761145A (zh) | 2022-07-15 |
DE102019132830A1 (de) | 2021-06-10 |
US20230005127A1 (en) | 2023-01-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2092311B1 (de) | Vorrichtung zur inspektion von flaschen oder dergleichen behältern | |
EP2479123B1 (de) | Vorrichtung und Verfahren zum Transport von Artikeln | |
EP2132129B1 (de) | Verfahren zur überwachung, steuerung und optimierung von abfüllanlagen | |
EP2295157B1 (de) | Vorrichtung und Verfahren zum Kontrollieren von Gefäßverschlüssen | |
DE60028756T2 (de) | Methode und vorrichtung zur handhabung von ausgeworfenen sprtizgussteilen | |
EP3625740B1 (de) | Vorrichtung und verfahren zum steuern eines materialflusses an einem materialfluss-knotenpunkt | |
WO2014170079A1 (de) | Verfahren zur überwachung und steuerung einer abfüllanlage und vorrichtung zur durchführung des verfahrens | |
EP2295156B1 (de) | Fördereinrichtung mit einer Einrichtung zur Erkennung von umgefallenen Artikeln und Verfahren zu deren Steuerung | |
EP3501676B1 (de) | Flaschenreinigungsmaschine zum reinigen von flaschen | |
DE102017006566B3 (de) | Vorrichtung und Verfahren zur optischen Überwachung von Oberflächen eines Körpers | |
EP0996531B1 (de) | Vorrichtung und verfahren zur herstellung von kunststoffhohlkörpern | |
EP1600764A1 (de) | Verfahren und Vorrichtung zum Untersuchen lichtdurchlässiger Objekten | |
WO2020003180A1 (de) | Vorrichtung und verfahren zur inspektion von hängend förderbaren transporttaschen | |
DE102018205005A1 (de) | Flaschenreinigungsmaschine | |
WO2018108782A1 (de) | Verfahren und vorrichtung zum erfassen von störungen beim objekttransport | |
WO2021110682A1 (de) | Sortiersystem und sortierverfahren für blatttabak | |
EP4159329A1 (de) | Beheben von fehlplatzierungen auf einem sorterelement | |
EP3544780B1 (de) | Verfahren und system zum transportieren und manipulieren von zu gebinden zusammengefassten getränkebehältnissen sowie zum einleiten mindestens eines bestimmten prozesses im falle einer aus einer handhabung resultierenden beschädigung eines gebindes | |
EP4069440A1 (de) | Verfahren und vorrichtung zur erkennung von umgefallenen und/oder beschädigten behältern in einem behältermassenstrom | |
DE19959623C2 (de) | Verfahren und Anordnung zum Lokalisieren von zylinderförmigen Objekten | |
EP2452291A1 (de) | Verfahren zur kontinuierlichen ermittlung einer greifposition | |
WO2024017426A1 (de) | Vorrichtung und verfahren zum entleeren von behältern | |
DE102022122554A1 (de) | Verfahren zur Steuerung von Prozessen des Behälterhandlings und Behälterbehandlungsanlage zur Produktion, Abfüllung, Handhabung, Verpackung und/oder Förderung von Behältern | |
EP4276450A1 (de) | Vorrichtung und verfahren zum inspizieren von behältnissen mit positionserfassung | |
DE20317567U1 (de) | Vorrichtung zum Ausschleusen von Flaschenkästen an einer Transportstrecke |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20220408 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20240513 |