EP3056288A1 - Selektives sortierverfahren - Google Patents

Selektives sortierverfahren Download PDF

Info

Publication number
EP3056288A1
EP3056288A1 EP16305055.2A EP16305055A EP3056288A1 EP 3056288 A1 EP3056288 A1 EP 3056288A1 EP 16305055 A EP16305055 A EP 16305055A EP 3056288 A1 EP3056288 A1 EP 3056288A1
Authority
EP
European Patent Office
Prior art keywords
objects
zone
cluster
nature
operator
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP16305055.2A
Other languages
English (en)
French (fr)
Other versions
EP3056288B1 (de
Inventor
Jérémy DOUBLET
Christophe Gambier
Jean-François REZEAU
Alexander Mallinson
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Veolia Environnement SA
Original Assignee
Veolia Environnement SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=53008698&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=EP3056288(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Veolia Environnement SA filed Critical Veolia Environnement SA
Publication of EP3056288A1 publication Critical patent/EP3056288A1/de
Application granted granted Critical
Publication of EP3056288B1 publication Critical patent/EP3056288B1/de
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C5/00Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
    • B07C5/34Sorting according to other particular properties
    • B07C5/342Sorting according to other particular properties according to optical properties, e.g. colour
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C5/00Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
    • B07C5/02Measures preceding sorting, e.g. arranging articles in a stream orientating
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C5/00Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
    • B07C5/34Sorting according to other particular properties
    • B07C5/3416Sorting according to other particular properties according to radiation transmissivity, e.g. for light, x-rays, particle radiation
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C7/00Sorting by hand only e.g. of mail
    • B07C7/005Computer assisted manual sorting, e.g. for mail

Definitions

  • the present invention generally relates to a selective sorting method for identifying and sorting material objects of different natures, sizes, masses and shapes.
  • the present invention also relates to a device capable of implementing such a sorting method.
  • the invention relates to a method for selectively sorting a set of objects in the form of a cluster.
  • PPE personal protective equipment
  • workstations ventilation and appropriate infrastructure in particular
  • hydraulic machines can be used in sorting areas.
  • Examples of hydraulic machines include construction machines such as cranes, or hydraulic excavators.
  • these machines do not make it possible to reach satisfactory levels of performance and productivity.
  • it is difficult to precisely control the extraction of an object in particular and to be able to observe it in its entirety during its displacement.
  • robotic systems are used to efficiently and quickly sort fruits and vegetables according to various predefined criteria, including physical criteria, such as the shape, size or maturity of a product. organic.
  • international demand WO 2009/068792 describes a method and a selective sorting device improving those described in WO 98/19799 , in particular by allowing a very high rate. More particularly, in the device of WO 2009/068792 , it is possible to change the visual appearance of an image of a targeted object on a video screen.
  • clusters in the sense of the present invention, means a set of heterogeneous objects entangled and arranged randomly on each other, said objects being waste.
  • said method being characterized in that the nature of the object grasped or to be grasped by the robot is defined and assigned between steps e) and i), and consists in capturing at least one two-dimensional image in which said object appears, using at least one two-dimensional image sensor and to broadcast at least one of said two-dimensional images on a display screen that can be observed by an operator, said operator assigning a nature to said object displayed.
  • the cluster of objects that can be sorted by the process according to the invention may for example contain, in a nonlimiting manner, bulky objects or objects small sizes, whatever waste, industrial or domestic.
  • the term "waste” means any object, or more generally any movable good, the holder of which is undone or whose intention or obligation it is to dispose of, for the purposes of recovery or disposal, whether the holder is an industrialist, a community or an individual.
  • the objects of the cluster that can be sorted by the method according to the invention are for example organic waste or not, electronic waste, construction waste, furniture waste, industrial waste etc. .
  • the objects that are to be sorted are conveyed to a processing center for recovery, for example for recycling.
  • the objects to be sorted are typically arranged in the form of bulk or clusters, which may comprise a greater or lesser number of randomly entangled objects, in a particular and predefined area of the processing center. Then, they are usually transferred to processing means and other specific devices. Their transfer from a particular and predefined area of the treatment center to treatment means is achieved using known transfer means, such as, for example, shovels or conveyors.
  • the method according to the invention is thus fed by these object transfer means to be sorted, said objects to be sorted generally being in the form of clusters.
  • the method according to the invention is implemented to identify and sort a succession of clusters consisting of material objects of different natures, shapes and sizes.
  • the first step a) of the method according to the invention consists in feeding a zone of vision into objects lying generally in the form of clusters, the viewing zone being in the zone of action of a robot provided with one or more gripping members.
  • the viewing zone of the method according to the invention can be confused with the aforementioned particular and predefined area of the processing center, the objects to be sorted then for example directly discharged into this area of vision by a collection vehicle.
  • the supply of this zone of vision into objects can be carried out either according to a batch supply, or according to a continuous supply.
  • batch feed means a batch feed.
  • the feeding of the viewing zone is discontinuous. Only one cluster of objects at a time is processed. In this configuration, until all the objects have been sorted, the viewing area is not powered. However, when the last object to be valorised previously contained in the cluster is grasped by at least one gripping member of said robot, another cluster is moved into the viewing zone to be sorted later.
  • continuous supply means a power supply without deactivation of the means making it possible to supply the vision zone with objects.
  • objects to be sorted are moved to the continuous viewing area.
  • This zone of vision comprises at least two sensors for measuring electromagnetic radiation. It is also possible to add in this zone a source of incident electromagnetic radiation in order to allow an emission level of electromagnetic radiation sufficient by the cluster of objects to capture images representative of the real cluster.
  • the electromagnetic radiation measuring sensors are used to identify the nature of the unitary object located in said receiving area.
  • these electromagnetic radiation measuring sensors can be directly attached to the articulated mechanical arm of the robot.
  • the one or more image captures, made by the electromagnetic radiation measuring sensors are performed when the unitary object is under the influence of one of the gripping members of the robot.
  • the assignment of a nature to the unitary object is made during its movement between the vision zone and the reception zone. During this identification, it is therefore not necessary for the unitary object to be deposited in a particular area.
  • unitary object means any object initially initially contained in the cluster of objects to be sorted and which has been extracted therefrom.
  • step b) of the method according to the invention make it possible, in step b) of the method according to the invention, to produce at least two two-dimensional images of the cluster present in said zone of vision.
  • These two-dimensional images make it possible to reconstruct one or more virtual or electronic images of the cluster of objects in the zone of vision that can be viewed on a screen.
  • step c) of the method according to the invention with the aim of identifying all the possible gripping zones for the gripping member (s) of the robot and of identifying the most gripping member adapted for each of the possible catch zones, said zones being associated with objects present in the cluster.
  • a gripping zone or specific zone, is understood to mean an area that can be grasped by any gripping member of a robot. It should also be noted that several capture zones can be associated with an object contained in the cluster.
  • the processing of these two-dimensional images can, for example, be done using computational software and image processing software.
  • step d) of the method according to the invention After all the gripping zones have been identified through the two-dimensional image processing and analysis, the gripping zones are located in position and orientation, step d) of the method according to the invention.
  • the nature of the object grasped or to be grasped by the robot is assigned between steps e) and i), and consists in capturing at least one two-dimensional image using at least one two-dimensional image sensor and to broadcast at least two-dimensional images on a display screen that can be observed by an operator, the operator assigning a nature to a unitary object displayed.
  • the step of defining the nature of the seized unit object is carried out in the reception zone, and more particularly between the steps h) and i) previously mentioned.
  • the choice of one of the gripping zones, step e) of the method according to the invention can advantageously be performed automatically through the use of an algorithm.
  • the selection of a specific area is performed through the use of a controller, which does not require the intervention of an operator.
  • the robot's trajectory can be calculated using computational software.
  • a particular input path may be associated with each tap area. The method is then advantageous because it is possible to quickly enter and deposit a unitary object.
  • the unitary object associated with this gripping zone is transferred from the viewing zone to a gripping zone. reception.
  • the step of defining the nature of the object in the reception zone can be advantageously achieved by capturing at least one two-dimensional image of the unitary object in the reception area at the reception area. using at least one electromagnetic radiation measuring sensor and diffusing at least one of these two-dimensional images of the unitary object onto a display screen that can be observed by an operator, who in real time attributes a nature to the unitary object displayed in the reception area. In this embodiment, the intervention of an operator is necessary.
  • the same object is moved from the reception area to an outlet according to the nature assigned to it beforehand.
  • the step consisting in defining the nature of the object seized or to be grasped can be performed between the steps e) and f) mentioned above, starting from the virtual image of the cluster of objects from step b), which is broadcast on at least one viewing screen that can be observed by an operator, said operator assigning a nature to said object to be entered within the cluster of objects displayed.
  • a nature is attributed to an object of the cluster when it is located in the zone of vision. In this embodiment, the intervention of an operator is then necessary.
  • one of the capture zones can be targeted by an operator on the viewing screen broadcasting said virtual image representing the cluster located in the viewing zone.
  • the operator can assign a nature to the preselected gripping area, corresponding to a particular object, interactively through the use of a touch screen, or a screen video associated with a voice recognition system or a keyboard type system or any other system allowing the particular selection of a specific area.
  • any of the gripping members of a robot can enter this preselected gripping zone so that move the object from the viewing area to the receiving area.
  • the robot's take-off trajectory can be advantageously calculated by the use of computational software.
  • a particular input path may be associated with each tap area. The method is advantageous because it is possible to quickly grab and deposit objects.
  • the unitary object, associated with this preselected gripping zone is transferred from the viewing zone to a receiving zone.
  • the object can be advantageously moved from the reception area to an outlet predefined according to this nature.
  • the viewing zones of the cluster of objects and the receiving zone of the unitary object are distinct zones, that is to say separate processing center volumes.
  • the totality of the objects contained in the initial cluster is sorted, that is to say that all the objects are grasped by the robot and transits from the zone. from vision to the reception area.
  • all the objects contained in the initial cluster is not necessarily sorted and therefore does not necessarily pass from the viewing area to the receiving area.
  • an outlet for receiving the objects not grasped by any gripping member may be placed near the viewing area to allow its evacuation.
  • all of the objects not grasped by any of the gripping members of the robot is moved to a particular outlet by the use of any transfer means.
  • the device according to the invention is advantageous because it allows sorting remotely avoiding contact between an operator and any object to be sorted.
  • interfaces can be used to allow an operator to control and control the remote sorting device.
  • the device according to the invention makes it possible to sort clusters of objects containing multiple material objects, including waste, which may be of different natures, sizes and shapes.
  • the device according to the invention comprises means for providing a flow of objects in the form of clusters.
  • these means may be belt conveyors or roller conveyors, roller conveyors, ball tables, vibrating tables, devices mechanical means comprising gripping means, or any other device for moving an object cluster from an initial point to another point.
  • Collection bins in which the objects to be sorted are placed can also be used.
  • the bucket is static during the sorting process and during the gripping of each of the objects it contains.
  • a bucket containing new objects to be sorted is conveyed into the viewing zone and thus replaces the first bucket. It is also possible that the bucket is directly filled by a collection truck which avoids the replacement of the bucket.
  • the flow of objects feeds an area of the device according to the invention, called vision zone, in clusters of objects.
  • the device according to the invention further comprises a mechanical robot provided with at least one gripping member making it possible, in a first step, to grasp an object contained in the cluster present beforehand in the zone of vision, each object of the cluster being defined by one or more gripping zones, and in a second step to move the object entered from the viewing zone to another zone, called receiving zone.
  • a mechanical robot provided with at least one gripping member making it possible, in a first step, to grasp an object contained in the cluster present beforehand in the zone of vision, each object of the cluster being defined by one or more gripping zones, and in a second step to move the object entered from the viewing zone to another zone, called receiving zone.
  • the mechanical robot by means of at least one of its gripping members, moves the object associated with this particular catch area of the viewing area to the receiving area.
  • the device according to the invention comprises, especially in the viewing zone, sensors for measuring electromagnetic radiation, which can be measurement sensors in the visible or non-visible spectrum, such as gamma radiation sensors, radioelectric sensors, sensors infrared sensors, ultraviolet sensors, X-ray sensors or cameras.
  • sensors for measuring electromagnetic radiation can be measurement sensors in the visible or non-visible spectrum, such as gamma radiation sensors, radioelectric sensors, sensors infrared sensors, ultraviolet sensors, X-ray sensors or cameras.
  • the electromagnetic radiation measuring sensors are visible spectrum cameras. It should be noted that the aforementioned sensors can also be used in combination.
  • electromagnetic radiation measurements allow the robot to grab a particular object through a preselected grip area. These electromagnetic radiation measurements can also be analyzed and processed by computational software and image processing to allow the identification and location of all possible capture areas of each object contained in the cluster.
  • the electromagnetic radiation measuring sensor (s) may advantageously be connected to image analysis means.
  • an object contained in the cluster may be associated with several gripping zones or specific zones, the calculating and image processing software having the objective of identifying gripping surfaces and not objects.
  • Measurements of electromagnetic radiation can allow the development of one or more two-dimensional images.
  • the device may comprise processing and calculation means capable of automatically associating with a gripping surface selected by an operator or an automaton, the most suitable robot member.
  • the device may advantageously comprise processing and calculation means capable of automatically defining a gripping trajectory of an object by the compatible specific zone, preselected by an operator or an automaton, for a particular gripping member of a mechanical robot.
  • the reception zone may comprise one or more sensors for measuring electromagnetic radiation.
  • reception zone means an area whose volume is accessible by a robot as previously described.
  • the electromagnetic radiation measurement sensors used are sensors making it possible to reproduce one or more virtual or electronic images of the object located in the reception zone.
  • these sensors may be infrared or near infrared radiation sensors, ultraviolet radiation sensors, X-ray sensors, electromagnetic radiation sensors according to the visible or non-visible spectrum, gamma radiation sensors, laser scanning distance sensors, and preferably the electromagnetic radiation measuring sensors are visible spectrum cameras.
  • the aforementioned sensors can also be used in combination.
  • the images recovered by one or more of the aforementioned sensors can be viewed on a touch screen.
  • This touch screen can for example comprise two zones.
  • a first zone can be used to display an image from electromagnetic radiation measuring sensors.
  • a second area of the screen includes sub-categories respectively corresponding to predetermined natures.
  • an operator can assign to the unit object that he displays on said touch screen a particular nature by selecting a particular subcategory.
  • the video screen is not tactile, but is associated with a voice recognition system or a keyboard type system or any other system for attributing a particular nature to the object displayed on said screen.
  • the seized object is evacuated from said receiving area to an outlet, by means of routing according to the nature that has been assigned to it.
  • conveying means may be belt conveyors or roller conveyors, roller conveyors, ball tables, vibrating tables, mechanical devices comprising gripping means, or any other device for moving a unitary object of an initial point to another point.
  • One of the advantageous means of routing is the same mechanical robot used to move the cluster object from the viewing area to the receiving area.
  • the object transiting on these conveyance means is moved into a predefined outlet according to the nature that has been assigned to it.
  • the outlets may include manipulator arms or robots adapted to the characteristics of the objects to be extracted. These outlets may also include pneumatic ejection devices conveyor belt jets, air nozzles, referral systems, pushers using jacks, traps, robots. Extraction means combining different aforementioned ejection devices can be applied.
  • the device according to the invention may comprise means for detecting and tracking the movements and positions of a particular object, between the gripping device of a robot and an outlet, as a function of time.
  • These means may comprise electromagnetic radiation measuring sensors as mentioned above.
  • FIGS. figures 1 and 2 robotic solutions marketed by the company Sileane or the company AKEO.
  • These devices comprise a robot which comprises a polyarticulate system provided with one or more gripping members capable of grasping an object by a specific area. To facilitate the reading of the figures, only one gripping member is shown on the figures 1 and 2 .
  • the robot may comprise at least two gripping members, the first using a so-called “suction cup” technology and the other a so-called “clamp” technology.
  • This robot is not the one illustrated in the figures.
  • the robot comprises a single gripping member using the "clamp" technology.
  • the figure 1 describes a device 10 according to the invention, for extracting particular objects contained in a cluster, depending on their nature.
  • the cluster of objects includes a bulk volume of heterogeneous objects placed randomly so that the objects become entangled.
  • the cluster of objects for processing is arranged on a first belt conveyor 11.
  • This first belt conveyor 11 is able to feed a zone, called the vision zone 12, into a cluster of material objects.
  • This viewing zone 12 is irradiated with electromagnetic radiation from radiation sources in order to produce one or more images of the cluster of objects situated in the viewing zone 12.
  • the device of the figure 1 comprises sensors for measuring electromagnetic radiation in order to produce one or more two-dimensional images of the cluster of objects situated in the viewing zone 12.
  • the electromagnetic radiation measuring sensors are configured to acquire successive two-dimensional images of the cluster located in the viewing zone 12. It should be noted that the captured images cover the entire cluster of objects.
  • the images are captured through the use of a camera 19a in visible spectrum.
  • One or more of said images of the cluster of captured objects are then processed and analyzed to allow the identification and location of each possible setting area by a gripping member 18 of the polyarticulate robot 14.
  • the sensors for measuring electromagnetic radiation are, for example, coupled to means of processing, which may be computers and other software, configured to process images from said sensors.
  • a gripping member 18 grasps the defined gripping zone.
  • the device of the figure 1 can also use computational software and image processing to also allow to define the fastest and shortest possible capture trajectory, for a given gripping member 18.
  • the speed of the flow of objects towards the viewing zone 12 is possibly not constant.
  • the speed of the flow of objects decreases, or even vanishes, so that the sensors present in the viewing zone 12 can capture at least two two-dimensional images representing the cluster of objects.
  • the sensors After each capture, the sensors capture new images of the cluster of objects. In this way, the object to be captured, which could be moved by the input of a previous object, will still be located and seized.
  • the first belt conveyor 11 can be put back into operation in order to bring into this viewing zone 12 a new cluster of objects to be sorted.
  • the totality of the objects contained in the initial cluster to be sorted passes from the viewing zone to the reception zone after having been grasped by the robot.
  • each object deposited in this reception zone 13 is assigned a particular nature. This allocation is made thanks to the intervention of an operator who can use an interface. This interface allows the operator to assign a nature to a given object at a given time.
  • the interface is coupled to one or more sensors for measuring electromagnetic radiation.
  • the two-dimensional image sensor is a camera 19b operating in the visible spectrum.
  • This camera 19b is connected to a screen and the captured images appear on this screen.
  • These images can be preprocessed by computer to present a certain contrast to the operator and thus give him indications for example on properties or the nature of said object ( figure 3A ).
  • an object is plastic
  • a specific color or texture that facilitates recognition of the operator can be assigned to the object. This step is particularly interesting for subjects whose effectiveness of recognition does not have good performance.
  • the second screen area may further comprise a subcategory 23 that does not correspond to any particular nature, but in which an operator can classify all objects regardless of their nature.
  • This subcategory 23 can then complement the other subcategories or be the only one available. In the latter case, possibly, it is possible to have only one outlet 16b in which are ejected all objects located on the second conveyor belt 15.
  • the object is directed to a predefined outlet 16, at first, thanks to the use of a second conveyor belt 15 and in a second step through the use of one or more extraction means 17.
  • the extraction means 17 make it possible to extract the objects situated on the second belt conveyor 15 and to route them to the appropriate outlets 16 intended to receive them.
  • the figure 1 shows that these outlets 16 comprise pneumatic ejection devices using jacks.
  • means can be used to record and follow the movements and positions of a particular object, between the gripping device 18 of a robot 14 and an outlet 16, as a function of time.
  • figure 2 describes a device 20 according to the invention for selecting in particular one or more objects contained in a cluster, according to the second embodiment of the method according to the invention.
  • the cluster of objects includes a bulk volume of heterogeneous objects placed randomly so that the objects become entangled.
  • the cluster of objects for processing is arranged on a first belt conveyor 11.
  • This first belt conveyor 11 is able to feed a zone, called the vision zone 12, into a cluster of material objects.
  • a visible spectrum camera 19c for producing one or more two-dimensional images of the cluster of objects situated in the viewing zone 12.
  • the images captured by the camera 19c cover the entire cluster of objects.
  • the camera 19c is connected to a screen and the captured images appear on this screen. These images can be preprocessed in order to present a certain contrast to the operator and thus give him indications, for example on properties or the nature of said object selected by the designation of a particular gripping area ( figure 3B ).
  • an object is plastic
  • a specific color or texture that facilitates recognition of the operator can be assigned to the object. This step is particularly interesting for subjects whose effectiveness of recognition does not have good performance.
  • the camera 19c is, for example, coupled to processing means, which may be computers and other software, configured to process the images from the camera 19c and thus allow the identification and location of all the zones take possible associated with each object of the cluster.
  • processing means which may be computers and other software, configured to process the images from the camera 19c and thus allow the identification and location of all the zones take possible associated with each object of the cluster.
  • the screen on which the images captured by the camera 19c appear is tactile.
  • the second screen area may include a sub-category 23 that does not correspond to any particular nature, but in which an operator can classify all the objects selected and previously contained in the cluster. In this last case, it is possible to have only one outlet 16b in which are ejected all the objects that have been deposited in the receiving area 13 located on the second conveyor belt 15.
  • the possible gripping zones are represented on the first zone of said touch screen 20 by virtual circles.
  • an operator may designate a finger pointing area on the touch screen 20.
  • the selection of a gripping zone 25 can be corrected. For example, selecting an already selected tap area 25 deselects it.
  • an operator can assign a nature to a given object.
  • the assignment of a nature to a particular engagement zone is as follows. An operator first selects a take zone 25 and then selects a subcategory 23 of allocation. According to this configuration, the selected gripping zone 25 is marked with a graphical sub-referent such as, for example, a colored circle 24.
  • unitary category that is to say that a single category is assigned automatically to all the objects designated by an operator after having selected a gripping zone. possible of an object contained in the cluster.
  • unitary category is previously defined by the operator.
  • each object has been assigned a nature.
  • this second conveyor belt 15 is in permanent operation, an object grasped by a gripper 18 of a robot 14, and then deposited on the second belt conveyor 15, is very quickly set in motion. In this way, each of the objects seized in the initial cluster succeeds throughout this second conveyor belt 15.
  • the second belt conveyor 15 thus makes it possible to extract the unitary object previously deposited in the receiving zone 13.
  • each object Since each object has been assigned a nature, each of them can be moved to a specific 16b outlet.
  • the outlet 16b comprises a robot provided with a gripping member capable of grasping any object located on the second conveyor belt 15 and to move it into a predefined outlet 16b according to the nature that has been assigned to it.
  • the robot 14 is used as extraction means 17.
  • means can be used to record and follow the movements and positions of a particular object, between the gripping device 18 of a robot 14 and an outlet 16b, as a function of time.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Toxicology (AREA)
  • Manipulator (AREA)
EP16305055.2A 2015-02-10 2016-01-21 Selektives sortierverfahren und entsprechende vorrichtung Active EP3056288B1 (de)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
FR1551084A FR3032366B1 (fr) 2015-02-10 2015-02-10 Procede de tri selectif

Publications (2)

Publication Number Publication Date
EP3056288A1 true EP3056288A1 (de) 2016-08-17
EP3056288B1 EP3056288B1 (de) 2018-03-14

Family

ID=53008698

Family Applications (1)

Application Number Title Priority Date Filing Date
EP16305055.2A Active EP3056288B1 (de) 2015-02-10 2016-01-21 Selektives sortierverfahren und entsprechende vorrichtung

Country Status (4)

Country Link
US (1) US9682406B2 (de)
EP (1) EP3056288B1 (de)
DK (1) DK3056288T3 (de)
FR (1) FR3032366B1 (de)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108144863A (zh) * 2016-12-02 2018-06-12 朱塞佩·普里泰利G.P.咨询简易两合公司 家具行业的面板生产线中精益生产的优化制造方法和系统
CN110076768A (zh) * 2018-01-25 2019-08-02 发那科株式会社 物品搬运系统以及机器人系统
CN110856846A (zh) * 2018-08-24 2020-03-03 胜宏科技(惠州)股份有限公司 一种pcb板自动分拣系统及方法
CN111805309A (zh) * 2020-07-02 2020-10-23 哈尔滨工业大学 用于硬脆单晶圆柱外圆超声振动辅助磨削的全自动磨床
US11660762B2 (en) 2018-05-11 2023-05-30 Mp Zenrobotics Oy Waste sorting robot

Families Citing this family (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
NL2011066C2 (en) * 2013-06-28 2015-01-05 Ig Specials B V Apparatus and method for sorting plant material.
FR3031048B1 (fr) * 2014-12-24 2016-12-30 Solystic Machine de tri postal avec une entree d'alimentation comprenant un bras robotise et un convoyeur a plat incline
FR3032365B1 (fr) * 2015-02-10 2017-02-03 Veolia Environnement-VE Procedure de tri selectif
CA2998544C (en) 2015-09-11 2023-05-09 Berkshire Grey, Inc. Robotic systems and methods for identifying and processing a variety of objects
US10625305B2 (en) 2015-12-04 2020-04-21 Berkshire Grey, Inc. Systems and methods for dynamic processing of objects
US10730078B2 (en) 2015-12-04 2020-08-04 Berkshire Grey, Inc. Systems and methods for dynamic sortation of objects
US9937532B2 (en) 2015-12-18 2018-04-10 Berkshire Grey Inc. Perception systems and methods for identifying and processing a variety of objects
CN110199231B (zh) * 2016-11-08 2023-12-15 伯克希尔格雷运营股份有限公司 用于处理物体的系统和方法
EP4299490A3 (de) 2016-11-28 2024-03-20 Berkshire Grey Operating Company, Inc. System zur vereinzelung von objekten zur verarbeitung
CA3155737C (en) 2016-12-06 2023-11-14 Berkshire Grey Operating Company, Inc. Systems and methods for providing for the processing of objects in vehicles
CN114132676A (zh) 2017-03-20 2022-03-04 伯克希尔格雷股份有限公司 用于处理物体的包括自动穿梭系统的系统和方法
CA3057334C (en) 2017-03-24 2023-10-31 Berkshire Grey, Inc. Systems and methods for processing objects, including automated processing
CA3186213A1 (en) 2017-04-18 2018-10-25 Berkshire Grey Operating Company, Inc. Systems and methods for processing objects including space efficient distribution stations and automated output processing
US11205059B2 (en) 2017-04-18 2021-12-21 Berkshire Grey, Inc. Systems and methods for separating objects using conveyor transfer with one or more object processing systems
US11301654B2 (en) 2017-04-18 2022-04-12 Berkshire Grey Operating Company, Inc. Systems and methods for limiting induction of objects to one or more object processing systems
US11055504B2 (en) 2017-04-18 2021-07-06 Berkshire Grey, Inc. Systems and methods for separating objects using a vacuum roller with one or more object processing systems
US11200390B2 (en) 2017-04-18 2021-12-14 Berkshire Grey, Inc. Systems and methods for separating objects using drop conveyors with one or more object processing systems
US11080496B2 (en) 2017-04-18 2021-08-03 Berkshire Grey, Inc. Systems and methods for separating objects using vacuum diverts with one or more object processing systems
US11373134B2 (en) 2018-10-23 2022-06-28 Berkshire Grey Operating Company, Inc. Systems and methods for dynamic processing of objects with data verification
US11416695B2 (en) 2017-04-18 2022-08-16 Berkshire Grey Operating Company, Inc. Systems and methods for distributing induction of objects to a plurality of object processing systems
CA3061181C (en) 2017-04-24 2023-10-03 Berkshire Grey, Inc. Systems and methods for providing singulation of objects for processing using object movement redistribution
JP2018203480A (ja) * 2017-06-07 2018-12-27 株式会社東芝 仕分装置および仕分システム
CN107661865A (zh) * 2017-06-19 2018-02-06 福建南方路面机械有限公司 建筑垃圾分拣前端检测系统
CN107214108A (zh) * 2017-06-19 2017-09-29 太仓弘杉环保科技有限公司 一种高效智能化生产加工系统的工作方法
SE543130C2 (en) 2018-04-22 2020-10-13 Zenrobotics Oy A waste sorting robot gripper
EP3870513A4 (de) * 2018-10-25 2022-08-10 And y Knot Innovation and Sales Inc. Vorrichtung zum stapeln und verpacken
EP3871172A1 (de) 2018-10-25 2021-09-01 Berkshire Grey, Inc. Systeme und verfahren zum lernen zur extrapolation optimaler zielrouting- und handhabungsparameter
CN109926354B (zh) * 2019-03-18 2020-12-25 佛山市利普达机械配件有限公司 一种用于物流分拣用传送带
CN110328148B (zh) * 2019-07-10 2021-06-08 广东华中科技大学工业技术研究院 一种屏幕检测筛选装置
CN110756464A (zh) * 2019-10-24 2020-02-07 北京京日东大食品有限公司 一种原料豆选豆工艺
CN111359908A (zh) * 2020-02-21 2020-07-03 苏州先迅检测科技有限公司 一种自动检测设备
CN111570326A (zh) * 2020-04-07 2020-08-25 西安航空学院 机电自动化物料分拣装置
CN111842178A (zh) * 2020-06-18 2020-10-30 苏州小优智能科技有限公司 一种流水线鞋坯3d扫描及智能分拣的设备及方法
CN112337807A (zh) * 2020-11-15 2021-02-09 李童 一种基于电子商务的自动分拣装置
WO2023042389A1 (ja) * 2021-09-17 2023-03-23 株式会社Pfu 物体処理装置
CN115138579A (zh) * 2022-06-28 2022-10-04 苏州启航电子有限公司 Ai视觉检测设备
FR3138331A1 (fr) 2022-07-26 2024-02-02 Tellux Procede de traitement automatique de deblais en forme de granulats sur convoyeur equipe d’un imageur hyperspectral
CN115365166B (zh) * 2022-10-26 2023-03-24 国家电投集团科学技术研究院有限公司 垃圾识别及分拣系统和分拣方法

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1998019799A1 (en) 1996-11-04 1998-05-14 National Recovery Technologies, Inc. Teleoperated robotic sorting system
WO2009068792A2 (fr) 2007-11-07 2009-06-04 Veolia Proprete Procede et dispositif de tri selectif d'objets

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
IT1258006B (it) * 1992-01-13 1996-02-20 Gd Spa Sistema e metodo per il prelievo automatico di oggetti
FR2725640B1 (fr) 1994-10-12 1997-01-10 Pellenc Sa Machine et procede pour le tri d'objets divers a l'aide d'au moins un bras robotise
US5675710A (en) 1995-06-07 1997-10-07 Lucent Technologies, Inc. Method and apparatus for training a text classifier
FR2842671B1 (fr) 2002-07-22 2005-03-04 Inst Nat Rech Inf Automat Compression de donnees numeriques robuste au bruit de transmission
US20070208455A1 (en) 2006-03-03 2007-09-06 Machinefabriek Bollegraaf Appingedam B.V. System and a method for sorting items out of waste material
US8177069B2 (en) * 2007-01-05 2012-05-15 Thomas A. Valerio System and method for sorting dissimilar materials
DE102007038837A1 (de) 2007-08-16 2009-02-19 BIBA - Bremer Institut für Produktion und Logistik GmbH Verfahren und Vorrichtung zur Umsetzung von Stückgut
FI20106090A0 (fi) 2010-10-21 2010-10-21 Zenrobotics Oy Menetelmä kohdeobjektin kuvien suodattamiseksi robottijärjestelmässä
US9067744B2 (en) 2011-10-17 2015-06-30 Kabushiki Kaisha Yaskawa Denki Robot system, robot, and sorted article manufacturing method
JP6000579B2 (ja) * 2012-03-09 2016-09-28 キヤノン株式会社 情報処理装置、情報処理方法
JP6364836B2 (ja) * 2014-03-14 2018-08-01 セイコーエプソン株式会社 ロボット、ロボットシステム、及び制御装置
US9266148B2 (en) * 2014-06-27 2016-02-23 Key Technology, Inc. Method and apparatus for sorting

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1998019799A1 (en) 1996-11-04 1998-05-14 National Recovery Technologies, Inc. Teleoperated robotic sorting system
WO2009068792A2 (fr) 2007-11-07 2009-06-04 Veolia Proprete Procede et dispositif de tri selectif d'objets

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108144863A (zh) * 2016-12-02 2018-06-12 朱塞佩·普里泰利G.P.咨询简易两合公司 家具行业的面板生产线中精益生产的优化制造方法和系统
CN110076768A (zh) * 2018-01-25 2019-08-02 发那科株式会社 物品搬运系统以及机器人系统
US11660762B2 (en) 2018-05-11 2023-05-30 Mp Zenrobotics Oy Waste sorting robot
CN110856846A (zh) * 2018-08-24 2020-03-03 胜宏科技(惠州)股份有限公司 一种pcb板自动分拣系统及方法
CN111805309A (zh) * 2020-07-02 2020-10-23 哈尔滨工业大学 用于硬脆单晶圆柱外圆超声振动辅助磨削的全自动磨床
CN111805309B (zh) * 2020-07-02 2021-07-20 哈尔滨工业大学 用于硬脆单晶圆柱外圆超声振动辅助磨削的全自动磨床

Also Published As

Publication number Publication date
FR3032366A1 (fr) 2016-08-12
FR3032366B1 (fr) 2017-02-03
EP3056288B1 (de) 2018-03-14
US20160228921A1 (en) 2016-08-11
US9682406B2 (en) 2017-06-20
DK3056288T3 (en) 2018-04-23

Similar Documents

Publication Publication Date Title
EP3056288B1 (de) Selektives sortierverfahren und entsprechende vorrichtung
EP3056289B1 (de) Selektives sortierverfahren und entsprechende vorrichtung
EP3134234B1 (de) Verfahren und anlage zum automatischen greifen eines objekts
US9008832B2 (en) Diamond sorting system
US20190075732A1 (en) Automated plant trimmer
FR2923165A1 (fr) Procede et dispositif de tri selectif d'objets
FR2725640A1 (fr) Machine et procede pour le tri d'objets divers a l'aide d'au moins un bras robotise
CN112113917A (zh) 用于处理收获的块根作物的方法和设备
FR3100140A1 (fr) Dispositif de tri pour produits agricoles, et procédé correspondant
CA3020716A1 (fr) Procede et dispositif d'orientation d'un fruit ombilique notamment en vue de son emballage
CN107138432A (zh) 非刚性物体分拣方法和装置
EP1082913A2 (de) Verfahren und Vorrichtung zur Behandlung von Pflanzgut nach der Ernte
CA2550440C (fr) Procede pour fusionner des lettres et des objets postaux de grand format et/ou non mecanisables dans une tournee unique du facteur
EP1703753B1 (de) Verfahren zur Betriebsanalyse eines zellularen Mobilfunknetzes
FR3044573A1 (fr) Procede et installation permettant de constituer un lot de pieces a partir de pieces situees dans des zones de stockage differentes
WO2011030042A1 (fr) Procédé d'aide à l'indentification de produits non conformes triés manuellement et installation pour sa mise en oeuvre
EP2812161B1 (de) Einheit und verfahren zum automatischen festhaken von teilen auf komplexen trägern
FR3017369A1 (fr) Procede et installation de depose de produits individuels sur des supports plans alveoles non indexes
EP4281930A1 (de) Verfahren zur optischen inspektion eines entlang einer produktionslinie bewegten elements
CN215218257U (zh) 自动抽梗分梗装置
FR2868971A1 (fr) Methode d'identification et/ou de tri par pieces, d'objets ou de produits utilisables pour la preparation de commande, le suivi de production ou la validation de commande
Liu et al. Laser point detection based on improved target matching method for application in home environment human-robot interaction
Mhamed et al. Advances in apple’s automated orchard equipment: A comprehensive research
EP3310502B1 (de) Verfahren und vorrichtung zum manuellen vereinen von postgegenständen mit einem stapel von bereits sortierten postgegenständen
WO2015014945A1 (fr) Système de programmation d'un système d'analyse de situation embarqué sur un porteur comportant au moins un système d'écoute embarqué

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

17P Request for examination filed

Effective date: 20160916

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20170203

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

INTG Intention to grant announced

Effective date: 20171103

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

Free format text: NOT ENGLISH

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

Ref country code: AT

Ref legal event code: REF

Ref document number: 978338

Country of ref document: AT

Kind code of ref document: T

Effective date: 20180315

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Free format text: LANGUAGE OF EP DOCUMENT: FRENCH

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 602016002041

Country of ref document: DE

Representative=s name: NOVAGRAAF BREVETS, FR

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602016002041

Country of ref document: DE

REG Reference to a national code

Ref country code: NL

Ref legal event code: FP

REG Reference to a national code

Ref country code: DK

Ref legal event code: T3

Effective date: 20180416

REG Reference to a national code

Ref country code: SE

Ref legal event code: TRGR

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180614

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 978338

Country of ref document: AT

Kind code of ref document: T

Effective date: 20180314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180614

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180615

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

REG Reference to a national code

Ref country code: DE

Ref legal event code: R026

Ref document number: 602016002041

Country of ref document: DE

PLBI Opposition filed

Free format text: ORIGINAL CODE: 0009260

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180716

PLAX Notice of opposition and request to file observation + time limit sent

Free format text: ORIGINAL CODE: EPIDOSNOBS2

26 Opposition filed

Opponent name: ZENROBOTICS OY

Effective date: 20181214

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

PLBB Reply of patent proprietor to notice(s) of opposition received

Free format text: ORIGINAL CODE: EPIDOSNOBS3

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190131

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190131

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20190121

PLBD Termination of opposition procedure: decision despatched

Free format text: ORIGINAL CODE: EPIDOSNOPC1

PLBP Opposition withdrawn

Free format text: ORIGINAL CODE: 0009264

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

REG Reference to a national code

Ref country code: DE

Ref legal event code: R100

Ref document number: 602016002041

Country of ref document: DE

PLBM Termination of opposition procedure: date of legal effect published

Free format text: ORIGINAL CODE: 0009276

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: OPPOSITION PROCEDURE CLOSED

27C Opposition proceedings terminated

Effective date: 20200409

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180714

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20160121

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20180314

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20230123

Year of fee payment: 8

Ref country code: DK

Payment date: 20230123

Year of fee payment: 8

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: SE

Payment date: 20230123

Year of fee payment: 8

Ref country code: BE

Payment date: 20230123

Year of fee payment: 8

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230601

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20240123

Year of fee payment: 9

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: LU

Payment date: 20240122

Year of fee payment: 9

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20240119

Year of fee payment: 9

Ref country code: GB

Payment date: 20240124

Year of fee payment: 9