WO2019132168A1 - 수술영상데이터 학습시스템 - Google Patents
수술영상데이터 학습시스템 Download PDFInfo
- Publication number
- WO2019132168A1 WO2019132168A1 PCT/KR2018/010333 KR2018010333W WO2019132168A1 WO 2019132168 A1 WO2019132168 A1 WO 2019132168A1 KR 2018010333 W KR2018010333 W KR 2018010333W WO 2019132168 A1 WO2019132168 A1 WO 2019132168A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- learning
- model
- client
- server
- surgical
- Prior art date
Links
- 238000012360 testing method Methods 0.000 claims description 33
- 238000000034 method Methods 0.000 claims description 30
- 238000001356 surgical procedure Methods 0.000 claims description 24
- 238000012549 training Methods 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 3
- 238000004891 communication Methods 0.000 description 10
- 238000010586 diagram Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 238000010801 machine learning Methods 0.000 description 6
- 238000002059 diagnostic imaging Methods 0.000 description 5
- 238000002432 robotic surgery Methods 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 4
- 230000033001 locomotion Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 239000000284 extract Substances 0.000 description 3
- 210000000056 organ Anatomy 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 238000004873 anchoring Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/14—Transformations for image registration, e.g. adjusting or mapping for alignment of images
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/20—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the management or administration of healthcare resources or facilities, e.g. managing hospital staff or surgery rooms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2065—Tracking using image or pattern recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
Definitions
- the present invention relates to a surgical image data learning system.
- Deep learning is defined as a set of machine learning algorithms that try to achieve high levels of abstraction (a task that summarizes key content or functions in large amounts of data or complex data) through a combination of several nonlinear transformation techniques. Deep learning can be viewed as a field of machine learning that teaches computers how people think in a big way.
- a problem to be solved by the present invention is to provide a surgical image data learning system.
- a surgical image data learning system including a server for learning one or more models for providing surgical guide information, at least one model learned in the server, And a client for recognizing a surgical image obtained during the operation using the model and providing surgical guide information corresponding to the recognized surgical image, wherein the client includes: first learning data obtained from the surgical image acquired during the operation; To the server, receives information for updating the stored model from the server, and updates the stored model based on the received information.
- the at least one model may include a first model for segmenting and recognizing a surgical image and a second model for calculating an optimized surgical method
- the first learning data may include at least one of a first model and a second model
- at least one model stored in the client may include the first model.
- the client may segment the surgical image acquired during the operation into one or more surgical steps, recognize the divided one or more surgical steps, and transmit the recognition result to the server as the first learning data .
- the client may assign a predetermined code to each of the divided one or more surgical steps, and may transmit the predetermined code assigned to each of the divided one or more surgical steps to the server.
- the client may compare the model stored in the client with the model stored in the server and update the model stored in the client from the server when the model stored in the server is an updated model than the model stored in the client And update the model stored in the client based on the received information.
- the server may acquire information for dividing at least one surgical image and the at least one surgical image, and may generate second learning data using the obtained information.
- the server receives the learning condition, extracts a data set corresponding to the input learning condition from the second learning data, and performs learning on the at least one model using the extracted data set can do.
- the server may generate a test data set corresponding to the input learning condition, perform a learning test on the at least one model using the generated test data set, Lt; / RTI >
- the server may also be configured to receive one or more new second learning data, to generate a data set that includes at least one new second learning data, and to use the generated data set for the at least one model Learning can be performed.
- the server may also acquire a test data set and perform a learning test on the at least one model learned using the generated data set using the test data set, Lt; / RTI >
- the server may generate the data set including the new second learning data and existing second learning data selected according to a predetermined criterion, and use the generated data set to generate the second learning data for the at least one model Learning can be performed.
- a method for learning surgical image data comprising: learning one or more models of a server providing surgical guide information; Receiving training data obtained from a surgical image obtained during surgery by the client from the client, updating the learned one or more models, and transmitting information about the updated at least one model to the client To the mobile station.
- a server comprising a memory for storing one or more instructions and a processor for executing the one or more instructions stored in the memory, the processor executing the one or more instructions
- the method comprising the steps of: learning one or more models providing surgical guide information; transmitting information about the learned at least one model to a client; transmitting learning data, which is obtained from a surgical image acquired by the client during surgery, , Updating the learned one or more models, and transmitting information about the updated at least one model to the client.
- a computer program stored in a recording medium readable by a computer, the computer program being capable of performing a surgical image data learning method according to an embodiment of the present invention in combination with a hardware computer.
- a surgical image data learning system based on a server and a client, so that it is possible to provide quick data processing and operation guide information during surgery, and, when the operation is not performed, And the model stored in the client can be updated.
- FIG. 1 is a simplified schematic diagram of a system capable of performing robotic surgery in accordance with the disclosed embodiments.
- FIG. 2 is a diagram illustrating a surgical image data learning system according to an embodiment.
- FIG. 3 is a flowchart briefly illustrating an operation of a surgical image data learning system according to an embodiment.
- FIG. 4 is a flowchart illustrating an operation of a surgical image data learning system according to an embodiment.
- FIG. 5 is a block diagram conceptually illustrating a method of scheduling learning according to an embodiment.
- FIG. 6 is a configuration diagram of a server according to an embodiment.
- the term “part” or “module” refers to a hardware component, such as a software, FPGA, or ASIC, and a “component” or “module” performs certain roles. However, “part” or “ module “ is not meant to be limited to software or hardware. A “module “ or “ module “ may be configured to reside on an addressable storage medium and configured to play back one or more processors. Thus, by way of example, “a” or " module " is intended to encompass all types of elements, such as software components, object oriented software components, class components and task components, Microcode, circuitry, data, databases, data structures, tables, arrays, and variables, as used herein. Or " modules " may be combined with a smaller number of components and "parts " or " modules " Can be further separated.
- FIG. 1 there is shown a simplified schematic representation of a system capable of performing robotic surgery in accordance with the disclosed embodiments.
- the robotic surgery system includes a medical imaging apparatus 10, a server 20, a control unit 30 provided in an operating room, a display 32, and a surgical robot 34.
- the medical imaging equipment 10 may be omitted from the robotic surgery system according to the disclosed embodiment.
- the surgical robot 34 includes a photographing device 36 and a surgical tool 38.
- robotic surgery is performed by the user controlling the surgical robot 34 using the control unit 30.
- robot surgery may be performed automatically by the control unit 30 without user control.
- the server 20 is a computing device including at least one processor and a communication unit.
- the control unit 30 includes a computing device including at least one processor and a communication unit. In one embodiment, the control unit 30 includes hardware and software interfaces for controlling the surgical robot 34.
- the photographing apparatus 36 includes at least one image sensor. That is, the photographing device 36 includes at least one camera device, and is used for photographing a body part of the object, that is, a surgical part. In one embodiment, the imaging device 36 includes at least one camera coupled with a surgical arm of the surgical robot 34.
- the image photographed at the photographing device 36 is displayed on the display 340.
- the surgical robot 34 includes one or more surgical tools 38 that can perform cutting, clipping, anchoring, grabbing, etc., of the surgical site.
- the surgical tool 38 is used in combination with the surgical arm of the surgical robot 34.
- the control unit 30 receives information necessary for surgery from the server 20, or generates information necessary for surgery and provides the information to the user. For example, the control unit 30 displays on the display 32 information necessary for surgery, which is generated or received.
- the user operates the control unit 30 while viewing the display 32 to perform the robot surgery by controlling the movement of the surgical robot 34.
- the server 20 generates information necessary for robot surgery using the medical image data of the object (patient) photographed beforehand from the medical imaging apparatus 10, and provides the generated information to the control unit 30.
- the control unit 30 provides the information received from the server 20 to the user by displaying the information on the display 32 or controls the surgical robot 34 using the information received from the server 20.
- the means that can be used in the medical imaging equipment 10 is not limited, and various other medical imaging acquiring means such as CT, X-Ray, PET, MRI and the like may be used.
- the surgical image obtained in the photographing device 36 is transmitted to the control section 30.
- control unit 30 may segment the surgical image obtained during the operation in real time.
- control unit 30 transmits a surgical image to the server 20 during or after surgery.
- the server 20 can divide and analyze the surgical image.
- FIG. 2 is a diagram illustrating a surgical image data learning system according to an embodiment.
- the surgical image data learning system includes a client 100 and a server 200.
- client 100 and server 200 are computing devices that include at least one processor.
- the client 100 may be a computing device in an operating room (surgical site).
- the client 100 may correspond to the control unit 30 shown in FIG.
- the client 100 may be provided in the operating room as a separate computing device from the control unit 30, and may be connected to the control unit 30.
- a connection includes not only a physical connection but also an electronic connection concept, and it can also be understood as a concept of connection that the communication state is mutually communicable.
- the client 100 may be connected to the control unit 30 by wire or wireless, and may be in a state where they can communicate with each other using short-range wireless communication or network communication.
- the client 100 determines the surgical situation in the operating room and provides corresponding surgical guide information.
- the client 100 uses the learned model to determine the surgical situation and provide corresponding surgical guide information.
- the learned model may be a learned model using machine learning.
- Machine learning can mean, but is not limited to, deep learning.
- the client 100 acquires a surgical image
- the server 200 performs learning using the acquired surgical image. Since machine learning is a task requiring a lot of resources, it may be difficult to perform in the client 100. Further, since a lot of learning data is required for the machine learning, it is difficult for sufficient learning data to be secured by the information collected by the client 100 alone.
- the server 200 constructs learning data using information collected from different sources, determines a surgical situation based on the learning data, and learns a model capable of providing surgical guide information.
- the client 100 must determine the operation status in real time and provide necessary surgical guide information. Therefore, in order to save time required for communication with the server 200 and to prevent errors due to communication errors, the information about the model learned by the server 200 is stored in the client 100 in advance.
- the client 100 can restore the model learned in the server 200 using the information obtained from the server 200 and provide the operation guide information in real time using the model.
- FIG. 3 is a flowchart briefly illustrating an operation of a surgical image data learning system according to an embodiment.
- the client 100 and the server 200 described in connection with FIG. 2 correspond to the client 100 and the server 200 in FIG. 3, respectively.
- step S210 the server 200 learns a model for providing surgical guide information using the learning data obtained from the surgical image.
- the model providing the surgical guide information includes a first model for recognizing the surgical image in a divided manner, and a second model for calculating the optimized surgical method.
- the training data may refer to the surgical image itself, or may be data obtained by analyzing the surgical image.
- the training data may include data obtained by dividing a surgical image into predetermined motion units.
- the learning data may include data labeled with a surgical image. The type of learning data is not limited to this.
- the server 200 transmits information about the learned model to the client 100.
- the information about the learned model may include information about the weight of the learned neural network.
- the information provided to the client 100 includes information about the first model, which divides and recognizes the surgical image.
- the server 200 transmits information about the learned model to the client 100 when the operation using the client 100 is not performed, i.e., when the client 100 is not assisting the operation . That is, during the operation, the operation guide information is provided based on the information pre-stored in the client 100, and when the operation is not performed (for example, before the operation is performed) May be transmitted from the server 200 to the client 100 in advance.
- the server 200 acquires a medical image including a surgical region of the patient from the medical image capturing apparatus 10.
- the server 200 performs 3D modeling on the surgical site of the patient based on the acquired medical image (step S222).
- the 3D modeling information of the patient's surgical site is used to determine the position and shape of each organ according to the angle of the camera during surgery, and to provide information on the invisible body part by the camera.
- the server 200 transmits the 3D modeling information generated in step S222 to the client 100 (step S224).
- the client 100 stores the 3D modeling information transmitted from the server 200.
- the client 100 acquires and stores the learned model based on the information transmitted from the server 200.
- the client 100 provides the operation guide information corresponding to the operation image acquired during the operation using the model stored in the client 100 during the operation.
- the client 100 can determine the operation status by dividing and recognizing the surgical image using the stored model, and provide the surgical guide information based on the judgment.
- the server 200 transmits 3D modeling information including the surgical site of the patient in advance to the client 100 before the surgery for a specific patient begins.
- the client 100 stores the 3D modeling information transmitted from the server 200 and uses the 3D modeling information stored in the client 100 and information about the learned model during the operation, Provide surgical guide information.
- the surgical guide information may include an optimal surgical procedure corresponding to the current surgical situation.
- the surgical guide information may include information about the body part of the patient that is not identified by the camera (e.g., hidden from view by another organ).
- the surgical guide information may include information on a position of a body part that should not be damaged during surgery such as a blood vessel or a nerve.
- the surgical guide information may include information on an event occurring during surgery or a surgical error, and information on a recognized situation or information on a recognized situation.
- step S120 after the operation is completed, the client 100 transmits the learning data obtained from the operation image acquired during the operation to the server.
- the training data includes a surgical image.
- the training data includes information that divides the surgical image acquired during surgery into one or more surgical steps and recognizes one or more surgical steps that have been split.
- the client 100 segments the surgical image acquired during surgery into one or more surgical steps.
- the client 100 may segment the surgical image into one or more predetermined surgical steps using the learned model.
- the predetermined operation step divides each operation included in the operation based on the position and the change of the organ and the position and movement of the surgical tool according to a predetermined criterion, and assigns a standardized name May be granted.
- predetermined codes may be assigned to each of the divided operations.
- the client 100 grants a predetermined code for each of the one or more surgical steps that have been divided.
- the client 100 may send to the server 100 predetermined codes assigned for each of the one or more surgical steps that were split instead of the surgical image.
- the server 100 may obtain information on the sequence of each surgical step and the overall surgical operation based on the received code. In this case, the amount of data transmission is greatly reduced compared with the case of transmitting the operation image to the server 200. [ Accordingly, it is possible to transmit learning data to the server 200 in real time (step S120) even during surgery, that is, while the client 100 provides the surgical guide information (step S110), if necessary.
- the server 200 updates the learned model based on the information received from the client 100.
- the server 200 can re-learn the model using the information received from the client 100 or fine tuning the model.
- the server 200 learns a second model that computes an optimized surgical procedure based on the information received from the client 100.
- the server 200 may learn a first model that divides and recognizes a surgical image based on information received from the client 100.
- the server 200 may learn a model for recognizing the movement of the surgical tool based on the information received from the client 100.
- the server 200 may learn a first model that divides and recognizes a surgical image based on information for dividing at least one surgical image and at least one surgical image obtained from the outside.
- step S240 the server 200 transmits the updated model to the client 100 through step S230.
- the client 100 receives information on the updated model from the server 200 when the operation is not performed, i.e., information for updating the model stored in the client 100, and transmits the updated information to the client 100 based on the received information.
- Lt; RTI ID 0.0 > stored < / RTI >
- the client 100 checks the update state of the model to the server 200 (step S130).
- the client 100 compares the learned model stored in the client 100 with the learned model stored in the server 200, and the learned model stored in the server 200 is stored in the client 100 If the model is an updated model than the learned model, the client 100 receives information for updating the learned model from the server 200, and updates the learned model stored in the client 100 based on the received information do.
- FIG. 5 is a block diagram conceptually illustrating a method of scheduling learning according to an embodiment.
- the scheduler 300, the database 320, the learning unit 330, the testing unit 340, and the model 350 may each comprise a module implemented in hardware or software in the server 200, one or more instructions The data stored in the memory, or the operations performed by the server 200. [0043] FIG.
- the learning information input unit 310 means a computing device capable of directly or indirectly communicating with the server 200 or the server 200.
- the learning information input unit 310 may input information for dividing a surgical image acquired by the server 200.
- the learning information input unit 310 may be a computing device that inputs information for a doctor to view a surgical image and directly divide a surgical image into predetermined operation units.
- the source of the surgical image used for the learning is not limited.
- the server 200 may provide various user interfaces through which the learning information input unit 310 can easily input information for dividing a surgical image by using an application or a web page.
- the information that divides the surgical image and the surgical image is stored in the database 320.
- the scheduler 300 plays a role of controlling the learning process of the server 200. In the past, it was common for the developer to set the data set for learning directly and to learn the model by learning instruction.
- the scheduler 300 automatically learns a model by automatically generating a learning data set, performing learning, and testing a learning result.
- the scheduler 300 may set a learning plan and determine whether to plan the next learning based on the progress level of the learning.
- the scheduler 300 performs learning on the basis of predetermined learning data, performs a test on the learning results, and when the learning over a predetermined threshold (for example, accuracy of 80% or more) is completed, And plan the next lesson.
- a predetermined threshold for example, accuracy of 80% or more
- the scheduler 300 can plan and perform the next learning based on the learning data newly obtained when the learning on the predetermined learning data is finished.
- the scheduler 300 extracts the learning data set from the learning data stored in the database 320. [ The scheduler 300 inputs the extracted learning data set to the learning unit 330 and performs learning.
- the scheduler 300 receives a predetermined learning condition for learning, extracts a learning data set corresponding to the learning condition input from the learning data stored in the database 320, Unit 330 to perform learning.
- the learning conditions may include, but are not limited to, gender, age group, type of disease, etc.
- the scheduler 300 tests the learned model using the test unit 340 when the learning is completed. For example, the scheduler 300 generates a predetermined test data set, and performs a learning test on the learned model using the generated test data set. The scheduler 300 provides feedback according to the results of the learning test. In addition, the scheduler 300 determines whether to finish the learning and whether to plan the next learning as described above according to the learning results.
- the generation of the learned model 350 can be terminated by assuming that the test has passed.
- the next learning for the learned model 350 can be planned and executed.
- the accuracy of the learning test result is less than the predetermined reference value, it is determined that the test is not passed, and the learning is again performed through the learning unit 330, or the learning condition can be reset.
- the accuracy of the already learned model can be compared with the accuracy of the model in which the additional learning is performed using the new data. If the accuracy of the model in which the additional learning is performed is higher, the scheduler 300 can update the learned model.
- test unit 340 when learning is performed based on a predetermined learning condition, the test unit 340 generates a test data set corresponding to a predetermined learning condition, Test the model.
- the scheduler 300 when the new learning data is received in the server 200, the scheduler 300 generates a learning data set including new learning data, inputs the generated data set to the learning unit 330, .
- the scheduler 300 may perform learning each time new learning data is received, and may generate learning data sets every time the learning data is accumulated over a predetermined amount to perform learning, It is not.
- the test unit 340 can generate a test data set for testing the learned model using new learning data, and test the learned model. If the accuracy of the learned model is higher than the accuracy of the existing model using the new learning data, the scheduler 300 can update the learned model.
- the scheduler 300 may perform learning based on rules for generating a learning data set that includes new learning data, so as not to forget previously learned results.
- the learning unit 330 when the learning unit 330 performs learning using a neural network, when learning is performed using new learning data, compared to performing learning using existing learning data, Do not change.
- the feature extracted from the existing learning data using the learned model using the new learning data should be close to the feature extracted from the existing learning data using the existing learned model.
- the scheduler 300 generates a learning data set that includes new learning data and existing learning data selected in accordance with a predetermined criterion, and transmits the learning data set to the learning unit 330 using the generated learning data set Learning.
- the scheduler 300 selects the existing learning data that can prevent the model learned by the new learning data set from forgetting the previously learned information.
- the client 100 compares the learned model 350 with the model stored in the client 100, and if the learned model 350 is an updated model than the model stored in the client 100, And updates the model stored in the client (100).
- FIG. 6 is a configuration diagram of a server 200 according to an embodiment.
- the processor 202 may include one or more cores (not shown) and a connection path (e.g., a bus, etc.) to transmit and receive signals to and / or from a graphics processing unit (not shown) .
- a connection path e.g., a bus, etc.
- the processor 202 in accordance with one embodiment performs the surgical video data learning method described with respect to Figures 1-5 by executing one or more instructions stored in the memory 204.
- the processor 202 may learn one or more models that provide surgical guide information by executing one or more instructions stored in memory, send information about the learned at least one model to a client, Receiving training data obtained from the surgical image acquired during surgery from the client, updating the learned one or more models, and transmitting information about the updated at least one model to the client.
- the processor 202 may include a random access memory (RAM) (not shown) and a read-only memory (ROM) for temporarily and / or permanently storing signals (or data) , Not shown).
- the processor 202 may be implemented as a system-on-chip (SoC) including at least one of a graphics processing unit, a RAM, and a ROM.
- SoC system-on-chip
- the memory 204 may store programs (one or more instructions) for processing and control of the processor 202.
- the programs stored in the memory 204 may be divided into a plurality of modules according to functions.
- the surgical image segmentation method may be implemented as a program (or an application) to be executed in combination with a hardware computer and stored in a medium.
- the above-described program may be stored in a computer-readable medium such as C, C ++, JAVA, machine language, or the like that can be read by the processor (CPU) of the computer through the device interface of the computer, And may include a code encoded in a computer language of the computer.
- code may include a functional code related to a function or the like that defines necessary functions for executing the above methods, and includes a control code related to an execution procedure necessary for the processor of the computer to execute the functions in a predetermined procedure can do.
- code may further include memory reference related code as to whether the additional information or media needed to cause the processor of the computer to execute the functions should be referred to at any location (address) of the internal or external memory of the computer have.
- the code may be communicated to any other computer or server remotely using the communication module of the computer
- a communication-related code for determining whether to communicate, what information or media should be transmitted or received during communication, and the like.
- the medium to be stored is not a medium for storing data for a short time such as a register, a cache, a memory, etc., but means a medium that semi-permanently stores data and is capable of being read by a device.
- examples of the medium to be stored include ROM, RAM, CD-ROM, magnetic tape, floppy disk, optical data storage, and the like, but are not limited thereto.
- the program may be stored in various recording media on various servers to which the computer can access, or on various recording media on the user's computer.
- the medium may be distributed to a network-connected computer system so that computer-readable codes may be stored in a distributed manner.
- the steps of a method or algorithm described in connection with the embodiments of the present invention may be embodied directly in hardware, in software modules executed in hardware, or in a combination of both.
- the software module may be a random access memory (RAM), a read only memory (ROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, a hard disk, a removable disk, a CD- May reside in any form of computer readable recording medium known in the art to which the invention pertains.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Molecular Biology (AREA)
- Veterinary Medicine (AREA)
- Heart & Thoracic Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Primary Health Care (AREA)
- Robotics (AREA)
- Epidemiology (AREA)
- General Business, Economics & Management (AREA)
- Business, Economics & Management (AREA)
- Pathology (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Radiology & Medical Imaging (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (14)
- 수술 가이드 정보를 제공하는 하나 이상의 모델을 학습시키는 서버; 및상기 서버에서 학습된 적어도 하나의 모델을 저장하고, 상기 저장된 모델을 이용하여 수술 중에 획득되는 수술 영상을 인식하고, 인식된 수술 영상에 대응하는 수술 가이드 정보를 제공하는 클라이언트; 를 포함하고,상기 클라이언트는,상기 수술 중에 획득된 수술 영상으로부터 획득되는 제1 학습용 데이터를 상기 서버에 전송하고, 상기 서버로부터 상기 저장된 모델을 업데이트하기위한 정보를 수신하고, 상기 수신된 정보에 기초하여 상기 저장된 모델을 업데이트하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 하나 이상의 모델은,수술 영상을 분할(segmentation)하여 인식하는 제1 모델; 및최적화된 수술방법을 산출하는 제2 모델; 을 포함하고,상기 제1 학습용 데이터는 상기 제1 모델 및 상기 제2 모델 중 적어도 하나를 학습시키는 데 이용되고,상기 클라이언트에 저장되는 적어도 하나의 모델은 상기 제1 모델을 포함하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 클라이언트는,상기 수술 중에 획득된 수술 영상을 하나 이상의 수술단계로 분할(segmentation)하고,상기 분할된 하나 이상의 수술단계를 인식하고,상기 인식 결과를 상기 제1 학습용 데이터로서 상기 서버에 전송하는, 수술영상데이터 학습시스템.
- 제3 항에 있어서,상기 클라이언트는,상기 분할된 하나 이상의 수술단계 각각에 대하여 기 설정된 코드를 부여하고,상기 분할된 하나 이상의 수술단계 각각에 대하여 부여된 상기 기 설정된 코드를 상기 서버에 전송하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 클라이언트는,상기 클라이언트에 저장된 모델과 상기 서버에 저장된 모델을 비교하고, 상기 서버에 저장된 모델이 상기 클라이언트에 저장된 모델보다 업데이트된 모델인 경우, 상기 서버로부터 상기 클라이언트에 저장된 모델을 업데이트하기위한 정보를 수신하고, 상기 수신된 정보에 기초하여 상기 클라이언트에 저장된 모델을 업데이트하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 서버는,하나 이상의 수술 영상 및 상기 하나 이상의 수술 영상을 분할하는 정보를 획득하고,상기 획득된 정보를 이용하여 제2 학습용 데이터를 생성하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 서버는,학습조건을 입력받고, 상기 제2 학습용 데이터로부터 상기 입력된 학습조건에 대응하는 데이터셋을 추출하고, 상기 추출된 데이터셋을 이용하여 상기 적어도 하나의 모델에 대한 학습을 수행하는, 수술영상데이터 학습시스템.
- 제7 항에 있어서,상기 서버는,상기 입력된 학습조건에 대응하는 테스트용 데이터셋을 생성하고, 상기 생성된 테스트용 데이터셋을 이용하여 상기 적어도 하나의 모델에 대한 학습 테스트를 수행하고, 상기 학습 테스트 수행결과에 따른 피드백을 제공하는, 수술영상데이터 학습시스템.
- 제1 항에 있어서,상기 서버는,하나 이상의 새로운 제2 학습용 데이터를 수신하고, 적어도 하나의 상기 새로운 제2 학습용 데이터를 포함하는 데이터셋을 생성하고, 상기 생성된 데이터셋을 이용하여 상기 적어도 하나의 모델에 대한 학습을 수행하는, 수술영상데이터 학습시스템.
- 제9 항에 있어서,상기 서버는,테스트용 데이터셋을 획득하고, 상기 테스트용 데이터셋을 이용하여 상기 생성된 데이터셋을 이용하여 학습된 상기 적어도 하나의 모델에 대한 학습 테스트를 수행하고, 상기 학습 테스트 수행결과에 따른 피드백을 제공하는, 수술영상데이터 학습시스템.
- 제9 항에 있어서,상기 서버는,상기 새로운 제2 학습용 데이터 및 소정의 기준에 따라 선정된 기존 제2 학습용 데이터를 포함하는 상기 데이터셋을 생성하고, 상기 생성된 데이터셋을 이용하여 상기 적어도 하나의 모델에 대한 학습을 수행하는, 수술영상데이터 학습시스템.
- 서버가 수술 가이드 정보를 제공하는 하나 이상의 모델을 학습시키는 단계;클라이언트에 상기 학습된 적어도 하나의 모델에 대한 정보를 전송하는 단계;상기 클라이언트가 수술 중에 획득한 수술 영상으로부터 획득되는 학습용 데이터를 상기 클라이언트로부터 수신하는 단계;상기 학습된 하나 이상의 모델을 업데이트하는 단계; 및상기 클라이언트에 업데이트된 상기 적어도 하나의 모델에 대한 정보를 전송하는 단계; 를 포함하는, 수술영상데이터 학습방법.
- 하나 이상의 인스트럭션을 저장하는 메모리; 및상기 메모리에 저장된 상기 하나 이상의 인스트럭션을 실행하는 프로세서를 포함하고,상기 프로세서는 상기 하나 이상의 인스트럭션을 실행함으로써,수술 가이드 정보를 제공하는 하나 이상의 모델을 학습시키는 단계;클라이언트에 상기 학습된 적어도 하나의 모델에 대한 정보를 전송하는 단계;상기 클라이언트가 수술 중에 획득한 수술 영상으로부터 획득되는 학습용 데이터를 상기 클라이언트로부터 수신하는 단계;상기 학습된 하나 이상의 모델을 업데이트하는 단계; 및상기 클라이언트에 업데이트된 상기 적어도 하나의 모델에 대한 정보를 전송하는 단계; 를 수행하는, 서버.
- 하드웨어인 컴퓨터와 결합되어, 제12 항의 방법을 수행할 수 있도록 컴퓨터에서 독출가능한 기록매체에 저장된 컴퓨터프로그램.
Applications Claiming Priority (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20170182900 | 2017-12-28 | ||
KR10-2017-0182899 | 2017-12-28 | ||
KR20170182898 | 2017-12-28 | ||
KR20170182899 | 2017-12-28 | ||
KR10-2017-0182898 | 2017-12-28 | ||
KR10-2017-0182900 | 2017-12-28 | ||
KR1020180013580 | 2018-02-02 | ||
KR10-2018-0013580 | 2018-02-02 | ||
KR10-2018-0026572 | 2018-03-06 | ||
KR1020180026572A KR101864380B1 (ko) | 2017-12-28 | 2018-03-06 | 수술영상데이터 학습시스템 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019132168A1 true WO2019132168A1 (ko) | 2019-07-04 |
Family
ID=62628386
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2018/010335 WO2019132170A1 (ko) | 2017-12-28 | 2018-09-05 | 학습용 데이터 관리방법, 장치 및 프로그램 |
PCT/KR2018/010333 WO2019132168A1 (ko) | 2017-12-28 | 2018-09-05 | 수술영상데이터 학습시스템 |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2018/010335 WO2019132170A1 (ko) | 2017-12-28 | 2018-09-05 | 학습용 데이터 관리방법, 장치 및 프로그램 |
Country Status (2)
Country | Link |
---|---|
KR (3) | KR101864380B1 (ko) |
WO (2) | WO2019132170A1 (ko) |
Families Citing this family (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2022502209A (ja) * | 2018-10-11 | 2022-01-11 | ジェイエルケー インコーポレイテッド | がん領域に対するディープラーニングモデル学習装置及び方法 |
KR102243644B1 (ko) * | 2018-12-07 | 2021-04-23 | 서울대학교 산학협력단 | 의료 영상 분획 딥러닝 모델 생성 장치 및 방법과, 그에 따라 생성된 의료 영상 분획 딥러닝 모델 |
KR102237009B1 (ko) * | 2018-12-11 | 2021-04-08 | 시너지에이아이 주식회사 | 다낭 신 또는 다낭 간의 진행 상태를 판단하는 방법 및 이를 수행하기 위한 의료용 전자 장치 |
KR102189761B1 (ko) * | 2018-12-21 | 2020-12-11 | 주식회사 엘지씨엔에스 | 딥러닝 학습 방법 및 서버 |
KR102186632B1 (ko) * | 2019-01-07 | 2020-12-02 | 재단법인대구경북과학기술원 | 의료 영상의 분석 모델을 학습시키는 학습 장치 및 그 학습 방법 |
WO2020159276A1 (ko) * | 2019-02-01 | 2020-08-06 | 주식회사 아이버티 | 수술 분석 장치, 수술영상 분석 및 인식 시스템, 방법 및 프로그램 |
US10791301B1 (en) | 2019-06-13 | 2020-09-29 | Verb Surgical Inc. | Method and system for synchronizing procedure videos for comparative learning |
US20220222807A1 (en) * | 2019-08-19 | 2022-07-14 | Lg Electronics Inc. | Ai-based new learning model generation system for vision inspection on product production line |
KR20210026623A (ko) * | 2019-08-30 | 2021-03-10 | 삼성전자주식회사 | 인공지능 모델을 학습시키는 시스템 및 방법 |
KR102542037B1 (ko) * | 2019-09-26 | 2023-06-12 | 주식회사 루닛 | 인공지능 모델을 사용 기관에 특화시키는 학습 방법, 이를 수행하는 장치 |
KR102119056B1 (ko) * | 2019-10-08 | 2020-06-05 | (주)제이엘케이 | 생성적 적대 신경망 기반의 의료영상 학습 방법 및 장치 |
KR102168558B1 (ko) * | 2019-10-24 | 2020-10-21 | 서울대학교산학협력단 | 액티브 러닝을 위한 학습용 데이터 선정 방법, 액티브 러닝을 위한 학습용 데이터 선정 장치 및 액티브 러닝을 이용한 영상 분석 방법 |
KR102334923B1 (ko) * | 2019-11-06 | 2021-12-03 | 주식회사 에이젠글로벌 | 인공지능을 이용한 대출 확대 가설 검정 시스템 및 이를 이용한 방법 |
KR102328154B1 (ko) * | 2019-11-11 | 2021-11-18 | 주식회사 테서 | 머신러닝에 기반한 의료데이터 수집 및 분석 서비스 제공 방법 및 시스템 |
CN113179659A (zh) * | 2019-11-27 | 2021-07-27 | 谷歌有限责任公司 | 利用封闭数据的个性化数据模型 |
KR102272573B1 (ko) * | 2019-11-28 | 2021-07-05 | 한국전자기술연구원 | 에너지 사용량 데이터의 비지도 학습 기반 부하 모니터링 방법 |
KR102356263B1 (ko) * | 2020-02-12 | 2022-01-28 | 한전케이디엔주식회사 | 전력설비 관리 시스템 및 방법 |
KR102296274B1 (ko) * | 2020-10-26 | 2021-09-01 | 주식회사 보고넷 | 사용자 학습 기반 딥러닝 객체 인식 서비스 제공 방법 |
KR102273494B1 (ko) * | 2020-11-05 | 2021-07-06 | 주식회사 카비랩 | 골절 수술용 의료기기 자동 디자인 생성 시스템 |
KR102287762B1 (ko) * | 2020-11-05 | 2021-08-10 | 주식회사 카비랩 | 앙상블 학습 기반 골절 수술 시뮬레이터 시스템 및 그 구동 방법 |
KR20220067732A (ko) * | 2020-11-18 | 2022-05-25 | 한국전자기술연구원 | 재학습 가능한 모바일 딥러닝 하드웨어 장치 |
KR102467047B1 (ko) * | 2020-11-18 | 2022-11-15 | (주)휴톰 | 어노테이션 평가 방법 및 장치 |
KR102336500B1 (ko) * | 2021-03-02 | 2021-12-07 | 국방과학연구소 | 의사-레이블 데이터 생성 방법 및 장치, 의사-레이블 데이터를 이용한 심층 신경망 학습 방법 및 장치 |
KR102515081B1 (ko) * | 2022-03-22 | 2023-03-29 | 한화시스템 주식회사 | 객체인식 소프트웨어 학습방법 및 학습장치 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008046729A (ja) * | 2006-08-11 | 2008-02-28 | Kddi Corp | 動画像話題分割装置 |
KR20100098055A (ko) * | 2009-02-27 | 2010-09-06 | 한국과학기술원 | 영상유도수술시스템 및 그 제어방법 |
KR101302595B1 (ko) * | 2012-07-03 | 2013-08-30 | 한국과학기술연구원 | 수술 진행 단계를 추정하는 시스템 및 방법 |
KR20160138174A (ko) * | 2014-03-21 | 2016-12-02 | 바이오레이즈, 인크. | 치과 레이저 인터페이스 시스템 및 방법 |
KR20170061016A (ko) * | 2015-11-25 | 2017-06-02 | 삼성전자주식회사 | 데이터 인식 모델 구축 장치 및 방법과 데이터 인식 장치 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3329202B2 (ja) * | 1996-08-07 | 2002-09-30 | ケイディーディーアイ株式会社 | ニューラルネットワーク学習方式 |
EP1356421B1 (en) * | 2001-01-23 | 2009-12-02 | Health Discovery Corporation | Computer-aided image analysis |
JP2009075737A (ja) * | 2007-09-19 | 2009-04-09 | Nec Corp | 半教師あり学習方法、半教師あり学習装置及び半教師あり学習プログラム |
KR100918361B1 (ko) * | 2008-02-26 | 2009-09-22 | 엔에이치엔(주) | 고속화 검색 모델링 시스템 및 방법 |
JP2009237923A (ja) * | 2008-03-27 | 2009-10-15 | Nec Corp | 学習方法およびシステム |
JP2010092266A (ja) * | 2008-10-08 | 2010-04-22 | Nec Corp | 学習装置、学習方法及びプログラム |
JP5881048B2 (ja) * | 2012-09-18 | 2016-03-09 | 株式会社日立製作所 | 情報処理システム、及び、情報処理方法 |
US9324022B2 (en) * | 2014-03-04 | 2016-04-26 | Signal/Sense, Inc. | Classifying data with deep learning neural records incrementally refined through expert input |
KR102239714B1 (ko) * | 2014-07-24 | 2021-04-13 | 삼성전자주식회사 | 신경망 학습 방법 및 장치, 데이터 처리 장치 |
US10223635B2 (en) * | 2015-01-22 | 2019-03-05 | Qualcomm Incorporated | Model compression and fine-tuning |
US20160358099A1 (en) * | 2015-06-04 | 2016-12-08 | The Boeing Company | Advanced analytical infrastructure for machine learning |
JP6182242B1 (ja) * | 2016-06-13 | 2017-08-16 | 三菱電機インフォメーションシステムズ株式会社 | データのラベリングモデルに係る機械学習方法、コンピュータおよびプログラム |
-
2018
- 2018-03-06 KR KR1020180026572A patent/KR101864380B1/ko active IP Right Grant
- 2018-03-06 KR KR1020180026575A patent/KR101864412B1/ko active IP Right Grant
- 2018-05-29 KR KR1020180061323A patent/KR20190088376A/ko not_active Application Discontinuation
- 2018-09-05 WO PCT/KR2018/010335 patent/WO2019132170A1/ko active Application Filing
- 2018-09-05 WO PCT/KR2018/010333 patent/WO2019132168A1/ko active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008046729A (ja) * | 2006-08-11 | 2008-02-28 | Kddi Corp | 動画像話題分割装置 |
KR20100098055A (ko) * | 2009-02-27 | 2010-09-06 | 한국과학기술원 | 영상유도수술시스템 및 그 제어방법 |
KR101302595B1 (ko) * | 2012-07-03 | 2013-08-30 | 한국과학기술연구원 | 수술 진행 단계를 추정하는 시스템 및 방법 |
KR20160138174A (ko) * | 2014-03-21 | 2016-12-02 | 바이오레이즈, 인크. | 치과 레이저 인터페이스 시스템 및 방법 |
KR20170061016A (ko) * | 2015-11-25 | 2017-06-02 | 삼성전자주식회사 | 데이터 인식 모델 구축 장치 및 방법과 데이터 인식 장치 |
Also Published As
Publication number | Publication date |
---|---|
KR101864412B1 (ko) | 2018-06-04 |
KR20190088376A (ko) | 2019-07-26 |
KR101864380B1 (ko) | 2018-06-04 |
WO2019132170A1 (ko) | 2019-07-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019132168A1 (ko) | 수술영상데이터 학습시스템 | |
WO2019132169A1 (ko) | 수술영상 재생제어 방법, 장치 및 프로그램 | |
KR102014385B1 (ko) | 수술영상 학습 및 학습 기반 수술동작 인식 방법 및 장치 | |
WO2020251217A1 (ko) | 사용자 맞춤형 피부 진단 시스템 및 방법 | |
WO2021045367A1 (ko) | 상담 대상자의 드로잉 과정을 통해 심리 상태를 판단하는 방법 및 컴퓨터 프로그램 | |
WO2017051945A1 (ko) | 질환 모델 기반의 의료 정보 서비스 제공 방법 및 장치 | |
WO2016171341A1 (ko) | 클라우드 기반 병리 분석 시스템 및 방법 | |
WO2019235828A1 (ko) | 투 페이스 질병 진단 시스템 및 그 방법 | |
WO2019132165A1 (ko) | 수술결과에 대한 피드백 제공방법 및 프로그램 | |
WO2020111754A9 (ko) | 세미 슈퍼바이즈드 학습을 이용한 진단 시스템 제공방법 및 이를 이용하는 진단 시스템 | |
WO2020032562A2 (ko) | 생체 이미지 진단 시스템, 생체 이미지 진단 방법, 및 이를 수행하기 위한 단말 | |
WO2013105815A1 (en) | Fetus modeling method and image processing apparatus therefor | |
WO2022131642A1 (ko) | 의료 영상 기반 질환 중증도 결정 장치 및 방법 | |
WO2021006472A1 (ko) | 임플란트 시술 계획 수립을 위한 다중 골밀도 표시방법 및 그 영상 처리장치 | |
WO2021034138A1 (ko) | 치매 평가 방법 및 이를 이용한 장치 | |
WO2024111914A1 (ko) | 범용성이 향상된 인공지능에 의한 의료영상 변환방법 및 그 장치 | |
WO2019221586A1 (ko) | 의료영상 관리 시스템, 방법 및 컴퓨터 판독 가능한 기록매체 | |
WO2021206518A1 (ko) | 수술 후 수술과정 분석 방법 및 시스템 | |
WO2019164273A1 (ko) | 수술영상을 기초로 수술시간을 예측하는 방법 및 장치 | |
WO2019240330A1 (ko) | 영상 기반 힘 예측 시스템 및 그 방법 | |
WO2020141751A1 (ko) | 신체 사이즈 측정용 사진 획득 방법 및 이를 이용한 신체 사이즈 측정 방법, 서버 및 프로그램 | |
WO2023282454A1 (ko) | Ai 학습을 위한 임플란트 클래스 분류 방법 | |
WO2022181919A1 (ko) | 가상 현실 기반의 수술 환경을 제공하는 장치 및 방법 | |
WO2017010612A1 (ko) | 의료 영상 분석 기반의 병리 진단 예측 시스템 및 방법 | |
WO2021015490A2 (ko) | 이미지의 특이 영역 분석 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18893411 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18893411 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 260121) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18893411 Country of ref document: EP Kind code of ref document: A1 |