WO2021054700A1 - Method for providing tooth lesion information, and device using same - Google Patents

Method for providing tooth lesion information, and device using same Download PDF

Info

Publication number
WO2021054700A1
WO2021054700A1 PCT/KR2020/012448 KR2020012448W WO2021054700A1 WO 2021054700 A1 WO2021054700 A1 WO 2021054700A1 KR 2020012448 W KR2020012448 W KR 2020012448W WO 2021054700 A1 WO2021054700 A1 WO 2021054700A1
Authority
WO
WIPO (PCT)
Prior art keywords
tooth
lesion
information
image
providing
Prior art date
Application number
PCT/KR2020/012448
Other languages
French (fr)
Korean (ko)
Inventor
김재영
이홍석
Original Assignee
주식회사 뷰노
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 뷰노 filed Critical 주식회사 뷰노
Publication of WO2021054700A1 publication Critical patent/WO2021054700A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7271Specific aspects of physiological measurement analysis
    • A61B5/7275Determining trends in physiological measurement data; Predicting development of a medical condition based on physiological measurements, e.g. determining a risk factor
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/05Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves 
    • A61B5/055Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves  involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/02Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
    • A61B6/03Computed tomography [CT]
    • A61B6/032Transmission computed tomography [CT]
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems

Definitions

  • the present invention relates to a method for providing tooth lesion information and an apparatus using the same.
  • Various X-ray images taken in various ways are widely used for dental diagnosis and treatment.
  • Panorama X-ray images are widely used in the establishment of treatment plans in that the oral structure can be confirmed as a whole and the anatomical structure of the oral cavity is properly revealed.
  • a method of predicting a lesion area by identifying an X-ray panoramic image with the naked eye is widely used in dental treatment, but techniques for automatically identifying a lesion have been developed to overcome the limitation of identification by the naked eye.
  • Research to establish a lesion candidate region in an X-ray image remains at the level of using the contrast change in an X-ray image.
  • An object of the present invention is to provide a means for effectively determining a lesion area in a tooth image by providing information on whether or not a lesion of each tooth occurs together with identification information of the corresponding tooth.
  • the present invention is to provide the visualized lesion information through a map that visualizes the possibility that each region corresponds to the lesion region in the input tooth image.
  • an object of the present invention is to provide a readout model capable of more accurately detecting a lesion area by using a loss function in consideration of a correlation between teeth in a process of learning a readout model for detecting a lesion area in a tooth image.
  • a characteristic configuration of the present invention for achieving the object of the present invention as described above and realizing the characteristic effects of the present invention described later is as follows.
  • a method for providing tooth lesion information performed by a computing device includes the steps of: (a) receiving a tooth image of a subject; (b) detecting a lesion area included in the tooth image through a pre-learned read model or supporting another device linked to the computing device to detect the lesion area; (c) generating tooth lesion information on the lesion area or supporting another device linked to the computing device to generate the tooth lesion information; And (d) providing lesion information to an external entity.
  • the step (c) includes the steps of (c1) generating a map for visualizing a lesion area in the tooth image; And (c2) generating the tooth lesion information through visualization information in which the lesion area is visualized in the tooth image based on the map.
  • the step (c) includes the steps of (c3) generating identification information for identifying the position or order of individual teeth included in the tooth image; (c4) generating matching information by matching the identification information with information on whether a lesion has occurred in an individual tooth corresponding to the identification information; And (c5) generating the tooth lesion information based on the matching information.
  • the readout model is pre-trained in a direction that minimizes the result of the loss function, and the loss function may be determined based on correlation information between individual teeth.
  • a computing device that provides tooth lesion information includes: a communication unit that receives a tooth image of a subject; And a processor for generating tooth lesion information for the tooth image, wherein the processor generates tooth lesion information on a lesion region included in the tooth image through a pre-learned read model, or other information linked to the computing device. It is possible to support the device to generate the lesion information and provide the tooth lesion information to an external entity.
  • a single tooth image may provide a means for reading whether or not lesions have occurred in individual teeth.
  • a medical image used in a conventional hospital can be used as it is, so it goes without saying that the method of the present invention is not dependent on a specific type of image or platform.
  • FIG. 1 is a conceptual diagram schematically showing an exemplary configuration of a computing device that performs a method for providing lesion information according to the present invention.
  • FIG. 2 is an exemplary block diagram showing hardware or software components of a computing device that performs a method for providing lesion information according to the present invention.
  • FIG. 3 is a flowchart illustrating a method of providing tooth lesion information according to an exemplary embodiment.
  • FIG. 4 is a diagram illustrating an example in which a method of providing lesion information according to an exemplary embodiment is performed.
  • FIG. 5 is a diagram illustrating an example in which a Pearson correlation matrix used in a process of determining a loss function is visualized.
  • image refers to multidimensional data composed of discrete image elements (eg, pixels in a 2D image and voxels in a 3D image).
  • image refers to an X-ray image, (cone-beam) computed tomography, magnetic resonance imaging (MRI), ultrasound, or any other medical image known in the art. It may be a medical image of a subject, that is, a subject collected by the system. Also, an image may be provided in a non-medical context, such as a remote sensing system, an electron microscopy, and the like.
  • image' refers to an image that can be seen by the eye (e.g., displayed on a video screen) or of an image (e.g., a file corresponding to a pixel output of a CT, MRI detector, etc.) It is a term referring to digital representations.
  • the'DICOM Digital Imaging and Communications in Medicine; Medical Digital Imaging and Communications
  • ACR American Radiological Society
  • NEMA American Electrical Industry Association
  • PES picture archiving and communication system
  • DICOM digital medical imaging equipment
  • 'learning' or'learning' is a term that refers to performing machine learning through computing according to procedures. It will be appreciated by those of skill in the art that it is not intended to be referred to.
  • the present invention covers all possible combinations of the embodiments indicated herein. It should be understood that the various embodiments of the present invention are different from each other, but need not be mutually exclusive. For example, specific shapes, structures, and characteristics described herein may be implemented in other embodiments without departing from the spirit and scope of the present invention in relation to one embodiment. In addition, it should be understood that the location or arrangement of individual components in each disclosed embodiment may be changed without departing from the spirit and scope of the present invention. Accordingly, the detailed description to be described below is not intended to be taken in a limiting sense, and the scope of the present invention, if appropriately described, is limited only by the appended claims, along with all ranges equivalent to those claimed by the claims. Like reference numerals in the drawings refer to the same or similar functions over several aspects.
  • FIG. 1 is a conceptual diagram schematically showing an exemplary configuration of a computing device that performs a method for providing lesion information according to the present invention.
  • a computing device 100 includes a communication unit 110 and a processor 120, and is directly or indirectly connected to an external computing device (not shown) through the communication unit 110. Can communicate with enemies.
  • the computing device 100 is a device that may include typical computer hardware (eg, computer processor, memory, storage, input devices and output devices, and other components of existing computing devices; routers, switches, etc.).
  • Electronic communication devices eg, electronic communication devices; electronic information storage systems such as network-attached storage (NAS) and storage area networks (SANs) and computer software (i.e., allowing the computing device to function in a specific way). Instructions) may be used to achieve the desired system performance.
  • NAS network-attached storage
  • SANs storage area networks
  • Instructions may be used to achieve the desired system performance.
  • the communication unit 110 of such a computing device can transmit and receive requests and responses to and from other computing devices to which it is linked.
  • requests and responses may be made by the same transmission control protocol (TCP) session.
  • TCP transmission control protocol
  • the present invention is not limited thereto, and may be transmitted/received as, for example, a user datagram protocol (UDP) datagram.
  • the communication unit 110 may include a keyboard, a mouse, other external input devices, printers, displays, and other external output devices for receiving commands or instructions.
  • the processor 120 of the computing device may include a micro processing unit (MPU), a central processing unit (CPU), a graphics processing unit (GPU) or a tensile processing unit (TPU), a cache memory, and a data bus. ), and the like.
  • MPU micro processing unit
  • CPU central processing unit
  • GPU graphics processing unit
  • TPU tensile processing unit
  • cache memory and a data bus.
  • data bus a data bus.
  • it may further include an operating system and a software configuration of an application that performs a specific purpose.
  • the image of the present invention targets a tooth image
  • the tooth image targets a tooth X-ray panoramic image as an example, but the scope of the present invention is not limited thereto, and all of the general form tooth image It will be readily understood by those of ordinary skill in the art that it can be applied and that it can be applied in the process of detecting an arbitrary lesion area included in a target image.
  • FIG. 2 is an exemplary block diagram showing hardware or software components of a computing device that performs a method for providing lesion information according to the present invention.
  • the individual modules shown in FIG. 2 may be implemented by interlocking the communication unit 110 or the processor 120 included in the computing device 100, or the communication unit 110 and the processor 120, for example. Technicians will be able to understand.
  • the computing device 100 may include an image acquisition module 210 as its constituent element.
  • the image acquisition module 210 may acquire a tooth image of a subject, which is previously stored in a database or obtained from a device dedicated to photographing an image.
  • the tooth image may be an X-ray panoramic image of a subject's teeth captured through a dedicated photographing device built into the computing device 100 or an external photographing dedicated device.
  • the tooth image acquired through the image acquisition module 210 may be transmitted to the tooth lesion region detection module 220.
  • the tooth lesion region detection module 220 may detect a lesion region included in the tooth image through a read model.
  • the readout model may be an artificial neural network that has been trained in advance to detect a lesion area in a tooth image.
  • the readout model included in the tooth lesion region detection module 220 is a deep learning-based artificial neural network that has been trained to output at least one of the location of the lesion region, the shape of the lesion region, and the probability corresponding to the lesion region from the input tooth image. I can.
  • the artificial neural network included in the tooth lesion region detection module 220 determines whether the region included in the input tooth image matches the detection target region (eg, bone loss region) based on the similarity, and outputs the determination result.
  • the detection target region eg, bone loss region
  • DNN Deep Neural Network
  • CNN Convolutional Neural Network
  • DCNN Deep Convolution Neural Network
  • RNN Recurrent Neural Network
  • RBM Restricted Boltzmann Machine
  • DBN Deep Belief Network
  • SSD Single Shot Detector
  • YOLO You Only Look Once
  • the type of artificial neural network that can be used in the tooth lesion region detection module 220 is not limited to the presented example, and may include any artificial neural network that can be learned to recognize the lesion region in the tooth image based on the labeled learning data. It will be understood by those of ordinary skill in the art that it can be.
  • the detection result of the lesion region performed through the tooth lesion region detection module 220 may be transmitted to the visualization module 230, and the visualization module 230 may generate a map for visualizing the lesion region in the tooth image.
  • the visualization module 230 may generate a heat map in which the color of a corresponding region is changed according to a probability that each region of a tooth corresponds to a lesion region (eg, a bone loss region). .
  • the method of generating the map by the visualization module 230 is not limited to the method described above, and an arbitrary method capable of visually expressing the probability that each region of the tooth image corresponds to the lesion region (e.g., It will be apparent to those of ordinary skill in the art that the corresponding probability may be expressed in a numerical manner, or the contrast may be adjusted according to the probability corresponding to the lesion area).
  • the detection result performed through the tooth lesion region detection module 220 may be transmitted to the classification module 240, and the classification module 240 may determine identification information of a tooth corresponding to the detected lesion region.
  • the classification module 240 may generate identification information for identifying the position or order of individual teeth included in the tooth image, and may determine identification information of the tooth in which the lesion area is detected based on the generated identification information and the detection result. .
  • the identification information for individual teeth may be determined based on the FDI World Dental Federation notation, which is a dental identification number most used by dentists around the world, but is not limited thereto, and is based on an arbitrary method by which individual teeth can be identified. Can be determined.
  • the classification module 240 determines the identification number of the tooth from which the lesion is detected among identification numbers generated for individual teeth, and based on the determination result, the identification number and It is possible to generate matching information by matching information on whether a lesion has occurred in an individual tooth.
  • Matching information is a method of providing a list of identification information of teeth determined to have lesions according to implementation, a method of matching identification information of all teeth and whether lesions have occurred, etc. It can be created in the way of
  • the tooth lesion information generation and transmission module 250 may generate tooth lesion information based on a map generated by the visualization module 230 or matching information generated through the classification module 240.
  • the tooth lesion information may be determined based on at least one of visualization information generated by applying the map to a tooth image or information on a tooth list in which a lesion is detected based on the matching information.
  • the tooth lesion information generated through the tooth lesion information generation and transmission module 250 may be stored in a database or may be provided to an external entity.
  • the tooth lesion information generation and transmission module 250 may provide an output image reflecting the tooth lesion information to the external entity using a predetermined display device or the like, or through an provided communication unit.
  • the external entity includes a user of the computing device 100, an administrator, a medical professional in charge of the subject, etc., but in addition to this, a subject who needs tooth lesion information for the input tooth image If so, it should be understood that any subject is included.
  • the external entity may be an external AI device including a separate artificial intelligence (AI) hardware and/or software module that utilizes the tooth lesion information.
  • AI artificial intelligence
  • tooth lesion information which is the result of the hardware and/or software module performing the procedure, is used to suggest that it can be used as input data for other methods. That is, the external entity may be the computing device 100 itself.
  • FIG. 2 Although the components shown in FIG. 2 are illustrated as being realized in one computing device for convenience of description, it will be understood that a plurality of computing devices 100 performing the method of the present invention may be configured to interlock with each other.
  • FIG. 3 is a flowchart illustrating a method of providing tooth lesion information according to an exemplary embodiment.
  • the computing device may receive a tooth image of a subject through a communication unit in step S100.
  • the tooth image may be a tooth X-ray panoramic image, but the present invention is not limited thereto, and may include an image of an arbitrary shape photographing the subject's teeth.
  • the computing device may detect a lesion area included in the tooth image through the reading model in step S200.
  • the readout model may include a pre-trained artificial neural network to detect a lesion area in a tooth image.
  • the loss function for training the readout model may be determined based on a maxillary loss function for a maxillary tooth and a mandibular loss function for a mandibular tooth in addition to a cross entropy function commonly used as a loss function.
  • the upper jaw may mean an upper jaw included in the tooth image
  • the mandible may mean a lower jaw included in the tooth image.
  • the maxillary loss function may mean a loss function determined corresponding to the upper jaw
  • the mandibular loss function may mean a loss function corresponding to the mandible.
  • individual loss functions for calculating the maxillary loss function and the mandibular loss function may be determined based on Equation 1.
  • Is the individual loss function Is a value indicating whether a tooth lesion has occurred, Is a value indicating whether a lesion of the j-th tooth of the i-th data sample has occurred, Is the predicted value of the readout model for the occurrence of a tooth lesion, Is the predicted value of the readout model for the occurrence of lesion of the j-th tooth in the i-th data sample, Is the pearson correlation matrix, Is each element of the Pearson correlation matrix and is the correlation coefficient value corresponding to the j-th tooth and the j'-th tooth, Is the correlation parameter determined based on the Pearson correlation matrix, Is the correlation parameter calculated for the j-th tooth of the i-th data sample, k is the number of teeth number, Is the number of data samples, Is Represents a value for normalizing between 0 and 1.
  • the correlation coefficient between the a-th tooth and the b-th tooth included in the Pearson correlation matrix of Equation 1 May be determined based on Equation 2 below.
  • Is the correlation coefficient between the a-th tooth and the b-th tooth Is a value indicating whether a lesion of the a-th tooth has occurred in the i-th training data
  • Equation 1 the maxillary loss function and the mandibular loss function may be calculated as shown in Equations 3 and 4.
  • Equation 5 the final loss function used for learning the readout model may be calculated as in Equation 5.
  • Is the cross entropy function May mean an adjustable parameter.
  • the final loss function is determined based on the maxillary loss function and the mandibular loss function, and the maxillary loss function and the mandibular loss function are determined based on the correlation between the teeth included in each. And It is determined on the basis of. In fact, since dental lesions often occur together between adjacent teeth, the correlation between adjacent teeth may be very high in relation to the occurrence of the lesion. Since the final loss function reflects the correlation between the teeth included in the maxilla and mandible, the accuracy of the prediction of the lesion area of the read model learned based on this is improved compared to the case of using the loss function that does not consider the correlation. Can be.
  • the computing device may generate tooth lesion information on the detected lesion area in step S300 or support another device interlocked with the computing device to generate tooth lesion information.
  • the tooth lesion information may be determined by at least one of visualization information in which a lesion area is visualized in a tooth image, and matching information in which identification information of individual teeth and whether or not a lesion occurs.
  • the computing device may generate a map for visualizing the lesion area detected in the tooth image based on the detection result in step S200.
  • the computing device may determine a weight for each region of the tooth image based on the output of the last layer of the read model, and generate a map representing different visual elements according to the determined weight. For example, the computing device represents an area corresponding to a high weight in the output of the final layer, and an area with a high probability of corresponding to a lesion area is expressed in a color close to red, and an area with a low probability of corresponding to a lesion area due to a low weight is expressed. It is possible to create a heat map that is represented by an area close to green.
  • the computing device applies the generated heat map to the input tooth image and visualizes the region with a high probability of corresponding to the lesion region with a color close to red, and visualizes the region with a low probability of corresponding to the lesion region with a color close to green.
  • Visualization information for a tooth image can be generated.
  • the map for generating the visualization information is not limited to the heat map and color configuration as an example, and it can be easily understood by those of ordinary skill in the art that each region can be implemented in an arbitrary manner to visualize the probability that it corresponds to the lesion region. There will be.
  • the computing device visualizes the area of the input tooth image with a high probability of corresponding to the lesion area with dark contrast, and the area with the low probability of corresponding to the lesion area with light intensity to visualize the lesion area of the input tooth image. can do.
  • the computing device generates identification information that identifies the position or sequence of individual teeth included in the tooth image, and information on whether a lesion has occurred in the individual tooth corresponding to the generated identification information and the identification information. It is possible to generate matching information that matches.
  • the identification information may be an identification number for an individual tooth according to the FDI World Dental Federation notation, and the computing device provides information on whether a lesion has occurred in the tooth based on the detection result in step S200. Matching information matching the identification number of can be generated.
  • the computing device may provide tooth lesion information to an external entity in step S400.
  • FIG. 4 is a diagram illustrating an example in which a method of providing lesion information according to an exemplary embodiment is performed.
  • the computing device 420 may provide tooth lesion information including visualization information 430 and matching information 440 of the input tooth image 410. To this end, the computing device 420 may detect a lesion area in the tooth image 410 through a pre-learned reading model. The computing device generates the visualization information 430 to which the heat maps 431, 432, 433 that visualize the detected lesion area is applied, or the matching information 440 that matches the identification information of the tooth and the information on whether the lesion has occurred. It is possible to provide tooth lesion information by creating and providing it to an external entity. Each of the heat maps 431, 432, and 433 may be expressed in different colors or different shades according to a probability corresponding to the lesion area. The identification number of the tooth included in the matching information 440 may be determined according to the FDI World Dental Federation notation shown in the identification information 441.
  • FIG. 5 is a diagram illustrating an example in which a Pearson correlation matrix is visualized according to an embodiment.
  • the present invention can generate a lesion prediction model with higher accuracy by using the final loss function corresponding to Equation 5 in which the correlation between adjacent teeth is reflected in the process of training the artificial neural network corresponding to the reading model.
  • Each of the matrices 510 and 520 corresponds to an example of visualizing the Pearson correlation matrix of the maxillary teeth and the Pearson correlation matrix of the mandibular teeth calculated based on Equation (2). Elements of each matrix may correspond to a degree of correlation between teeth. An element visualized with a darker color (or darker shade) corresponds to a higher correlation, and an element visualized with a lighter color (or lighter shade) corresponds to a lower correlation. Referring to the matrices 510 and 520, it can be seen that the correlation between teeth adjacent to each other is calculated high, which corresponds to the above-mentioned.
  • the hardware may include a general-purpose computer and/or a dedicated computing device, or a specific computing device or special features or components of a specific computing device.
  • the processes may be realized by one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices, with internal and/or external memory.
  • the processes can be configured to process application specific integrated circuits (ASICs), programmable gate arrays, programmable array logic (PAL) or electronic signals.
  • ASICs application specific integrated circuits
  • PAL programmable array logic
  • the machine-readable recording medium may include program instructions, data files, data structures, etc. alone or in combination.
  • the program instructions recorded on the machine-readable recording medium may be specially designed and configured for the present invention, or may be known to and usable by a person skilled in the computer software field.
  • Examples of machine-readable recording media include magnetic media such as hard disks, floppy disks and magnetic tapes, optical recording media such as CD-ROM, DVD, Blu-ray, and magnetic-optical media such as floptical disks.
  • magnetic-optical media and a hardware device specially configured to store and execute program instructions such as ROM, RAM, flash memory, and the like.
  • program instructions include a processor, a processor architecture, or a heterogeneous combination of different hardware and software combinations, as well as any one of the aforementioned devices, or storing and compiling or interpreting to be executed on a machine capable of executing any other program instructions.
  • the method and combinations of methods may be implemented as executable code that performs the respective steps.
  • the method may be implemented as systems that perform the steps, and the methods may be distributed in several ways across devices or all functions may be integrated into one dedicated, standalone device or other hardware.
  • the means for performing the steps associated with the processes described above may include any hardware and/or software described above. All such sequential combinations and combinations are intended to be within the scope of this disclosure.
  • the hardware device may be configured to operate as one or more software modules to perform the processing according to the present invention, and vice versa.
  • the hardware device may include a processor such as an MPU, CPU, GPU, or TPU that is combined with a memory such as ROM/RAM for storing program instructions and configured to execute instructions stored in the memory, and external devices and signals It may include a communication unit that can send and receive.
  • the hardware device may include a keyboard, a mouse, and other external input devices for receiving commands written by developers.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Public Health (AREA)
  • Biomedical Technology (AREA)
  • Physics & Mathematics (AREA)
  • Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Veterinary Medicine (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Epidemiology (AREA)
  • Artificial Intelligence (AREA)
  • Optics & Photonics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Pulmonology (AREA)
  • Primary Health Care (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physiology (AREA)
  • Psychiatry (AREA)
  • Signal Processing (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

The present disclosure relates to a method for providing tooth lesion information carried out by a computing device. The method for providing tooth lesion information, carried out by a computing device, may comprise the steps of: (a) receiving an image of a tooth of a subject; (b) detecting, via a pre-trained interpretation model, a lesion area included in the tooth image, or, supporting such that another device, linked to the computing device, detects the lesion area; (c) generating tooth lesion information for the lesion area, or, supporting such that another device, linked to the computing device, generates the tooth lesion information; and (d) providing the lesion information to an external entity.

Description

치아 병변 정보 제공 방법 및 이를 이용한 장치Method for providing tooth lesion information and device using the same
본 발명은 치아 병변 정보 제공 방법 및 이를 이용한 장치에 관한 것이다.The present invention relates to a method for providing tooth lesion information and an apparatus using the same.
다양한 방식으로 촬영된 다양한 X선 영상이 치과 진단 및 치료용으로 널리 활용되고 있다. X선 파노라마 영상(panorama image)은 구강 구조를 전체적으로 확인할 수 있고, 구강의 해부학적 구조를 적절히 드러내는 점에서 치료 계획 수립에 널리 사용되고 있다. 보통 X선 파노라마 영상을 육안으로 식별하여 병변 영역을 예상하는 방식이 치과 치료에 널리 사용되고 있으나, 육안에 의한 식별의 한계를 극복하기 위하여 병변을 자동으로 식별하는 기술들이 개발되고 있다. X선 영상에서 병변 후보 영역을 설정하기 위한 연구는 X선 영상에서의 콘트라스트(Contrast) 변화를 이용하는 수준에 머물러 있다.Various X-ray images taken in various ways are widely used for dental diagnosis and treatment. Panorama X-ray images are widely used in the establishment of treatment plans in that the oral structure can be confirmed as a whole and the anatomical structure of the oral cavity is properly revealed. Usually, a method of predicting a lesion area by identifying an X-ray panoramic image with the naked eye is widely used in dental treatment, but techniques for automatically identifying a lesion have been developed to overcome the limitation of identification by the naked eye. Research to establish a lesion candidate region in an X-ray image remains at the level of using the contrast change in an X-ray image.
본 발명은 각각의 치아의 병변 발생 여부에 대한 정보를 해당 치아의 식별 정보와 함께 제공함으로써 치아 영상에서 병변 영역을 효과적으로 판단할 수 있는 수단을 제공하는 것으로 목표로 한다.An object of the present invention is to provide a means for effectively determining a lesion area in a tooth image by providing information on whether or not a lesion of each tooth occurs together with identification information of the corresponding tooth.
또한, 본 발명은 입력 치아 영상에 각 영역이 병변 영역에 해당할 가능성을 시각화하는 맵을 통한 시각화된 병변 정보를 제공하고자 한다.In addition, the present invention is to provide the visualized lesion information through a map that visualizes the possibility that each region corresponds to the lesion region in the input tooth image.
그리고, 본원 발명은 치아 영상에서 병변 영역 검출하는 판독 모델을 학습시키는 과정에서 치아들 사이의 상관 관계를 고려한 손실 함수를 이용함으로써, 보다 정확하게 병변 영역을 검출할 수 있는 판독 모델을 제공하고자 한다.In addition, an object of the present invention is to provide a readout model capable of more accurately detecting a lesion area by using a loss function in consideration of a correlation between teeth in a process of learning a readout model for detecting a lesion area in a tooth image.
상기한 바와 같은 본 발명의 목적을 달성하고, 후술하는 본 발명의 특징적인 효과를 실현하기 위한 본 발명의 특징적인 구성은 하기와 같다.A characteristic configuration of the present invention for achieving the object of the present invention as described above and realizing the characteristic effects of the present invention described later is as follows.
본 발명의 일 태양(aspect)에 따르면, 컴퓨팅 장치에 의해 수행되는 치아 병변 정보 제공 방법은 (a) 피검자의 치아 영상을 수신하는 단계; (b) 미리 학습된 판독 모델을 통해 상기 치아 영상에 포함되는 병변 영역을 검출하거나 상기 컴퓨팅 장치에 연동되는 타 장치로 하여금 상기 병변 영역을 검출하도록 지원하는 단계; (c) 상기 병변 영역에 대한 치아 병변 정보를 생성하거나 상기 컴퓨팅 장치에 연동되는 타 장치로 하여금 상기 치아 병변 정보를 생성하도록 지원하는 단계; 및 (d) 병변 정보를 외부 엔티티에 제공하는 단계를 포함할 수 있다.According to an aspect of the present invention, a method for providing tooth lesion information performed by a computing device includes the steps of: (a) receiving a tooth image of a subject; (b) detecting a lesion area included in the tooth image through a pre-learned read model or supporting another device linked to the computing device to detect the lesion area; (c) generating tooth lesion information on the lesion area or supporting another device linked to the computing device to generate the tooth lesion information; And (d) providing lesion information to an external entity.
일 태양에 따르면, 상기 (c) 단계는 (c1) 상기 치아 영상에서 병변 영역을 시각화하는 맵을 생성하는 단계; 및 (c2) 상기 맵에 기초하여 상기 치아 영상에서 병변 영역이 시각화된 시각화 정보를 통해 상기 치아 병변 정보를 생성하는 단계를 포함할 수 있다.According to an aspect, the step (c) includes the steps of (c1) generating a map for visualizing a lesion area in the tooth image; And (c2) generating the tooth lesion information through visualization information in which the lesion area is visualized in the tooth image based on the map.
다른 태양에 따르면, 상기 (c) 단계는 (c3) 상기 치아 영상에 포함되는 개별 치아의 위치 또는 순서를 식별하는 식별 정보를 생성하는 단계; (c4) 상기 식별 정보와 상기 식별 정보에 대응되는 개별 치아에 병변이 발생하였는지 여부에 대한 정보를 매칭한 매칭 정보를 생성하는 단계; 및 (c5) 상기 매칭 정보에 기초하여 상기 치아 병변 정보를 생성하는 단계를 포함할 수 있다.According to another aspect, the step (c) includes the steps of (c3) generating identification information for identifying the position or order of individual teeth included in the tooth image; (c4) generating matching information by matching the identification information with information on whether a lesion has occurred in an individual tooth corresponding to the identification information; And (c5) generating the tooth lesion information based on the matching information.
일 태양에 따르면 상기 판독 모델은 손실 함수의 결과를 최소화하는 방향으로 미리 학습되고, 상기 손실 함수는 개별 치아들 사이의 상관 관계 정보에 기초하여 결정될 수 있다.According to an aspect, the readout model is pre-trained in a direction that minimizes the result of the loss function, and the loss function may be determined based on correlation information between individual teeth.
일 태양에 따른 치아 병변 정보를 제공하는 컴퓨팅 장치는 피검자의 치아 영상을 수신하는 통신부; 및 상기 치아 영상에 대한 치아 병변 정보를 생성하는 프로세서를 포함하고, 상기 프로세서는 미리 학습된 판독 모델을 통해 상기 치아 영상에 포함되는 병변 영역에 대한 치아 병변 정보를 생성하거나 상기 컴퓨팅 장치에 연동되는 타 장치로 하여금 상기 병변 정보를 생성하도록 지원하고, 상기 치아 병변 정보를 외부 엔티티에 제공할 수 있다.A computing device that provides tooth lesion information according to an aspect includes: a communication unit that receives a tooth image of a subject; And a processor for generating tooth lesion information for the tooth image, wherein the processor generates tooth lesion information on a lesion region included in the tooth image through a pre-learned read model, or other information linked to the computing device. It is possible to support the device to generate the lesion information and provide the tooth lesion information to an external entity.
본 발명의 일 실시 예에 의하면, 치아들 사이에 상관 관계를 고려한 손실함수에 기반하여 판독 모델을 학습시킴으로써, 치아 영상에 대한 병변 영역 검출의 정확도를 획기적으로 향상시킬 수 있다.According to an embodiment of the present invention, by learning a readout model based on a loss function in consideration of a correlation between teeth, it is possible to dramatically improve the accuracy of detecting a lesion area for a tooth image.
또한, 치아 영상의 병변 검출 결과를 시각화하는 맵을 통해 치아 영상에서 병변 영역을 효과적으로 가시화하는 수단을 제공하고, 치아 식별 번호와 해당 치아에 병변 발생 여부에 대한 정보를 함께 제공하는 인터페이스를 통해 판독자가 단일 치아 영상으로 개별 치아들에 병변 발생 여부를 판독할 수 있는 수단을 제공할 수 있다.In addition, it provides a means to effectively visualize the lesion area in the tooth image through a map that visualizes the lesion detection result of the tooth image, and through an interface that provides information on whether a lesion has occurred in the tooth and a tooth identification number, the reader can A single tooth image may provide a means for reading whether or not lesions have occurred in individual teeth.
본 발명에 따르면 궁극적으로 의료진의 진단의 정확도를 향상시키고, 단일 영상을 통해 전체 치아의 병변 발생 여부를 판독할 수 있도록 함으로써, 의료 현장에서의 워크플로(workflow)를 혁신할 수 있게 되는 잠재적 효과가 있다.According to the present invention, the potential effect of being able to innovate the workflow in the medical field by ultimately improving the accuracy of diagnosis by the medical staff and allowing the entire tooth to be able to read whether or not lesions have occurred through a single image. have.
그리고 본 발명은, 종래에 병원에서 이용하고 있는 의료 영상이 그대로 활용될 수 있는바, 본 발명의 방법이 특정 형식의 영상이나 플랫폼에 종속되지 않음은 물론이다.In addition, in the present invention, a medical image used in a conventional hospital can be used as it is, so it goes without saying that the method of the present invention is not dependent on a specific type of image or platform.
본 발명의 실시 예의 설명에 이용되기 위하여 첨부된 아래 도면들은 본 발명의 실시 예들 중 단지 일부일 뿐이며, 본 발명이 속한 기술분야에서 통상의 지식을 가진 사람(이하 "통상의 기술자"라고 함)에게 있어서는 발명에 이르는 노력 없이 이 도면들에 기초하여 다른 도면들이 얻어질 수 있다.The accompanying drawings, which are attached to be used in the description of the embodiments of the present invention, are only some of the embodiments of the present invention, and for those of ordinary skill in the technical field to which the present invention belongs (hereinafter referred to as "common technician") Other drawings can be obtained on the basis of these drawings without efforts to reach the invention.
도 1은 본 발명에 따른 병변 정보 제공 방법을 수행하는 컴퓨팅 장치의 예시적 구성을 개략적으로 도시한 개념도이다.1 is a conceptual diagram schematically showing an exemplary configuration of a computing device that performs a method for providing lesion information according to the present invention.
도 2는 본 발명에 따른 병변 정보 제공 방법을 수행하는 컴퓨팅 장치의 하드웨어 또는 소프트웨어 구성요소를 도시한 예시적 블록도이다.2 is an exemplary block diagram showing hardware or software components of a computing device that performs a method for providing lesion information according to the present invention.
도 3은 일 실시 예에 따른 치아 병변 정보 제공 방법을 설명하기 위한 흐름도이다.3 is a flowchart illustrating a method of providing tooth lesion information according to an exemplary embodiment.
도 4는 일 실시 예에 따른 병변 정보 제공 방법이 수행되는 일례를 도시하는 도면이다.4 is a diagram illustrating an example in which a method of providing lesion information according to an exemplary embodiment is performed.
도 5는 손실 함수 결정 과정에서 사용되는 피어슨 상관 행렬이 시각화된 일례를 도시하는 도면이다.5 is a diagram illustrating an example in which a Pearson correlation matrix used in a process of determining a loss function is visualized.
후술하는 본 발명에 대한 상세한 설명은, 본 발명의 목적들, 기술적 해법들 및 장점들을 분명하게 하기 위하여 본 발명이 실시될 수 있는 특정 실시 예를 예시로서 도시하는 첨부 도면을 참조한다. 이들 실시 예는 통상의 기술자가 본 발명을 실시할 수 있기에 충분하도록 상세히 설명된다.Detailed description of the present invention to be described later, in order to clarify the objects, technical solutions and advantages of the present invention, reference is made to the accompanying drawings, which illustrate specific embodiments in which the present invention may be practiced as an example. These embodiments are described in detail sufficient to enable a person skilled in the art to practice the present invention.
본 발명의 상세한 설명 및 청구항들에 걸쳐 이용된 "영상" 또는 "영상 데이터"라는 용어는 이산적 영상 요소들(예컨대, 2차원 영상에 있어서는 픽셀, 3차원 영상에 있어서는 복셀)로 구성된 다차원 데이터를 지칭한다. 예를 들어 "영상"은 X선 영상, (콘-빔형; cone-beam) 전산화 단층 촬영(computed tomography), MRI(magnetic resonance imaging), 초음파 또는 본 발명의 기술분야에서 공지된 임의의 다른 의료 영상 시스템의 의하여 수집된 피사체, 즉 피검체(subject)의 의료 영상일 수 있다. 또한 영상은 비의료적 맥락에서 제공될 수도 있는바, 예를 들어 원격 감지 시스템(remote sensing system), 전자현미경(electron microscopy) 등등이 있을 수 있다.The term "image" or "image data" used throughout the detailed description and claims of the present invention refers to multidimensional data composed of discrete image elements (eg, pixels in a 2D image and voxels in a 3D image). Refers to. For example, "image" refers to an X-ray image, (cone-beam) computed tomography, magnetic resonance imaging (MRI), ultrasound, or any other medical image known in the art. It may be a medical image of a subject, that is, a subject collected by the system. Also, an image may be provided in a non-medical context, such as a remote sensing system, an electron microscopy, and the like.
본 발명의 상세한 설명 및 청구항들에 걸쳐, '영상'은 (예컨대, 비디오 화면에 표시된) 눈으로 볼 수 있는 영상 또는 (예컨대, CT, MRI 검출기 등의 픽셀 출력에 대응되는 파일과 같은) 영상의 디지털 표현물을 지칭하는 용어이다.Throughout the detailed description and claims of the present invention,'image' refers to an image that can be seen by the eye (e.g., displayed on a video screen) or of an image (e.g., a file corresponding to a pixel output of a CT, MRI detector, etc.) It is a term referring to digital representations.
본 발명의 상세한 설명 및 청구항들에 걸쳐 'DICOM(Digital Imaging and Communications in Medicine; 의료용 디지털 영상 및 통신)' 표준은 의료용 기기에서 디지털 영상 표현과 통신에 이용되는 여러 가지 표준을 총칭하는 용어인바, DICOM 표준은 미국 방사선 의학회(ACR)와 미국 전기 공업회(NEMA)에서 구성한 연합 위원회에서 발표한다.Throughout the detailed description and claims of the present invention, the'DICOM (Digital Imaging and Communications in Medicine; Medical Digital Imaging and Communications)' standard is a generic term for various standards used for digital image expression and communication in medical devices. Standards are published by a coalition committee formed by the American Radiological Society (ACR) and the American Electrical Industry Association (NEMA).
또한, 본 발명의 상세한 설명 및 청구항들에 걸쳐 '의료영상 저장 전송 시스템(PACS; picture archiving and communication system)'은 DICOM 표준에 맞게 저장, 가공, 전송하는 시스템을 지칭하는 용어이며, X선, CT, MRI와 같은 디지털 의료영상 장비를 이용하여 획득된 의료영상 이미지는 DICOM 형식으로 저장되고 네트워크를 통하여 병원 내외의 단말로 전송이 가능하며, 이에는 판독 결과 및 진료 기록이 추가될 수 있다.In addition, throughout the detailed description and claims of the present invention,'medical image storage and transmission system (PACS; picture archiving and communication system)' is a term that refers to a system that stores, processes, and transmits according to the DICOM standard, and X-ray, CT , Medical image images acquired using digital medical imaging equipment such as MRI are stored in DICOM format and can be transmitted to terminals inside and outside the hospital through a network, and reading results and medical records can be added to this.
그리고 본 발명의 상세한 설명 및 청구항들에 걸쳐 '학습' 혹은 '러닝'은 절차에 따른 컴퓨팅(computing)을 통하여 기계 학습(machine learning)을 수행함을 일컫는 용어인바, 인간의 교육 활동과 같은 정신적 작용을 지칭하도록 의도된 것이 아님을 통상의 기술자는 이해할 수 있을 것이다.And throughout the detailed description and claims of the present invention,'learning' or'learning' is a term that refers to performing machine learning through computing according to procedures. It will be appreciated by those of skill in the art that it is not intended to be referred to.
그리고 본 발명의 상세한 설명 및 청구항들에 걸쳐, '포함하다'라는 단어 및 그 변형은 다른 기술적 특징들, 부가물들, 구성요소들 또는 단계들을 제외하는 것으로 의도된 것이 아니다. 또한, '하나' 또는 '한'은 하나 이상의 의미로 쓰인 것이며, '또 다른'은 적어도 두 번째 이상으로 한정된다.And throughout the detailed description and claims of the present invention, the word'comprise' and its variations are not intended to exclude other technical features, additions, components or steps. In addition,'one' or'one' is used in one or more meanings, and'another' is limited to at least a second or more.
통상의 기술자에게 본 발명의 다른 목적들, 장점들 및 특성들이 일부는 본 설명서로부터, 그리고 일부는 본 발명의 실시로부터 드러날 것이다. 아래의 예시 및 도면은 실례로서 제공되며, 본 발명을 한정하는 것으로 의도된 것이 아니다. 따라서, 특정 구조나 기능에 관하여 본 명세서에 개시된 상세 사항들은 한정하는 의미로 해석되어서는 아니되고, 단지 통상의 기술자가 실질적으로 적합한 임의의 상세 구조들로써 본 발명을 다양하게 실시하도록 지침을 제공하는 대표적인 기초 자료로 해석되어야 할 것이다.Other objects, advantages, and features of the present invention to those skilled in the art will appear, in part, from this description, and in part from the practice of the present invention. The examples and drawings below are provided by way of example and are not intended to limit the invention. Therefore, the details disclosed in this specification with respect to a specific structure or function are not to be construed in a limiting sense, but a representative of providing guidance for a person skilled in the art to variously implement the present invention with any detailed structures that are substantially suitable. It should be interpreted as basic data.
더욱이 본 발명은 본 명세서에 표시된 실시 예들의 모든 가능한 조합들을 망라한다. 본 발명의 다양한 실시 예는 서로 다르지만 상호 배타적일 필요는 없음이 이해되어야 한다. 예를 들어, 여기에 기재되어 있는 특정 형상, 구조 및 특성은 일 실시 예에 관련하여 본 발명의 사상 및 범위를 벗어나지 않으면서 다른 실시 예로 구현될 수 있다. 또한, 각각의 개시된 실시 예 내의 개별 구성요소의 위치 또는 배치는 본 발명의 사상 및 범위를 벗어나지 않으면서 변경될 수 있음이 이해되어야 한다. 따라서, 후술하는 상세한 설명은 한정적인 의미로서 취하려는 것이 아니며, 본 발명의 범위는, 적절하게 설명된다면, 그 청구항들이 주장하는 것과 균등한 모든 범위와 더불어 첨부된 청구항에 의해서만 한정된다. 도면에서 유사한 참조부호는 여러 측면에 걸쳐서 동일하거나 유사한 기능을 지칭한다.Moreover, the present invention covers all possible combinations of the embodiments indicated herein. It should be understood that the various embodiments of the present invention are different from each other, but need not be mutually exclusive. For example, specific shapes, structures, and characteristics described herein may be implemented in other embodiments without departing from the spirit and scope of the present invention in relation to one embodiment. In addition, it should be understood that the location or arrangement of individual components in each disclosed embodiment may be changed without departing from the spirit and scope of the present invention. Accordingly, the detailed description to be described below is not intended to be taken in a limiting sense, and the scope of the present invention, if appropriately described, is limited only by the appended claims, along with all ranges equivalent to those claimed by the claims. Like reference numerals in the drawings refer to the same or similar functions over several aspects.
본 명세서에서 달리 표시되거나 분명히 문맥에 모순되지 않는 한, 단수로 지칭된 항목은, 그 문맥에서 달리 요구되지 않는 한, 복수의 것을 아우른다. 또한, 본 발명을 설명함에 있어, 관련된 공지 구성 또는 기능에 대한 구체적인 설명이 본 발명의 요지를 흐릴 수 있다고 판단되는 경우에는 그 상세한 설명은 생략한다.Unless otherwise indicated in this specification or clearly contradicting the context, items referred to in the singular encompass the plural unless otherwise required by that context. In addition, in describing the present invention, when it is determined that a detailed description of a related known configuration or function may obscure the subject matter of the present invention, a detailed description thereof will be omitted.
이하, 통상의 기술자가 본 발명을 용이하게 실시할 수 있도록 하기 위하여, 본 발명의 바람직한 실시 예들에 관하여 첨부된 도면을 참조하여 상세히 설명하기로 한다.Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings in order to enable those skilled in the art to easily implement the present invention.
도 1은 본 발명에 따른 병변 정보 제공 방법을 수행하는 컴퓨팅 장치의 예시적 구성을 개략적으로 도시한 개념도이다.1 is a conceptual diagram schematically showing an exemplary configuration of a computing device that performs a method for providing lesion information according to the present invention.
도 1을 참조하면, 본 발명의 일 실시 예에 따른 컴퓨팅 장치(100)는, 통신부(110) 및 프로세서(120)를 포함하며, 상기 통신부(110)를 통하여 외부 컴퓨팅 장치(미도시)와 직간접적으로 통신할 수 있다.Referring to FIG. 1, a computing device 100 according to an embodiment of the present invention includes a communication unit 110 and a processor 120, and is directly or indirectly connected to an external computing device (not shown) through the communication unit 110. Can communicate with enemies.
구체적으로, 상기 컴퓨팅 장치(100)는, 전형적인 컴퓨터 하드웨어(예컨대, 컴퓨터 프로세서, 메모리, 스토리지, 입력 장치 및 출력 장치, 기타 기존의 컴퓨팅 장치의 구성요소들을 포함할 수 있는 장치; 라우터, 스위치 등과 같은 전자 통신 장치; 네트워크 부착 스토리지(NAS; network-attached storage) 및 스토리지 영역 네트워크(SAN; storage area network)와 같은 전자 정보 스토리지 시스템)와 컴퓨터 소프트웨어(즉, 컴퓨팅 장치로 하여금 특정의 방식으로 기능하게 하는 명령어들)의 조합을 이용하여 원하는 시스템 성능을 달성하는 것일 수 있다.Specifically, the computing device 100 is a device that may include typical computer hardware (eg, computer processor, memory, storage, input devices and output devices, and other components of existing computing devices; routers, switches, etc.). Electronic communication devices; electronic information storage systems such as network-attached storage (NAS) and storage area networks (SANs) and computer software (i.e., allowing the computing device to function in a specific way). Instructions) may be used to achieve the desired system performance.
이와 같은 컴퓨팅 장치의 통신부(110)는 연동되는 타 컴퓨팅 장치와 요청과 응답을 송수신할 수 있는바, 일 예시로서 그러한 요청과 응답은 동일한 TCP(transmission control protocol) 세션(session)에 의하여 이루어질 수 있지만, 이에 한정되지는 않는바, 예컨대 UDP(user datagram protocol) 데이터그램(datagram)으로서 송수신될 수도 있을 것이다. 덧붙여, 넓은 의미에서 상기 통신부(110)는 명령어 또는 지시 등을 전달받기 위한 키보드, 마우스, 기타 외부 입력장치, 프린터, 디스플레이, 기타 외부 출력장치를 포함할 수 있다.The communication unit 110 of such a computing device can transmit and receive requests and responses to and from other computing devices to which it is linked. As an example, such requests and responses may be made by the same transmission control protocol (TCP) session. However, the present invention is not limited thereto, and may be transmitted/received as, for example, a user datagram protocol (UDP) datagram. In addition, in a broad sense, the communication unit 110 may include a keyboard, a mouse, other external input devices, printers, displays, and other external output devices for receiving commands or instructions.
또한, 컴퓨팅 장치의 프로세서(120)는 MPU(micro processing unit), CPU(central processing unit), GPU(graphics processing unit) 또는 TPU(tensor processing unit), 캐시 메모리(cache memory), 데이터 버스(data bus) 등의 하드웨어 구성을 포함할 수 있다. 또한, 운영체제, 특정 목적을 수행하는 애플리케이션의 소프트웨어 구성을 더 포함할 수도 있다.In addition, the processor 120 of the computing device may include a micro processing unit (MPU), a central processing unit (CPU), a graphics processing unit (GPU) or a tensile processing unit (TPU), a cache memory, and a data bus. ), and the like. In addition, it may further include an operating system and a software configuration of an application that performs a specific purpose.
하기의 설명에서는 본원 발명의 영상은 치아 영상을 대상으로 하고, 치아 영상은 일례로 치아 X선 파노라마 영상을 대상으로 하고 있으나, 본원 발명의 권리범위는 이에 한정되지 않고, 일반적인 형태의 치아 영상에 모두 적용이 가능하고, 대상이 되는 영상에 포함되는 임의의 병변 영역을 검출하는 과정에서 적용될 수 있다는 점은 통상의 기술자가 용이하게 이해할 것이다.In the following description, the image of the present invention targets a tooth image, and the tooth image targets a tooth X-ray panoramic image as an example, but the scope of the present invention is not limited thereto, and all of the general form tooth image It will be readily understood by those of ordinary skill in the art that it can be applied and that it can be applied in the process of detecting an arbitrary lesion area included in a target image.
도 2는 본 발명에 따른 병변 정보 제공 방법을 수행하는 컴퓨팅 장치의 하드웨어 또는 소프트웨어 구성요소를 도시한 예시적 블록도이다.2 is an exemplary block diagram showing hardware or software components of a computing device that performs a method for providing lesion information according to the present invention.
도 2에 도시된 개별 모듈들은, 예컨대, 컴퓨팅 장치(100)에 포함된 통신부(110)나 프로세서(120), 또는 상기 통신부(110) 및 프로세서(120)의 연동에 의하여 구현될 수 있음은 통상의 기술자가 이해할 수 있을 것이다.The individual modules shown in FIG. 2 may be implemented by interlocking the communication unit 110 or the processor 120 included in the computing device 100, or the communication unit 110 and the processor 120, for example. Technicians will be able to understand.
도 2를 참조하여 본 발명에 따른 방법 및 장치의 구성을 간략히 개관하면, 컴퓨팅 장치(100)는 그 구성요소로서 영상 획득 모듈(210)을 포함할 수 있다. 이 영상 획득 모듈(210)은 데이터베이스에 미리 저장되거나 영상 촬영을 위한 촬영 전용 기기로부터 획득되는 피검자의 치아 영상을 획득할 수 있다. 치아 영상은 컴퓨팅 장치(100)에 내장된 촬영 전용 기기 또는 외부의 촬영 전용 기기를 통해 촬영된 피검자의 치아에 대한 X선 파노라마 영상일 수 있다.Referring to FIG. 2 to briefly overview the configuration of the method and apparatus according to the present invention, the computing device 100 may include an image acquisition module 210 as its constituent element. The image acquisition module 210 may acquire a tooth image of a subject, which is previously stored in a database or obtained from a device dedicated to photographing an image. The tooth image may be an X-ray panoramic image of a subject's teeth captured through a dedicated photographing device built into the computing device 100 or an external photographing dedicated device.
영상 획득 모듈(210)을 통해 획득한 치아 영상은 치아 병변 영역 검출 모듈(220)로 전송될 수 있다. 치아 병변 영역 검출 모듈(220)은 판독 모델을 통해 치아 영상에 포함된 병변 영역을 검출할 수 있다. 판독 모델은 치아 영상에서 병변 영역을 검출하도록 미리 학습된 인공 신경망일 수 있다. 치아 병변 영역 검출 모듈(220)에 포함된 판독 모델은 입력된 치아 영상에서 병변 영역의 위치, 병변 영역의 형태, 병변 영역에 해당될 확률 중 적어도 하나를 출력하도록 학습된 딥 러닝 기반의 인공 신경망일 수 있다. 치아 병변 영역 검출 모듈(220)에 포함되는 인공 신경망은 입력 치아 영상에 포함되는 영역과 탐지 대상 영역(예를 들어 골 소실 영역)이 일치하는지 여부를 그 유사도에 기초하여 결정하고, 결정 결과를 출력하도록 학습되는 것으로, 예를 들어, DNN (Deep Neural Network), CNN (Convolutional Neural Network), DCNN (Deep Convolution Neural Network), RNN (Recurrent Neural Network), RBM (Restricted Boltzmann Machine), DBN (Deep Belief Network), SSD (Single Shot Detector), YOLO (You Only Look Once) 등이 이용될 수 있음은 통상의 기술자가 이해할 것이다. 치아 병변 영역 검출 모듈(220)에서 사용될 수 있는 인공 신경망의 종류는 제시된 예시에 한정되지 않고, 라벨링된 학습 데이터에 기반하여 치아 영상에서 병변 영역을 인식하도록 학습될 수 있는 임의의 인공 신경망을 포함할 수 있다는 점은 통상의 기술자가 이해할 수 있을 것이다.The tooth image acquired through the image acquisition module 210 may be transmitted to the tooth lesion region detection module 220. The tooth lesion region detection module 220 may detect a lesion region included in the tooth image through a read model. The readout model may be an artificial neural network that has been trained in advance to detect a lesion area in a tooth image. The readout model included in the tooth lesion region detection module 220 is a deep learning-based artificial neural network that has been trained to output at least one of the location of the lesion region, the shape of the lesion region, and the probability corresponding to the lesion region from the input tooth image. I can. The artificial neural network included in the tooth lesion region detection module 220 determines whether the region included in the input tooth image matches the detection target region (eg, bone loss region) based on the similarity, and outputs the determination result. For example, DNN (Deep Neural Network), CNN (Convolutional Neural Network), DCNN (Deep Convolution Neural Network), RNN (Recurrent Neural Network), RBM (Restricted Boltzmann Machine), DBN (Deep Belief Network) ), SSD (Single Shot Detector), YOLO (You Only Look Once), etc. can be used, it will be understood by those of ordinary skill in the art. The type of artificial neural network that can be used in the tooth lesion region detection module 220 is not limited to the presented example, and may include any artificial neural network that can be learned to recognize the lesion region in the tooth image based on the labeled learning data. It will be understood by those of ordinary skill in the art that it can be.
치아 병변 영역 검출 모듈(220)을 통해 수행된 병변 영역에 대한 검출 결과는 시각화 모듈(230)에 전달될 수 있고, 시각화 모듈(230)은 치아 영상에서 병변 영역을 시각화하는 맵을 생성할 수 있다. 예를 들어, 시각화 모듈(230)은 치아의 각 영역이 병변 영역(예를 들어 골 소실 영역)에 해당될 확률에 따라 대응되는 영역의 색상을 달리하는 히트맵(hit map)을 생성할 수 있다. 시각화 모듈(230)이 맵을 생성하는 방식은 앞서 설명된 방식에 한정되지 않고, 치아 영상의 각 영역이 병변 영역에 해당할 확률을 시각적으로 표현할 수 있는 임의의 방식(예를 들어, 병변 영역에 해당할 확률을 수치화하여 표현하거나, 병변 영역에 해당할 확률에 따라 명암을 조절하는 등)을 포함할 수 있음은 통상의 기술자가 자명하게 이해할 수 있을 것이다.The detection result of the lesion region performed through the tooth lesion region detection module 220 may be transmitted to the visualization module 230, and the visualization module 230 may generate a map for visualizing the lesion region in the tooth image. . For example, the visualization module 230 may generate a heat map in which the color of a corresponding region is changed according to a probability that each region of a tooth corresponds to a lesion region (eg, a bone loss region). . The method of generating the map by the visualization module 230 is not limited to the method described above, and an arbitrary method capable of visually expressing the probability that each region of the tooth image corresponds to the lesion region (e.g., It will be apparent to those of ordinary skill in the art that the corresponding probability may be expressed in a numerical manner, or the contrast may be adjusted according to the probability corresponding to the lesion area).
치아 병변 영역 검출 모듈(220)을 통해 수행된 검출 결과는 분류 모듈(240)에 전달될 수 있고, 분류 모듈(240)은 검출된 병변 영역에 대응되는 치아의 식별 정보를 결정할 수 있다. 분류 모듈(240)은 치아 영상에 포함된 개별 치아의 위치 또는 순서를 식별하는 식별 정보를 생성하고, 생성된 식별 정보 및 상기 검출 결과에 기초하여 병변 영역이 검출된 치아의 식별 정보를 결정할 수 있다. 개별 치아에 대한 식별 정보는 전 세계 치과의사들이 가장 많이 사용하는 치아 식별 번호인 FDI World Dental Federation notation에 기초하여 결정될 수 있으나, 이에 한정되지 않고, 개별 치아를 식별할 수 있는 임의의 방식에 기초하여 결정될 수 있다. 예를 들어, 식별 정보로 치아 식별 번호가 이용되는 경우, 분류 모듈(240)은 개별 치아에 대해 생성된 식별 번호 중 병변이 검출된 치아의 식별 번호를 결정하고, 결정 결과에 기초하여 식별 번호와 개별 치아에 병변이 발생하였는지 여부에 대한 정보를 매칭한 매칭 정보를 생성할 수 있다. 매칭 정보는 구현에 따라 병변이 발생된 것으로 결정된 치아들의 식별 정보를 리스트로 제공하는 방식, 모든 치아의 식별 정보와 병변 발생 여부를 매칭하는 방식 등 식별 정보와 병변 발생 여부에 대한 정보를 매칭하는 임의의 방식으로 생성될 수 있다.The detection result performed through the tooth lesion region detection module 220 may be transmitted to the classification module 240, and the classification module 240 may determine identification information of a tooth corresponding to the detected lesion region. The classification module 240 may generate identification information for identifying the position or order of individual teeth included in the tooth image, and may determine identification information of the tooth in which the lesion area is detected based on the generated identification information and the detection result. . The identification information for individual teeth may be determined based on the FDI World Dental Federation notation, which is a dental identification number most used by dentists around the world, but is not limited thereto, and is based on an arbitrary method by which individual teeth can be identified. Can be determined. For example, when a tooth identification number is used as the identification information, the classification module 240 determines the identification number of the tooth from which the lesion is detected among identification numbers generated for individual teeth, and based on the determination result, the identification number and It is possible to generate matching information by matching information on whether a lesion has occurred in an individual tooth. Matching information is a method of providing a list of identification information of teeth determined to have lesions according to implementation, a method of matching identification information of all teeth and whether lesions have occurred, etc. It can be created in the way of
치아 병변 정보 생성 및 전송 모듈(250)은 시각화 모듈(230)에 의해 생성된 맵 또는 분류 모듈(240)을 통해 생성된 매칭 정보에 기초하여 치아 병변 정보를 생성할 수 있다. 치아 병변 정보는 상기 맵을 치아 영상에 적용함으로써 생성되는 시각화 정보 또는 상기 매칭 정보에 기초하여 병변이 검출된 치아 리스트에 대한 정보 중 적어도 하나에 기초하여 결정될 수 있다.The tooth lesion information generation and transmission module 250 may generate tooth lesion information based on a map generated by the visualization module 230 or matching information generated through the classification module 240. The tooth lesion information may be determined based on at least one of visualization information generated by applying the map to a tooth image or information on a tooth list in which a lesion is detected based on the matching information.
치아 병변 정보 생성 및 전송 모듈(250)을 통해 생성된 치아 병변 정보는 데이터베이스에 저장되거나, 외부 엔티티에 제공될 수 있다. 치아 병변 정보를 외부 엔티티에 제공되는 때에는 치아 병변 정보 생성 및 전송 모듈(250)은 소정의 디스플레이 장치 등을 이용하거나, 구비된 통신부를 통해 외부 엔티티에 치아 병변 정보가 반영된 출력 영상을 제공할 수 있다. 여기에서 외부 엔티티라고 함은, 상기 컴퓨팅 장치(100)의 사용자, 관리자, 상기 피검체를 담당하는 담당 의료 전문가 등을 포함하나, 이 이외에도 상기 입력된 치아 영상에 대한 치아 병변 정보를 필요로 하는 주체라면 어느 주체라도 포함되는 것으로 이해되어야 할 것이다. 예를 들어, 상기 외부 엔티티는 상기 치아 병변 정보를 활용하는 별도의 AI(artificial intelligence; 인공지능) 하드웨어 및/또는 소프트웨어 모듈을 포함하는 외부의 AI 장치일 수도 있다. 또한, 외부 엔티티에서의 '외부(external)'는 상기 치아 병변 정보를 이용하는 AI 하드웨어 및/또는 소프트웨어 모듈이 상기 컴퓨팅 장치(100)에 일체화되는 실시 예를 배제하도록 의도된 것이 아니라, 본 발명의 방법을 수행하는 하드웨어 및/또는 소프트웨어 모듈의 결과물인 치아 병변 정보가 타 방법의 입력 데이터로 활용될 수 있음을 시사하도록 이용된 것임을 밝혀둔다. 즉, 상기 외부 엔티티는 컴퓨팅 장치(100) 자체일 수도 있다.The tooth lesion information generated through the tooth lesion information generation and transmission module 250 may be stored in a database or may be provided to an external entity. When the tooth lesion information is provided to an external entity, the tooth lesion information generation and transmission module 250 may provide an output image reflecting the tooth lesion information to the external entity using a predetermined display device or the like, or through an provided communication unit. . Here, the external entity includes a user of the computing device 100, an administrator, a medical professional in charge of the subject, etc., but in addition to this, a subject who needs tooth lesion information for the input tooth image If so, it should be understood that any subject is included. For example, the external entity may be an external AI device including a separate artificial intelligence (AI) hardware and/or software module that utilizes the tooth lesion information. In addition,'external' in an external entity is not intended to exclude an embodiment in which an AI hardware and/or software module using the tooth lesion information is integrated into the computing device 100, but the method of the present invention. It should be noted that the tooth lesion information, which is the result of the hardware and/or software module performing the procedure, is used to suggest that it can be used as input data for other methods. That is, the external entity may be the computing device 100 itself.
도 2에 나타난 구성요소들은 설명의 편의상 하나의 컴퓨팅 장치에서 실현되는 것으로 예시되었으나, 본 발명의 방법을 수행하는 컴퓨팅 장치(100)는 복수개가 서로 연동되도록 구성될 수도 있다는 점이 이해될 것이다.Although the components shown in FIG. 2 are illustrated as being realized in one computing device for convenience of description, it will be understood that a plurality of computing devices 100 performing the method of the present invention may be configured to interlock with each other.
이제 본 발명에 따른 영상 제공 방법의 일 실시 예를 도 3 내지 도 5를 참조하여 더 구체적으로 설명하기로 한다.Now, an embodiment of an image providing method according to the present invention will be described in more detail with reference to FIGS. 3 to 5.
도 3은 일 실시 예에 따른 치아 병변 정보 제공 방법을 설명하기 위한 흐름도이다.3 is a flowchart illustrating a method of providing tooth lesion information according to an exemplary embodiment.
도 3을 참조하면, 컴퓨팅 장치는 단계(S100)에서 통신부를 통해 피검자의 치아 영상을 수신할 수 있다. 치아 영상은 앞서 설명한 바와 같이 치아 X선 파노라마 영상일 수 있으나, 이에 한정되지 않고, 피검자의 치아를 촬영한 임의의 형태의 영상을 포함할 수 있다.Referring to FIG. 3, the computing device may receive a tooth image of a subject through a communication unit in step S100. As described above, the tooth image may be a tooth X-ray panoramic image, but the present invention is not limited thereto, and may include an image of an arbitrary shape photographing the subject's teeth.
컴퓨팅 장치는 단계(S200)에서 판독 모델을 통해 치아 영상에 포함되는 병변 영역을 검출할 수 있다. 판독 모델은 치아 영상에 병변 영역을 검출하도록 미리 학습된 인공 신경망을 포함할 수 있다.The computing device may detect a lesion area included in the tooth image through the reading model in step S200. The readout model may include a pre-trained artificial neural network to detect a lesion area in a tooth image.
판독 모델을 학습시키기 위한 손실 함수는 손실 함수로써 통상적으로 사용되는 크로스 엔트로피(cross entropy) 함수와 더불어 상악 치아에 대한 상악 손실 함수 및 하악 치아에 대한 하악 손실 함수에 기초하여 결정될 수 있다. 상악이란 치아 영상에 포함된 위쪽 턱을 의미하고, 하악이란 치아 영상에 포함된 아래쪽 턱을 의미할 수 있다. 이에 따라 상악 손실 함수는 상악에 대응하여 결정되는 손실 함수를 의미할 수 있고, 하악 손실 함수는 하악에 대응되는 손실 함수를 의미할 수 있다.The loss function for training the readout model may be determined based on a maxillary loss function for a maxillary tooth and a mandibular loss function for a mandibular tooth in addition to a cross entropy function commonly used as a loss function. The upper jaw may mean an upper jaw included in the tooth image, and the mandible may mean a lower jaw included in the tooth image. Accordingly, the maxillary loss function may mean a loss function determined corresponding to the upper jaw, and the mandibular loss function may mean a loss function corresponding to the mandible.
보다 구체적으로, 상악 손실 함수 및 하악 손실 함수를 산출하기 위한 개별 손실 함수는 수학식 1에 기초하여 결정될 수 있다.More specifically, individual loss functions for calculating the maxillary loss function and the mandibular loss function may be determined based on Equation 1.
Figure PCTKR2020012448-appb-img-000001
Figure PCTKR2020012448-appb-img-000001
Figure PCTKR2020012448-appb-img-000002
는 개별 손실 함수,
Figure PCTKR2020012448-appb-img-000003
는 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000004
는 i번째 데이터 샘플의 j번째 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000005
는 치아의 병변 발생 여부에 대한 판독 모델의 예측 값,
Figure PCTKR2020012448-appb-img-000006
는 i번째 데이터 샘플의 j번째 치아의 병변 발생 여부에 대한 판독 모델의 예측 값,
Figure PCTKR2020012448-appb-img-000007
는 피어슨 상관 행렬(pearson correlation matrix),
Figure PCTKR2020012448-appb-img-000008
는 피어슨 상관 행렬의 각 요소로써 j번째 치아와 j'번째 치아에 상응하는 상관 계수 값,
Figure PCTKR2020012448-appb-img-000009
는 피어슨 상관 행렬에 기초하여 결정되는 상관 파라미터,
Figure PCTKR2020012448-appb-img-000010
는 i번째 데이터 샘플의 j번째 치아에 대해 산출된 상관 파라미터, k는 치아 번호의 개수,
Figure PCTKR2020012448-appb-img-000011
은 데이터 샘플의 개수,
Figure PCTKR2020012448-appb-img-000012
Figure PCTKR2020012448-appb-img-000013
를 0 및 1사이로 정규화하기 위한 값을 나타낸다.
Figure PCTKR2020012448-appb-img-000002
Is the individual loss function,
Figure PCTKR2020012448-appb-img-000003
Is a value indicating whether a tooth lesion has occurred,
Figure PCTKR2020012448-appb-img-000004
Is a value indicating whether a lesion of the j-th tooth of the i-th data sample has occurred,
Figure PCTKR2020012448-appb-img-000005
Is the predicted value of the readout model for the occurrence of a tooth lesion,
Figure PCTKR2020012448-appb-img-000006
Is the predicted value of the readout model for the occurrence of lesion of the j-th tooth in the i-th data sample,
Figure PCTKR2020012448-appb-img-000007
Is the pearson correlation matrix,
Figure PCTKR2020012448-appb-img-000008
Is each element of the Pearson correlation matrix and is the correlation coefficient value corresponding to the j-th tooth and the j'-th tooth,
Figure PCTKR2020012448-appb-img-000009
Is the correlation parameter determined based on the Pearson correlation matrix,
Figure PCTKR2020012448-appb-img-000010
Is the correlation parameter calculated for the j-th tooth of the i-th data sample, k is the number of teeth number,
Figure PCTKR2020012448-appb-img-000011
Is the number of data samples,
Figure PCTKR2020012448-appb-img-000012
Is
Figure PCTKR2020012448-appb-img-000013
Represents a value for normalizing between 0 and 1.
수학식 1의 피어슨 상관 행렬에 포함되는 a번째 치아와 b번째 치아 사이의 상관 계수
Figure PCTKR2020012448-appb-img-000014
는 하기의 수학식 2에 기초하여 결정될 수 있다.
The correlation coefficient between the a-th tooth and the b-th tooth included in the Pearson correlation matrix of Equation 1
Figure PCTKR2020012448-appb-img-000014
May be determined based on Equation 2 below.
Figure PCTKR2020012448-appb-img-000015
Figure PCTKR2020012448-appb-img-000015
Figure PCTKR2020012448-appb-img-000016
는 a번째 치아와 b번째 치아의 상관 계수,
Figure PCTKR2020012448-appb-img-000017
는 i번째 학습 데이터에서 a번째 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000018
는i번째 학습 데이터에서 b번째 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000019
Figure PCTKR2020012448-appb-img-000020
의 평균,
Figure PCTKR2020012448-appb-img-000021
Figure PCTKR2020012448-appb-img-000022
의 평균을 의미할 수 있다.
Figure PCTKR2020012448-appb-img-000016
Is the correlation coefficient between the a-th tooth and the b-th tooth,
Figure PCTKR2020012448-appb-img-000017
Is a value indicating whether a lesion of the a-th tooth has occurred in the i-th training data,
Figure PCTKR2020012448-appb-img-000018
Is a value indicating whether a lesion of the b-th tooth has occurred in the i-th training data,
Figure PCTKR2020012448-appb-img-000019
Is
Figure PCTKR2020012448-appb-img-000020
Mean of,
Figure PCTKR2020012448-appb-img-000021
Is
Figure PCTKR2020012448-appb-img-000022
Can mean the average of.
예를 들어, a번째 치아에 병변 발생 여부에 대한 값(병변이 발생하는 경우
Figure PCTKR2020012448-appb-img-000023
, 병변이 발생하지 않는 경우
Figure PCTKR2020012448-appb-img-000024
로 출력된다.)은 a = {1, 0, 0, 0, 0, 1, 1}이고, b = {1, 1, 0, 0, 0, 1, 1}로 학습 데이터가 7개인 경우,
Figure PCTKR2020012448-appb-img-000025
는 0.428,
Figure PCTKR2020012448-appb-img-000026
는 0.571로 산출되고, 수학식 2에 따라
Figure PCTKR2020012448-appb-img-000027
는 0.75로 산출될 수 있다.
For example, the value of whether a lesion has occurred on the a-th tooth (if a lesion occurs
Figure PCTKR2020012448-appb-img-000023
, If the lesion does not occur
Figure PCTKR2020012448-appb-img-000024
Is outputted as a = {1, 0, 0, 0, 0, 1, 1}, and b = {1, 1, 0, 0, 0, 1, 1} with 7 training data,
Figure PCTKR2020012448-appb-img-000025
Is 0.428,
Figure PCTKR2020012448-appb-img-000026
Is calculated as 0.571, and according to Equation 2
Figure PCTKR2020012448-appb-img-000027
Can be calculated as 0.75.
수학식 2에 기초하여 산출된 상관 계수가 시각화된 일례는 이하 첨부되는 도 5와 같다.An example in which the correlation coefficient calculated based on Equation 2 is visualized is shown in FIG. 5 attached below.
수학식 1에 기초하여 상악 손실 함수 및 하악 손실 함수는 수학식 3 및 수학식 4와 같이 산출될 수 있다.Based on Equation 1, the maxillary loss function and the mandibular loss function may be calculated as shown in Equations 3 and 4.
Figure PCTKR2020012448-appb-img-000028
Figure PCTKR2020012448-appb-img-000028
Figure PCTKR2020012448-appb-img-000029
는 상악 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000030
는 상악 치아의 병변 발생 여부에 대한 판독 모델의 예측 값,
Figure PCTKR2020012448-appb-img-000031
는 상악 치아에 대한 상관 파라미터를 나타낸다.
Figure PCTKR2020012448-appb-img-000029
Is a value indicating whether a lesion of the maxillary tooth has occurred,
Figure PCTKR2020012448-appb-img-000030
Is the predicted value of the readout model for the occurrence of lesions of the maxillary teeth,
Figure PCTKR2020012448-appb-img-000031
Represents the correlation parameter for the maxillary teeth.
Figure PCTKR2020012448-appb-img-000032
Figure PCTKR2020012448-appb-img-000032
Figure PCTKR2020012448-appb-img-000033
는 하악 치아의 병변 발생 여부를 나타내는 값,
Figure PCTKR2020012448-appb-img-000034
는 하악 치아의 병변 발생 여부에 대한 판독 모델의 예측 값,
Figure PCTKR2020012448-appb-img-000035
는 하악 치아에 대한 상관 파라미터를 나타낸다.
Figure PCTKR2020012448-appb-img-000033
Is a value indicating whether a lesion of the mandibular tooth has occurred,
Figure PCTKR2020012448-appb-img-000034
Is the predicted value of the readout model for the occurrence of lesion of the mandibular tooth,
Figure PCTKR2020012448-appb-img-000035
Represents the correlation parameter for the mandibular tooth.
수학식 3 및 수학식 4에 기반하여, 판독 모델의 학습에 사용되는 최종 손실 함수는 수학식 5와 같이 산출될 수 있다.Based on Equation 3 and Equation 4, the final loss function used for learning the readout model may be calculated as in Equation 5.
Figure PCTKR2020012448-appb-img-000036
Figure PCTKR2020012448-appb-img-000036
Figure PCTKR2020012448-appb-img-000037
는 크로스 엔트로피 함수,
Figure PCTKR2020012448-appb-img-000038
는 조절 가능한 파라미터를 의미할 수 있다.
Figure PCTKR2020012448-appb-img-000037
Is the cross entropy function,
Figure PCTKR2020012448-appb-img-000038
May mean an adjustable parameter.
최종 손실 함수는 상악 손실 함수 및 하악 손실 함수에 기초하여 결정되고, 상악 손실 함수 및 하악 손실 함수는 각각에 포함되는 치아들 사이의 상관 관계에 기초하여 결정되는
Figure PCTKR2020012448-appb-img-000039
Figure PCTKR2020012448-appb-img-000040
에 기초하여 결정된다. 실제로 치아 병변의 경우 인접한 치아들 사이에 함께 발생되는 경우가 많으므로, 병변 발생과 관련하여 인접한 치아들 사이에 상관 관계는 매우 높을 수 있다. 최종 손실 함수는 상악 및 하악에 포함되는 치아들 사이의 상관 관계가 반영되어 있으므로, 이에 기반하여 학습된 판독 모델의 병변 영역 예측에 대한 정확도는 상관 관계를 고려하지 않는 손실 함수를 이용하는 경우에 비해 향상될 수 있다. 보다 구체적으로, 치아들 사이의 상관 관계를 고려하지 않는 손실 함수에 기초하여 판독 모델이 학습되는 경우, 병변이 48번 치아에 발생하였으나, 47번 치아에 병변이 발생한 것으로 예측한 경우와 41번 치아에 병변이 발생한 것으로 예측하는 경우 오류는 동일하게 측정될 수 있다. 하지만, 수학식 5에 따른 최종 손실 함수에 기반하여 학습이 진행되는 경우, 치아들 사이의 상관 관계가 학습 과정에서 반영되어 있기 때문에, 상기 상황에서 41번 치아로 잘못 예측한 경우에 더 큰 패널티가 부여될 수 있고, 이를 통해 보다 예측 정확도가 높은 판독 모델이 생성될 수 있다.
The final loss function is determined based on the maxillary loss function and the mandibular loss function, and the maxillary loss function and the mandibular loss function are determined based on the correlation between the teeth included in each.
Figure PCTKR2020012448-appb-img-000039
And
Figure PCTKR2020012448-appb-img-000040
It is determined on the basis of. In fact, since dental lesions often occur together between adjacent teeth, the correlation between adjacent teeth may be very high in relation to the occurrence of the lesion. Since the final loss function reflects the correlation between the teeth included in the maxilla and mandible, the accuracy of the prediction of the lesion area of the read model learned based on this is improved compared to the case of using the loss function that does not consider the correlation. Can be. More specifically, when the readout model is learned based on the loss function that does not take into account the correlation between teeth, the lesion occurred on tooth 48, but predicted that the lesion occurred on tooth 47 and tooth 41 The error can be measured in the same way when predicting that a lesion has occurred. However, when learning is performed based on the final loss function according to Equation 5, since the correlation between teeth is reflected in the learning process, a larger penalty is incurred in the case of incorrectly predicting tooth 41 in the above situation. May be given, and through this, a readout model with higher prediction accuracy may be generated.
컴퓨팅 장치는 단계(S300)에서 상기 검출된 병변 영역에 대한 치아 병변 정보를 생성하거나 컴퓨팅 장치에 연동되는 타 장치로 하여금 치아 병변 정보를 생성하도록 지원할 수 있다. 치아 병변 정보는 치아 영상에서 병변 영역이 시각화된 시각화 정보 및 개별 치아의 식별 정보와 병변 발생 여부가 매칭된 매칭 정보 중 적어도 하나에 의해 결정될 수 있다.The computing device may generate tooth lesion information on the detected lesion area in step S300 or support another device interlocked with the computing device to generate tooth lesion information. The tooth lesion information may be determined by at least one of visualization information in which a lesion area is visualized in a tooth image, and matching information in which identification information of individual teeth and whether or not a lesion occurs.
일 실시예에 따르면, 컴퓨팅 장치는 단계(S200)의 검출 결과에 기반하여, 치아 영상에서 검출된 병변 영역을 시각화하는 맵을 생성할 수 있다. 컴퓨팅 장치는 판독 모델의 마지막 레이어의 출력에 기초하여, 치아 영상의 영역 별 가중치를 결정하고, 결정된 가중치에 따라 시각적 요소를 달리 표현하는 맵을 생성할 수 있다. 예를 들어, 컴퓨팅 장치는 최종 레이어의 출력 중 높은 가중치에 대응되는 영역으로써 병변 영역에 해당할 확률이 높은 영역은 적색에 가까운 색으로 표현하고, 가중치가 낮아 병변 영역에 해당할 확률이 낮은 영역을 녹색에 가까운 영역으로 표현하는 히트맵을 생성할 수 있다. 컴퓨팅 장치는 생성된 히트맵을 입력 치아 영상에 적용하여 병변 영역에 해당될 가능성이 높은 영역은 적색에 가까운 색상으로 시각화하고, 병변 영역에 해당될 가능성이 낮은 영역은 녹색에 가까운 색상으로 시각함으로써, 치아 영상에 대한 시각화 정보를 생성할 수 있다. 시각화 정보를 생성하기 위한 맵은 예시로 든 히트맵 및 색상 구성에 한정되지 않고, 각 영역이 병변 영역에 해당될 확률을 시각화하는 임의의 방식으로 구현될 수 있음은 통상의 기술자가 용이하게 이해할 수 있을 것이다. 예를 들어, 컴퓨팅 장치는 입력 치아 영상에서 병변 영역에 해당할 확률이 높은 영역은 어두운 명암으로 시각화하고, 병변 영역에 해당할 확률이 낮은 영역은 밝은 명암으로 시각화함으로써 입력 치아 영상의 병변 영역을 시각화할 수 있다. According to an embodiment, the computing device may generate a map for visualizing the lesion area detected in the tooth image based on the detection result in step S200. The computing device may determine a weight for each region of the tooth image based on the output of the last layer of the read model, and generate a map representing different visual elements according to the determined weight. For example, the computing device represents an area corresponding to a high weight in the output of the final layer, and an area with a high probability of corresponding to a lesion area is expressed in a color close to red, and an area with a low probability of corresponding to a lesion area due to a low weight is expressed. It is possible to create a heat map that is represented by an area close to green. The computing device applies the generated heat map to the input tooth image and visualizes the region with a high probability of corresponding to the lesion region with a color close to red, and visualizes the region with a low probability of corresponding to the lesion region with a color close to green. Visualization information for a tooth image can be generated. The map for generating the visualization information is not limited to the heat map and color configuration as an example, and it can be easily understood by those of ordinary skill in the art that each region can be implemented in an arbitrary manner to visualize the probability that it corresponds to the lesion region. There will be. For example, the computing device visualizes the area of the input tooth image with a high probability of corresponding to the lesion area with dark contrast, and the area with the low probability of corresponding to the lesion area with light intensity to visualize the lesion area of the input tooth image. can do.
다른 실시예에 따르면, 컴퓨팅 장치는 치아 영상에 포함되는 개별 치아의 위치 또는 순서를 식별하는 식별 정보를 생성하고, 생성된 식별 정보와 식별 정보에 대응되는 개별 치아에 병변이 발생하였는지 여부에 대한 정보를 매칭한 매칭 정보를 생성할 수 있다. 예를 들어, 식별 정보는 FDI World Dental Federation notation에 따른 개별 치아에 대한 식별 번호일 수 있고, 컴퓨팅 장치는 단계(S200)의 검출 결과에 기초하여 치아에 병변이 발생하였는지 여부에 대한 정보를 해당 치아의 식별 번호에 매칭한 매칭 정보를 생성할 수 있다.According to another embodiment, the computing device generates identification information that identifies the position or sequence of individual teeth included in the tooth image, and information on whether a lesion has occurred in the individual tooth corresponding to the generated identification information and the identification information. It is possible to generate matching information that matches. For example, the identification information may be an identification number for an individual tooth according to the FDI World Dental Federation notation, and the computing device provides information on whether a lesion has occurred in the tooth based on the detection result in step S200. Matching information matching the identification number of can be generated.
컴퓨팅 장치는 단계(S400)에서 치아 병변 정보를 외부 엔티티에 제공할 수 있다.The computing device may provide tooth lesion information to an external entity in step S400.
도 4는 일 실시 예에 따른 병변 정보 제공 방법이 수행되는 일례를 도시하는 도면이다.4 is a diagram illustrating an example in which a method of providing lesion information according to an exemplary embodiment is performed.
컴퓨팅 장치(420)는 입력된 치아 영상(410)의 시각화 정보(430) 및 매칭 정보(440)을 포함하는 치아 병변 정보를 제공할 수 있다. 이를 위하여 컴퓨팅 장치(420)은 미리 학습된 판독 모델을 통해 치아 영상(410)에서 병변 영역을 검출할 수 있다. 컴퓨팅 장치는 검출된 병변 영역을 시각화한 히트맵(431, 432, 433)이 적용된 시각화 정보(430)를 생성하거나, 치아의 식별 정보와 병변 발생 여부에 대한 정보를 매칭한 매칭 정보(440)을 생성하여 외부 엔티티에 제공함으로써 치아 병변 정보를 제공할 수 있다. 히트맵(431, 432, 433)은 병변 영역에 해당할 확률에 따라 서로 다른 색 또는 서로 다른 명암으로 각각의 영역들이 표현될 수 있다. 매칭 정보(440)에 포함된 치아의 식별 번호는 식별 정보(441)에 도시되는 FDI World Dental Federation notation에 따라 결정될 수 있다.The computing device 420 may provide tooth lesion information including visualization information 430 and matching information 440 of the input tooth image 410. To this end, the computing device 420 may detect a lesion area in the tooth image 410 through a pre-learned reading model. The computing device generates the visualization information 430 to which the heat maps 431, 432, 433 that visualize the detected lesion area is applied, or the matching information 440 that matches the identification information of the tooth and the information on whether the lesion has occurred. It is possible to provide tooth lesion information by creating and providing it to an external entity. Each of the heat maps 431, 432, and 433 may be expressed in different colors or different shades according to a probability corresponding to the lesion area. The identification number of the tooth included in the matching information 440 may be determined according to the FDI World Dental Federation notation shown in the identification information 441.
도 5는 일 실시예에 따른 피어슨 상관 행렬이 시각화된 일례를 도시하는 도면이다.5 is a diagram illustrating an example in which a Pearson correlation matrix is visualized according to an embodiment.
치아에 발생하는 병변들은 인접한 치아들 사이에 연속적으로 발생할 가능성이 높다. 따라서, 치아의 병변 검출과 관련하여, 인접한 치아들은 매우 높은 상관 관계를 가질 수 있다. 본원 발명은 판독 모델에 해당하는 인공 신경망을 학습시키는 과정에서 인접한 치아들 사이의 상관 관계가 반영된 수학식 5에 해당하는 최종 손실 함수를 이용함으로써, 보다 정확도 높은 병변 예측 모델을 생성할 수 있다.Lesions that occur on teeth are more likely to occur continuously between adjacent teeth. Thus, with regard to the detection of a lesion of a tooth, adjacent teeth can have a very high correlation. The present invention can generate a lesion prediction model with higher accuracy by using the final loss function corresponding to Equation 5 in which the correlation between adjacent teeth is reflected in the process of training the artificial neural network corresponding to the reading model.
행렬(510, 520) 각각은 수학식 2에 기초하여 산출된 상악 치아의 피어슨 상관 행렬 및 하악 치아의 피어슨 상관 행렬을 시각화한 일례에 해당한다. 각 행렬의 요소들은 치아들 사이의 상관도에 대응될 수 있다. 보다 짙은 색(또는 짙은 명암)으로 시각화된 요소는 상관도가 높게 산출된 것이고, 보다 옅은 색(또는 옅은 명암)으로 시각화된 요소는 상관도가 낮게 산출된 것에 해당한다. 행렬(510, 520)을 참조하면, 상호 인접한 치아들 사이에 상관도가 높게 산출된 것을 확인할 수 있고, 이는 앞서 언급한 바에 대응된다.Each of the matrices 510 and 520 corresponds to an example of visualizing the Pearson correlation matrix of the maxillary teeth and the Pearson correlation matrix of the mandibular teeth calculated based on Equation (2). Elements of each matrix may correspond to a degree of correlation between teeth. An element visualized with a darker color (or darker shade) corresponds to a higher correlation, and an element visualized with a lighter color (or lighter shade) corresponds to a lower correlation. Referring to the matrices 510 and 520, it can be seen that the correlation between teeth adjacent to each other is calculated high, which corresponds to the above-mentioned.
위 실시 예의 설명에 기초하여 해당 기술분야의 통상의 기술자는, 본 발명의 방법 및/또는 프로세스들, 그리고 그 단계들이 하드웨어, 소프트웨어 또는 특정 용례에 적합한 하드웨어 및 소프트웨어의 임의의 조합으로 실현될 수 있다는 점을 명확하게 이해할 수 있다. 상기 하드웨어는 범용 컴퓨터 및/또는 전용 컴퓨팅 장치 또는 특정 컴퓨팅 장치 또는 특정 컴퓨팅 장치의 특별한 모습 또는 구성요소를 포함할 수 있다. 상기 프로세스들은 내부 및/또는 외부 메모리를 가지는, 하나 이상의 마이크로프로세서, 마이크로컨트롤러, 임베디드 마이크로컨트롤러, 프로그래머블 디지털 신호 프로세서 또는 기타 프로그래머블 장치에 의하여 실현될 수 있다. 게다가, 혹은 대안으로서, 상기 프로세스들은 주문형 집적회로(application specific integrated circuit; ASIC), 프로그래머블 게이트 어레이(programmable gate array), 프로그래머블 어레이 로직(Programmable Array Logic; PAL) 또는 전자 신호들을 처리하기 위해 구성될 수 있는 임의의 다른 장치 또는 장치들의 조합으로 실시될 수 있다. 더욱이 본 발명의 기술적 해법의 대상물 또는 선행 기술들에 기여하는 부분들은 다양한 컴퓨터 구성요소를 통하여 수행될 수 있는 프로그램 명령어의 형태로 구현되어 기계 판독 가능한 기록 매체에 기록될 수 있다. 상기 기계 판독 가능한 기록 매체는 프로그램 명령어, 데이터 파일, 데이터 구조 등을 단독으로 또는 조합하여 포함할 수 있다. 상기 기계 판독 가능한 기록 매체에 기록되는 프로그램 명령어는 본 발명을 위하여 특별히 설계되고 구성된 것들이거나 컴퓨터 소프트웨어 분야의 통상의 기술자에게 공지되어 사용 가능한 것일 수도 있다. 기계 판독 가능한 기록 매체의 예에는, 하드 디스크, 플로피 디스크 및 자기 테이프와 같은 자기 매체, CD-ROM, DVD, Blu-ray와 같은 광기록 매체, 플롭티컬 디스크(floptical disk)와 같은 자기-광 매체(magneto-optical media), 및 ROM, RAM, 플래시 메모리 등과 같은 프로그램 명령어를 저장하고 수행하도록 특별히 구성된 하드웨어 장치가 포함된다. 프로그램 명령어의 예에는, 전술한 장치들 중 어느 하나뿐만 아니라 프로세서, 프로세서 아키텍처 또는 상이한 하드웨어 및 소프트웨어의 조합들의 이종 조합, 또는 다른 어떤 프로그램 명령어들을 실행할 수 있는 기계 상에서 실행되기 위하여 저장 및 컴파일 또는 인터프리트될 수 있는, C와 같은 구조적 프로그래밍 언어, C++ 같은 객체지향적 프로그래밍 언어 또는 고급 또는 저급 프로그래밍 언어(어셈블리어, 하드웨어 기술 언어들 및 데이터베이스 프로그래밍 언어 및 기술들)를 사용하여 만들어질 수 있는바, 기계어 코드, 바이트코드뿐만 아니라 인터프리터 등을 사용해서 컴퓨터에 의해서 실행될 수 있는 고급 언어 코드도 이에 포함된다. Based on the description of the above embodiment, a person skilled in the art will know that the method and/or processes of the present invention, and the steps thereof, can be implemented in hardware, software, or any combination of hardware and software suitable for a specific application. Can understand the point clearly. The hardware may include a general-purpose computer and/or a dedicated computing device, or a specific computing device or special features or components of a specific computing device. The processes may be realized by one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices, with internal and/or external memory. In addition, or as an alternative, the processes can be configured to process application specific integrated circuits (ASICs), programmable gate arrays, programmable array logic (PAL) or electronic signals. May be implemented with any other device or combination of devices. Furthermore, the objects of the technical solution of the present invention or parts contributing to the prior arts may be implemented in the form of program instructions that can be executed through various computer components and recorded in a machine-readable recording medium. The machine-readable recording medium may include program instructions, data files, data structures, etc. alone or in combination. The program instructions recorded on the machine-readable recording medium may be specially designed and configured for the present invention, or may be known to and usable by a person skilled in the computer software field. Examples of machine-readable recording media include magnetic media such as hard disks, floppy disks and magnetic tapes, optical recording media such as CD-ROM, DVD, Blu-ray, and magnetic-optical media such as floptical disks. (magneto-optical media), and a hardware device specially configured to store and execute program instructions such as ROM, RAM, flash memory, and the like. Examples of program instructions include a processor, a processor architecture, or a heterogeneous combination of different hardware and software combinations, as well as any one of the aforementioned devices, or storing and compiling or interpreting to be executed on a machine capable of executing any other program instructions. Can be created using a structured programming language such as C, an object-oriented programming language such as C++, or a high-level or low-level programming language (assembly, hardware description languages and database programming languages and technologies), machine code, This includes not only bytecode but also high-level language code that can be executed by a computer using an interpreter or the like.
따라서 본 발명에 따른 일 태양에서는, 앞서 설명된 방법 및 그 조합들이 하나 이상의 컴퓨팅 장치들에 의하여 수행될 때, 그 방법 및 방법의 조합들이 각 단계들을 수행하는 실행 가능한 코드로서 실시될 수 있다. 다른 일 태양에서는, 상기 방법은 상기 단계들을 수행하는 시스템들로서 실시될 수 있고, 방법들은 장치들에 걸쳐 여러 가지 방법으로 분산되거나 모든 기능들이 하나의 전용, 독립형 장치 또는 다른 하드웨어에 통합될 수 있다. 또 다른 일 태양에서는, 위에서 설명한 프로세스들과 연관된 단계들을 수행하는 수단들은 앞서 설명한 임의의 하드웨어 및/또는 소프트웨어를 포함할 수 있다. 그러한 모든 순차 결합 및 조합들은 본 개시서의 범위 내에 속하도록 의도된 것이다.Thus, in an aspect according to the present invention, when the above-described method and combinations thereof are performed by one or more computing devices, the method and combinations of methods may be implemented as executable code that performs the respective steps. In another aspect, the method may be implemented as systems that perform the steps, and the methods may be distributed in several ways across devices or all functions may be integrated into one dedicated, standalone device or other hardware. In yet another aspect, the means for performing the steps associated with the processes described above may include any hardware and/or software described above. All such sequential combinations and combinations are intended to be within the scope of this disclosure.
예를 들어, 상기 하드웨어 장치는 본 발명에 따른 처리를 수행하기 위해 하나 이상의 소프트웨어 모듈로서 작동하도록 구성될 수 있으며, 그 역도 마찬가지이다. 상기 하드웨어 장치는, 프로그램 명령어를 저장하기 위한 ROM/RAM 등과 같은 메모리와 결합되고 상기 메모리에 저장된 명령어들을 실행하도록 구성되는 MPU, CPU, GPU, TPU와 같은 프로세서를 포함할 수 있으며, 외부 장치와 신호를 주고 받을 수 있는 통신부를 포함할 수 있다. 덧붙여, 상기 하드웨어 장치는 개발자들에 의하여 작성된 명령어들을 전달받기 위한 키보드, 마우스, 기타 외부 입력장치를 포함할 수 있다.For example, the hardware device may be configured to operate as one or more software modules to perform the processing according to the present invention, and vice versa. The hardware device may include a processor such as an MPU, CPU, GPU, or TPU that is combined with a memory such as ROM/RAM for storing program instructions and configured to execute instructions stored in the memory, and external devices and signals It may include a communication unit that can send and receive. In addition, the hardware device may include a keyboard, a mouse, and other external input devices for receiving commands written by developers.
이상에서 본 발명이 구체적인 구성요소 등과 같은 특정 사항들과 한정된 실시 예 및 도면에 의해 설명되었으나, 이는 본 발명의 보다 전반적인 이해를 돕기 위해서 제공된 것일 뿐, 본 발명이 상기 실시 예들에 한정되는 것은 아니며, 본 발명이 속하는 기술분야에서 통상적인 지식을 가진 사람이라면 이러한 기재로부터 다양한 수정 및 변형을 꾀할 수 있다.In the above, the present invention has been described by specific matters such as specific components and limited embodiments and drawings, but this is provided only to help a more general understanding of the present invention, and the present invention is not limited to the above embodiments, Anyone with ordinary knowledge in the technical field to which the present invention pertains can make various modifications and variations from these descriptions.
따라서, 본 발명의 사상은 상기 설명된 실시 예에 국한되어 정해져서는 아니되며, 후술하는 특허청구범위뿐만 아니라 이 특허청구범위와 균등하게 또는 등가적으로 변형된 모든 것들은 본 발명의 사상의 범주에 속한다고 할 것이다.Therefore, the spirit of the present invention is limited to the above-described embodiments and should not be defined, and all modifications that are equally or equivalent to the claims as well as the claims to be described later fall within the scope of the spirit of the present invention. I would say.
그와 같이 균등하게 또는 등가적으로 변형된 것에는, 예컨대 본 발명에 따른 방법을 실시한 것과 동일한 결과를 낼 수 있는, 논리적으로 동치(logically equivalent)인 방법이 포함될 것인바, 본 발명의 진의 및 범위는 전술한 예시들에 의하여 제한되어서는 아니되며, 법률에 의하여 허용 가능한 가장 넓은 의미로 이해되어야 한다.Such equivalently or equivalently modified ones will include, for example, a logically equivalent method capable of producing the same result as that of carrying out the method according to the present invention, the true meaning and scope of the present invention. Is not to be limited by the above-described examples, and should be understood in the broadest possible meaning by law.

Claims (8)

  1. 컴퓨팅 장치에 의해 수행되는 치아 병변 정보 제공 방법에 있어서,In the method of providing tooth lesion information performed by a computing device,
    피검자의 치아 영상을 수신하는 단계;Receiving an image of the subject's teeth;
    미리 학습된 판독 모델을 통해 상기 치아 영상에 포함되는 병변 영역을 검출하는 단계;Detecting a lesion region included in the tooth image through a pre-learned read model;
    상기 병변 영역에 대한 치아 병변 정보를 생성하는 단계; 및Generating tooth lesion information for the lesion area; And
    병변 정보를 외부 엔티티에 제공하는 단계Providing lesion information to an external entity
    를 포함하는, 치아 병변 정보 제공 방법.Containing, dental lesion information providing method.
  2. 제1항에 있어서,The method of claim 1,
    상기 생성하는 단계는,The generating step,
    상기 치아 영상에서 병변 영역을 시각화하는 맵을 생성하는 단계; 및Generating a map for visualizing a lesion area in the tooth image; And
    상기 맵에 기초하여 상기 치아 병변 정보를 생성하는 단계Generating the tooth lesion information based on the map
    를 포함하는, 치아 병변 정보 제공 방법.Containing, dental lesion information providing method.
  3. 제1항에 있어서,The method of claim 1,
    상기 생성하는 단계는,The generating step,
    상기 치아 영상에 포함되는 개별 치아의 위치 또는 순서를 식별하는 식별 정보를 생성하는 단계;Generating identification information for identifying the position or order of individual teeth included in the tooth image;
    상기 식별 정보와 상기 식별 정보에 대응되는 개별 치아에 병변이 발생하였는지 여부에 대한 정보를 매칭한 매칭 정보를 생성하는 단계; 및Generating matching information by matching information on whether a lesion has occurred in the individual tooth corresponding to the identification information and the identification information; And
    상기 매칭 정보에 기초하여 상기 치아 병변 정보를 생성하는 단계Generating the tooth lesion information based on the matching information
    를 포함하는, 치아 병변 정보 제공 방법.Containing, dental lesion information providing method.
  4. 제1항에 있어서,The method of claim 1,
    상기 판독 모델은,The reading model,
    손실 함수의 결과를 최소화하는 방향으로 미리 학습되고,It is learned in advance in the direction of minimizing the result of the loss function,
    상기 손실 함수는,The loss function is,
    개별 치아들 사이의 상관 관계 정보에 기초하여 결정되는, 치아 병변 정보 제공 방법.A method of providing tooth lesion information, which is determined based on correlation information between individual teeth.
  5. 제4항에 있어서,The method of claim 4,
    상기 손실 함수는,The loss function is,
    개별 치아들에 대한 상관 파라미터와 상기 판독 모델의 예측 결과 사이의 연산에 기초하여 산출되는 개별 손실 함수를 포함하는, 치아 병변 제공 방법.A method for providing a dental lesion comprising an individual loss function calculated based on an operation between a correlation parameter for individual teeth and a prediction result of the readout model.
  6. 제5항에 있어서,The method of claim 5,
    상기 상관 파라미터는,The correlation parameter,
    학습 데이터에 포함된 개별 치아들의 병변 발생 여부에 대한 정보들 사이의 연산에 기초하여 산출되는, 치아 병변 정보 제공 방법.A method of providing tooth lesion information that is calculated based on an operation between information on whether or not lesions of individual teeth included in the learning data have occurred.
  7. 컴퓨팅 장치로 하여금, 제1항의 방법을 수행하도록 구현된 명령어(instructions)를 포함하는, 기계 판독 가능한 비일시적 기록 매체.A machine-readable non-transitory recording medium comprising instructions embodied to cause a computing device to perform the method of claim 1.
  8. 치아 병변 정보를 제공하는 컴퓨팅 장치에 있어서,In the computing device for providing tooth lesion information,
    피검자의 치아 영상을 수신하는 통신부; 및A communication unit for receiving an image of a subject's teeth; And
    상기 치아 영상에 대한 치아 병변 정보를 생성하는 프로세서Processor for generating tooth lesion information for the tooth image
    를 포함하고,Including,
    상기 프로세서는,The processor,
    미리 학습된 판독 모델을 통해 상기 치아 영상에 포함되는 병변 영역에 대한 치아 병변 정보를 생성하고,Generates tooth lesion information for a lesion region included in the tooth image through a pre-learned reading model,
    상기 치아 병변 정보를 외부 엔티티에 제공하는, 컴퓨팅 장치.Providing the tooth lesion information to an external entity.
PCT/KR2020/012448 2019-09-18 2020-09-15 Method for providing tooth lesion information, and device using same WO2021054700A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2019-0114578 2019-09-18
KR1020190114578A KR102186709B1 (en) 2019-09-18 2019-09-18 Method for providing tooth leison information and apparatus using the same

Publications (1)

Publication Number Publication Date
WO2021054700A1 true WO2021054700A1 (en) 2021-03-25

Family

ID=73776712

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2020/012448 WO2021054700A1 (en) 2019-09-18 2020-09-15 Method for providing tooth lesion information, and device using same

Country Status (2)

Country Link
KR (1) KR102186709B1 (en)
WO (1) WO2021054700A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102568247B1 (en) * 2021-05-17 2023-08-21 오스템임플란트 주식회사 Method and apparatus for simulatiing of teeth treatment, computer-readable storage medium and computer program
CN115100210B (en) * 2022-08-29 2022-11-18 山东艾克赛尔机械制造有限公司 Anti-counterfeiting identification method based on automobile parts

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140091176A (en) * 2013-01-10 2014-07-21 삼성전자주식회사 Apparatus and method for lesion diagnosis
KR20160112493A (en) * 2015-03-19 2016-09-28 (주)바텍이우홀딩스 Effective oral lesion onset detection method and system in Computed Tomography
KR20180045551A (en) * 2016-10-26 2018-05-04 고려대학교 산학협력단 System and system for diagnosis of oral lesion
KR101943011B1 (en) * 2018-01-22 2019-01-28 주식회사 뷰노 Method for facilitating medical image reading and apparatus using the same

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011140663A1 (en) * 2010-05-13 2011-11-17 Quantum Dental Technologies Inc. Method of processing and displaying oral health diagnostic data
KR101903424B1 (en) 2017-01-10 2018-11-13 한국광기술원 Three dimensions intraoral scanner based on optical coherence tomography and diagnosing method of dental condition using same

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140091176A (en) * 2013-01-10 2014-07-21 삼성전자주식회사 Apparatus and method for lesion diagnosis
KR20160112493A (en) * 2015-03-19 2016-09-28 (주)바텍이우홀딩스 Effective oral lesion onset detection method and system in Computed Tomography
KR20180045551A (en) * 2016-10-26 2018-05-04 고려대학교 산학협력단 System and system for diagnosis of oral lesion
KR101943011B1 (en) * 2018-01-22 2019-01-28 주식회사 뷰노 Method for facilitating medical image reading and apparatus using the same

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZANELLA-CALZADA LAURA, GALVAN-TEJADA CARLOS, CHAVEZ-LAMAS NUBIA, RIVAS-GUTIERREZ JESUS, MAGALLANES-QUINTANAR RAFAEL, CELAYA-PADILL: "Deep Artificial Neural Networks for the Diagnostic of Caries Using Socioeconomic and Nutritional Features as Determinants: Data from NHANES 2013–2014", BIOENGINEERING, vol. 5, no. 2, 47, 18 June 2018 (2018-06-18), pages 1 - 20, XP055793073, ISSN: 2306-5354, DOI: 10.3390/bioengineering5020047 *

Also Published As

Publication number Publication date
KR102186709B1 (en) 2020-12-04

Similar Documents

Publication Publication Date Title
WO2019103440A1 (en) Method for supporting reading of medical image of subject and device using same
KR101898575B1 (en) Method for predicting future state of progressive lesion and apparatus using the same
WO2019143177A1 (en) Method for reconstructing series of slice images and apparatus using same
WO2019208848A1 (en) Three-dimensional eyeball movement measurement method and automatic deep learning based dizziness diagnosis system
WO2021034138A1 (en) Dementia evaluation method and apparatus using same
WO2021054700A1 (en) Method for providing tooth lesion information, and device using same
WO2019143021A1 (en) Method for supporting viewing of images and apparatus using same
WO2019143179A1 (en) Method for automatically detecting same regions of interest between images of same object taken with temporal interval, and apparatus using same
WO2019098415A1 (en) Method for determining whether subject has developed cervical cancer, and device using same
WO2019231104A1 (en) Method for classifying images by using deep neural network and apparatus using same
WO2022131642A1 (en) Apparatus and method for determining disease severity on basis of medical images
JP2024107095A (en) Estimation device, learning model, learning model generation method, and computer program
WO2019124836A1 (en) Method for mapping region of interest of first medical image onto second medical image, and device using same
KR102290799B1 (en) Method for providing tooth leison information and apparatus using the same
WO2022231329A1 (en) Method and device for displaying bio-image tissue
CN113516639B (en) Training method and device for oral cavity abnormality detection model based on panoramic X-ray film
Zhao et al. Recognition and segmentation of teeth and mandibular nerve canals in panoramic dental X-rays by Mask RCNN
WO2019164277A1 (en) Method and device for evaluating bleeding by using surgical image
WO2021145607A1 (en) Dental medical record device and dental medical record method thereof
Xiong et al. Simultaneous detection of dental caries and fissure sealant in intraoral photos by deep learning: a pilot study
KR102392312B1 (en) Apparatus and method for dental medical record
WO2017010612A1 (en) System and method for predicting pathological diagnosis on basis of medical image analysis
WO2022190891A1 (en) Information processing system and information processing method
WO2019164273A1 (en) Method and device for predicting surgery time on basis of surgery image
WO2023027248A1 (en) Data generation method, and training method and apparatus using same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20865841

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20865841

Country of ref document: EP

Kind code of ref document: A1