WO2021201272A1 - 医療画像処理装置、内視鏡システム及び医療画像処理装置の作動方法並びに医療画像処理装置用プログラム - Google Patents
医療画像処理装置、内視鏡システム及び医療画像処理装置の作動方法並びに医療画像処理装置用プログラム Download PDFInfo
- Publication number
- WO2021201272A1 WO2021201272A1 PCT/JP2021/014299 JP2021014299W WO2021201272A1 WO 2021201272 A1 WO2021201272 A1 WO 2021201272A1 JP 2021014299 W JP2021014299 W JP 2021014299W WO 2021201272 A1 WO2021201272 A1 WO 2021201272A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- medical image
- highlighting
- image processing
- visibility
- image
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/045—Control thereof
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
- A61B1/000094—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope extracting biological structures
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00043—Operational features of endoscopes provided with output arrangements
- A61B1/00045—Display arrangement
- A61B1/0005—Display arrangement combining images e.g. side-by-side, superimposed or tiled
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/30—Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/361—Image-producing devices, e.g. surgical cameras
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00043—Operational features of endoscopes provided with output arrangements
- A61B1/00055—Operational features of endoscopes provided with output arrangements for alerting the user
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00022—Sensing or detecting at the treatment site
- A61B2017/00057—Light
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00199—Electrical control of surgical instruments with a console, e.g. a control panel with a display
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/373—Surgical systems with images on a monitor during operation using light, e.g. by using optical scanners
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10068—Endoscopic image
Definitions
- the present invention relates to a medical image processing device for detecting a region of interest such as a lesion, an operation method of an endoscopic system and a medical image processing device, and a program for the medical image processing device.
- medical images such as endoscopic images, X-ray images, CT (Computed Tomography) images, and MR (Magnetic Resonanse) images are used to diagnose the patient's medical condition and perform diagnostic imaging such as follow-up. ing. Based on such diagnostic imaging, doctors and the like make decisions on treatment policies.
- Patent Documents 1 and 2 describe a medical image processing apparatus that performs image processing based on the detection information when a region of interest such as a lesion is detected from a medical image.
- a highlighting process for superimposing a highlighting for emphasizing a region of interest on a medical image is performed.
- Patent Documents 1 and 2 since the medical image processing apparatus described in Patent Documents 1 and 2 does not consider the visibility of highlighting, it depends on the color of the subject in the medical image, the presence or absence of an object existing in the subject, and the like. , The highlighting may be assimilated with the surroundings or may not be noticeable to the surrounding parts. Such reduced visibility of highlighting may cause the physician to go unnoticed in the area of interest.
- An object of the present invention is to provide a medical image processing device, an endoscopic system, an operation method of the medical image processing device, and a program for the medical image processing device capable of recognizing a decrease in visibility of highlighting.
- the present invention is a medical image processing apparatus including a processor, in which the processor acquires a medical image, detects a region of interest in the medical image, sets highlighting for highlighting the detected region of interest, and highlights the region of interest.
- the display is superimposed on the medical image, and the visibility of the highlighting is judged from the image information acquired from the medical image in which the region of interest is detected and the set highlighting, and the judgment result of the visibility is given to the user. Notify.
- the processor preferably acquires image information from inside the highlighting in the medical image.
- the processor preferably acquires image information from outside the highlighting in the medical image.
- the processor acquires the color difference between the medical image and the highlighting from the color information calculated from the image information and the color information calculated from the highlighting, and determines the visibility from the color difference.
- the processor preferably calculates the average value calculated from the image information as the color information.
- the processor displays a frame-shaped figure surrounding the region of interest as a highlight and determines visibility from the thickness of the line of the frame shape with respect to the region of interest. Further, it is preferable that the processor displays a frame-shaped figure surrounding the region of interest as highlighting and determines the visibility from the similarity of the frame shape with respect to the region of interest.
- the processor preferably displays the determination result on the display screen. Further, it is preferable that the processor calculates a quantified index value as a visibility determination result and displays it on the display screen, and the processor preferably displays the index value as a notification when the index value is equal to or less than a preset threshold value. It is more preferable to display.
- the processor preferably uses the color difference calculated from the image information and highlighting as an index value. Further, the processor may calculate an index value obtained by quantifying the visibility determination result and display the identification information or the identification figure according to the index value.
- the processor determines the visibility based on the presence or absence of an object not to be detected inside the highlighting.
- the processor does not detect an object when the area ratio of the portion where the brightness or brightness inside the highlighting is equal to or greater than the second threshold value is equal to or greater than the third threshold value with respect to the range inside the highlighting in the medical image. It is more preferable to determine that there is.
- the processor may display the determination result on a display screen different from the display screen for displaying the medical image.
- the processor preferably automatically saves the medical image in which the region of interest is detected, and the processor preferably has the index value equal to or less than the preset first threshold value. In the case of, it is preferable to warn the user.
- the endoscope system of the present invention includes a light source device, an endoscope, a processor, and a display unit.
- the processor acquires the medical image, detects the area of interest in the medical image, sets the highlighting for highlighting the detected area of interest, superimposes the highlighting on the medical image, displays it on the monitor, and displays the area of interest on the monitor.
- the visibility of the highlighting is determined from the image information acquired from the detected medical image and the set highlighting, and the determination result of the visibility is notified to the user.
- the light source device emits illumination light for illuminating the observation target.
- the endoscope has an imaging sensor that captures an observation object illuminated by illumination light.
- the monitor displays a medical image obtained by signal-processing the image signal output by the image sensor.
- a step of acquiring a medical image, a step of detecting a region of interest in the acquired medical image, and a highlighting for highlighting the detected region of interest are set.
- a step of superimposing the highlighting on the medical image, a step of determining the visibility of the highlighting from the image information acquired from the medical image in which the region of interest is detected, and the set highlighting, and the visibility It includes a step of notifying the user of the determination result.
- the medical image processing device program of the present invention is installed in a medical image processing device that acquires a medical image and performs image processing on the medical image, and has a function of acquiring the medical image on a computer and a region of interest from within the medical image.
- a function of determining the visibility of the highlighting from the set highlighting and a function of notifying the user of the determination result of the visibility are realized.
- the user can recognize the decrease in visibility of highlighting.
- This is an example of a display screen when the display control unit highlights the area of interest and displays notification information.
- It is a flowchart which shows the series flow of attention area detection mode.
- It is explanatory drawing which shows the display state, and is explanatory drawing which shows the example (B) which detects the lesion part from the endoscopic image, and superimposes the figure as a highlight on the endoscopic image.
- the endoscope system 10 includes an endoscope 12, a light source device 14, a processor device 16, a monitor 18 (display unit), and a console 19.
- the endoscope 12 is optically connected to the light source device 14 and electrically connected to the processor device 16.
- the endoscope 12 has an insertion portion 12a to be inserted into the subject, an operation portion 12b provided at the base end portion of the insertion portion 12a, and a curved portion 12c and a tip portion 12d provided on the tip end side of the insertion portion 12a. doing.
- the angle knob 13a of the operation unit 12b By operating the angle knob 13a of the operation unit 12b, the bending unit 12c bends. This bending motion directs the tip 12d in a desired direction.
- the tip portion 12d has an illumination window, an observation window, an air supply / water supply nozzle, and a forceps outlet on the tip surface (none of which is shown).
- the illumination window is for irradiating the observation site with illumination light.
- the observation window is for taking in light from the observation site.
- the air supply / water supply nozzle is for cleaning the illumination window and the observation window.
- the forceps outlet is for performing various treatments using a forceps and a treatment tool such as an electric knife.
- the operation unit 12b includes a still image acquisition unit 13b used for still image acquisition operation, a mode switching unit 13c used for observation mode switching operation, and a zoom operation unit 13d used for zoom magnification changing operation. Is provided.
- the still image acquisition unit 13b can perform a freeze operation for displaying the still image to be observed on the monitor 18 and a release operation for saving the still image in the storage.
- the endoscope system 10 has a normal mode, a special mode, and a region of interest detection mode as observation modes.
- the observation mode is the normal mode
- the normal light obtained by combining the light of a plurality of colors with the light amount ratio Lc for the normal mode is emitted.
- the observation mode is the special mode
- the special light obtained by combining the light of a plurality of colors with the light amount ratio Ls for the special mode is emitted.
- the illumination light for the attention area detection mode is emitted.
- normal light is emitted as the illumination light for the region of interest detection mode, but special light may be emitted.
- the processor device 16 is electrically connected to the monitor 18 and the console 19.
- the monitor 18 outputs and displays an image to be observed, information incidental to the image, and the like.
- the console 19 functions as a user interface that accepts input operations such as designation of a region of interest (ROI: RegionOfInterest) and function settings.
- ROI region of interest
- the light source device 14 includes a light source unit 20 that emits illumination light used for illuminating an observation target, and a light source control unit 22 that controls the light source unit 20.
- the light source unit 20 is a semiconductor light source such as a multi-color LED (Light Emitting Diode).
- the light source control unit 22 controls the amount of light emitted from the illumination light by turning on / off the LED and the like, and adjusting the drive current and the drive voltage of the LED and the like. Further, the light source control unit 22 controls the wavelength band of the illumination light by changing the optical filter or the like.
- the light source unit 20 includes a V-LED (VioletLightEmittingDiode) 20a, a B-LED (BlueLightEmittingDiode) 20b, a G-LED (GreenLightEmittingDiode) 20c, and an R-LED (Red). It has a 4-color LED of LightEmittingDiode) 20d and a wavelength cut filter 23. As shown in FIG. 3, the V-LED 20a emits purple light V having a wavelength band of 380 nm to 420 nm.
- the B-LED20b emits blue light B having a wavelength band of 420 nm to 500 nm.
- the blue light B emitted from the B-LED 23b at least the wavelength side longer than the peak wavelength of 450 nm is cut by the wavelength cut filter 23.
- the blue light Bx after passing through the wavelength cut filter 23 is in the wavelength range of 420 to 460 nm.
- the light in the wavelength region longer than 460 nm is cut because the light in the wavelength region longer than 460 nm reduces the vascular contrast of the blood vessel to be observed. Because there is.
- the wavelength cut filter 23 may dimming the light in the wavelength region longer than 460 nm instead of cutting the light in the wavelength region longer than 460 nm.
- the G-LED20c emits green light G having a wavelength band of 480 nm to 600 nm.
- the R-LED20d emits red light R having a wavelength band of 600 nm to 650 nm.
- the light emitted from each of the LEDs 20a to 20d may have the same center wavelength and the peak wavelength, or may be different from each other.
- the light source control unit 22 adjusts the emission timing, emission period, light amount, and spectral spectrum of the illumination light by independently controlling the lighting and extinguishing of the LEDs 20a to 20d and the amount of light emitted at the time of lighting.
- the control of turning on and off in the light source control unit 22 is different for each observation mode.
- the reference brightness can be set by the brightness setting unit of the light source device 14, the console 19, or the like.
- the light source control unit 22 lights all the V-LED20a, B-LED20b, G-LED20c, and R-LED20d.
- the peak of the light intensity of the blue light Bx is the purple light V, the green light G.
- red light R are set to be larger than the peak of any of the light intensities.
- the multicolored light for the normal mode or the attention region detection mode including the purple light V, the blue light Bx, the green light G, and the red light R is usually emitted from the light source device 14. As light, is emitted. Normal light is almost white because it has a certain intensity or more from the blue band to the red band.
- the light source control unit 22 lights all the V-LED20a, B-LED20b, G-LED20c, and R-LED20d.
- the light intensity ratio Ls between the purple light V, the blue light B, the green light G, and the red light R has a peak of the light intensity of the purple light V, which is the blue light Bx and the green light G.
- red light R are set to be larger than the peak of any of the light intensities.
- the peaks of the light intensities of the green light G and the red light R are set to be smaller than the peaks of the light intensities of the purple light V and the blue light Bx.
- the light source device 14 emits multicolored light for the special mode including purple light V, blue light Bx, green light G, and red light R as special light.
- the special light is bluish because the proportion of purple light V is large.
- the special light does not have to include light of all four colors, and may include light from at least one of the four color LEDs 20a to 20d. Further, the special light preferably has a main wavelength range of 450 nm or less, for example, a peak wavelength or a center wavelength.
- the illumination light emitted by the light source unit 20 is incident on the light guide 24 inserted into the insertion unit 12a via an optical path coupling portion (not shown) formed by a mirror, a lens, or the like.
- the light guide 24 is built in the endoscope 12 and the universal cord, and propagates the illumination light to the tip portion 12d of the endoscope 12.
- the universal cord is a cord that connects the endoscope 12, the light source device 14, and the processor device 16.
- a multimode fiber can be used as the light guide 24.
- a fine fiber cable having a core diameter of 105 ⁇ m, a clad diameter of 125 ⁇ m, and a diameter of ⁇ 0.3 mm to ⁇ 0.5 mm including a protective layer serving as an outer skin can be used for the light guide 24.
- the tip portion 12d of the endoscope 12 is provided with an illumination optical system 30a and an imaging optical system 30b.
- the illumination optical system 30a has an illumination lens 32.
- the observation target is illuminated by the illumination light propagating through the illumination lens 32 and propagating through the light guide 24.
- the imaging optical system 30b includes an objective lens 34, a magnifying optical system 36, and an imaging sensor 38 (corresponding to the “imaging unit” of the present invention).
- Various types of light such as reflected light, scattered light, and fluorescence from the observation target are incident on the image pickup sensor 38 through the objective lens 34 and the magnifying optical system 36. As a result, an image to be observed is formed on the image sensor 38.
- the magnifying optical system 36 includes a zoom lens 36a that magnifies the observation target and a lens driving unit 36b that moves the zoom lens 36a in the optical axis direction CL.
- the zoom lens 36a is freely moved between the telephoto end and the wide-angle end according to the zoom control by the lens driving unit 36b, thereby enlarging or reducing the observation target imaged on the image sensor 38.
- the image sensor 38 is a color image sensor that captures an observation target irradiated with illumination light.
- Each pixel of the image sensor 38 is provided with any one of an R (red) color filter, a G (green) color filter, and a B (blue) color filter.
- the image pickup sensor 38 receives purple to blue light from the B pixel provided with the B color filter, receives green light from the G pixel provided with the G color filter, and is provided with the R color filter.
- the existing R pixel receives red light.
- the image signals of each RGB color are output from the pixels of each color.
- the image sensor 38 transmits the output image signal to the CDS circuit 40.
- the image sensor 38 In the normal mode or the region of interest detection mode, the image sensor 38 outputs a Bc image signal from the B pixel and outputs a Gc image signal from the G pixel by imaging an observation target illuminated with normal light, and R The Rc image signal is output from the pixels. Further, in the special mode, the image sensor 38 outputs a Bs image signal from the B pixel, outputs a Gs image signal from the G pixel, and Rs from the R pixel by imaging the observation target illuminated with the special light. Output the image signal.
- a CCD (Charge Coupled Device) image sensor, a CMOS (Complementary Metal-Oxide Semiconductor) image sensor, or the like can be used.
- a complementary color imaging sensor provided with complementary color filters of C (cyan), M (magenta), Y (yellow) and G (green) may be used. good.
- the image signals of four colors of CMYG are output. Therefore, by converting the image signals of the four colors of CMYG into the image signals of the three colors of RGB by the complementary color-primary color conversion, it is possible to obtain the image signals of each RGB color similar to the image sensor 38.
- a monochrome sensor without a color filter may be used.
- the CDS circuit 40 performs correlated double sampling (CDS: Correlated Double Sampling) on the analog image signal received from the image sensor 38.
- CDS Correlated Double Sampling
- the image signal that has passed through the CDS circuit 40 is input to the AGC circuit 42.
- the AGC circuit 40 performs automatic gain control (AGC: Automatic Gain Control) on the input image signal.
- a / D (Analog to Digital) conversion circuit 44 converts an analog image signal that has passed through the AGC circuit 42 into a digital image signal.
- the A / D conversion circuit 44 inputs the digital image signal after the A / D conversion to the processor device 16.
- the processor device 16 includes an image signal acquisition unit 50, a DSP (Digital Signal Processor) 52, a noise reduction unit 54, an image processing unit 56, and a display control unit 58.
- DSP Digital Signal Processor
- the processor device 16 includes a function as a medical image processing device, and as will be described later, the image processing unit 56 acquires an endoscopic image and detects a region of interest in the observation target from the endoscopic image. Then, the display control unit 58 highlights the region of interest with respect to the endoscopic image 75.
- the image signal acquisition unit 50 acquires a digital image signal corresponding to the observation mode from the endoscope 12.
- the Bc image signal, the Gc image signal, and the Rc image signal are acquired.
- the Bs image signal, the Gs image signal, and the Rs image signal are acquired.
- one frame of Bc image signal, Gc image signal, and Rc image signal is acquired when the normal light is illuminated, and one frame of Bs image signal, Gs image signal, is acquired when the special light is illuminated. Acquire the Rs image signal.
- the DSP 52 performs various signal processing such as defect correction processing, offset processing, gain correction processing for DSP, linear matrix processing, gamma conversion processing, and demosaic processing on the image signal acquired by the image signal acquisition unit 50.
- the defect correction process corrects the signal of the defective pixel of the image sensor 38.
- the offset processing removes the dark current component from the defect-corrected image signal and sets an accurate zero level.
- the DSP gain correction process adjusts the signal level by multiplying the offset-processed image signal by a specific DSP gain.
- the linear matrix processing enhances the color reproducibility of the image signal that has been gain-corrected for DSP.
- the gamma conversion process adjusts the brightness and saturation of the image signal processed by the linear matrix.
- the gamma-converted image signal is subjected to demosaic processing (also referred to as isotropic processing or simultaneous processing) to generate a signal of a color lacking in each pixel by interpolation. By this demosaic processing, all the pixels have RGB signals of each color.
- the noise reduction unit 54 reduces noise by performing noise reduction processing by, for example, a moving average method, a median filter method, or the like on an image signal that has undergone demosaic processing or the like by DSP 52.
- the image signal after noise reduction is input to the image processing unit 56.
- the image processing unit 56 includes a normal mode image processing unit 60, a special mode image processing unit 62, and a region of interest detection mode image processing unit 64.
- the normal mode image processing unit 60 operates when the normal mode is set, and performs color conversion processing, color enhancement processing, and structure enhancement processing on the received Bc image signal, Gc image signal, and Rc image signal. conduct.
- the RGB image signal is subjected to color conversion processing by 3 ⁇ 3 matrix processing, gradation conversion processing, three-dimensional LUT (Look Up Table) processing, or the like.
- the color enhancement process is performed on the RGB image signal that has undergone the color conversion process.
- the structure enhancement process is a process for emphasizing the structure of the observation target, and is performed on the RGB image signal after the color enhancement process.
- a normal image can be obtained by performing various image processing and the like as described above. Since the normal image is an image obtained based on normal light in which purple light V, blue light Bx, green light G, and red light R are emitted in a well-balanced manner, it is an image having a natural hue.
- the normal image is input to the display control unit 58.
- the special mode image processing unit 62 operates when the special mode is set.
- the special mode image processing unit 62 performs color conversion processing, color enhancement processing, and structure enhancement processing on the received Bs image signal, Gs image signal, and Rs image signal.
- the processing contents of the color conversion processing, the color enhancement processing, and the structure enhancement processing are the same as those of the normal mode image processing unit 60.
- a special image can be obtained by performing various image processing as described above.
- the special image is an image obtained based on special light in which purple light V, which has a high absorption coefficient of hemoglobin in blood vessels, emits a larger amount of light than blue light Bx, green light G, and red light R of other colors. Therefore, the resolution of the vascular structure and the ductal structure is higher than that of other structures.
- the special image is input to the display control unit 58.
- the attention area detection mode image processing unit 64 operates when the attention area detection mode is set. As shown in FIG. 6, the attention area detection mode image processing unit 64 includes a detection image processing unit 70, an attention area detection unit 71, a visibility determination unit 72, and a visibility notification control unit 73. There is.
- the detection image processing unit 70 sequentially acquires the endoscopic image 75 from the received Bc image signal, Gc image signal, and Rc image signal by the same image processing as the normal mode image processing unit 60 such as color conversion processing. ..
- the attention area detection unit 71 analyzes the endoscopic image 75 and performs the attention area detection process for detecting the attention area in the observation target.
- the region of interest detection unit 71 detects a lesion (for example, a tumor, inflammation, etc.) in the observation target as the region of interest.
- the region of interest detection unit 71 first divides the endoscopic image 75 into a plurality of small regions, for example, a square region for several pixels. Next, the image feature amount is calculated from the divided endoscopic image 75. Subsequently, based on the calculated feature amount, whether or not each small region is a lesion is recognized and processed.
- a machine learning algorithm such as a convolutional neural network or deep learning is preferable.
- the feature amount calculated from the endoscopic image 75 by the region of interest detection unit 71 is preferably a value obtained from the shape and color of a predetermined portion in the observation target or those shapes and colors.
- a value obtained from the shape and color of a predetermined portion in the observation target or those shapes and colors is preferable.
- the value is at least one of the length, the degree of tortuosity of the blood vessel, and the color information, or a combination of two or more of them.
- the region of interest detection unit 71 associates information such as the position information, size, and type of lesion of the extracted lesion portion with the endoscopic image 75 as detection information 76.
- the attention area detection mode image processing unit 64 outputs the endoscope image 75 associated with the detection information 76 to the display control unit 58.
- the display control unit 58 performs display control for displaying the image or data from the image processing unit 56 on the monitor 18.
- the display control unit 58 controls to display the normal image on the monitor 18.
- the display control unit 58 controls to display the special image on the monitor 18.
- the display control unit 58 highlights the attention area detected by the attention area detection unit 71 on the endoscopic image 75.
- the display control unit 58 first bases on the endoscope image 75 output from the region of interest detection mode image processing unit 64 and the detection information 76 associated with the endoscope image 75. Set the emphasis area to emphasize the area of interest.
- the display control unit 58 has an area larger than that of the lesion 77 and is an emphasized region 78 including the lesion 77, based on the detection information 76 such as the position, size, and type of the lesion 77.
- a square area is set as the emphasis area 78.
- the emphasized region 78 has, for example, a square outer circumference that is set at a predetermined distance from the outer circumference of the lesion portion 77.
- the emphasized region 78 is not limited to this, and may be set to a square in contact with the outer circumference of the lesion portion 77.
- the display control unit 58 highlights the highlighted area 78 set as described above. That is, the display control unit 58 superimposes and displays the graphic as the highlight display on the position of the highlight region 78 in the endoscope image 75.
- the display control unit 58 displays the square frame-shaped (frame-shaped) figure 79 surrounding the lesion portion 77 in accordance with the position of the emphasized region 78.
- the display control unit 58 resets the emphasized area 78 according to the amount of fluctuation of the lesion portion 77 in the endoscopic image 75, and shapes the figure according to the position of the reset emphasized area 78. 79 is displayed.
- the graphic 79 as a highlighting is a display mode different from other parts of the endoscopic image 75, and the display control unit 58 is, for example, a color generally included in a large amount in the endoscopic image. Displayes the figure 79 in a color having a different hue. Further, the color of the figure 79 may be set according to the input operation of the user.
- the display control unit 58 outputs the setting information 81 of the graphic 79 as highlighting to the image processing unit 56.
- the setting information 81 includes position information, color information, and the like of the figure 79 with respect to the endoscopic image 75.
- the setting information 81 is tagged with the information of the endoscopic image 75 from which the lesion 77 was detected.
- the visibility determination unit 72 determines the visibility of the highlighting from the image information acquired from the endoscopic image 75 in which the lesion portion 77 is detected and the highlighting setting information 81 set by the display control unit 58.
- the quantified index value is calculated as the judgment result.
- the visibility determination unit 72 calculates the color information from the image information acquired from the endoscopic image 75 in which the lesion portion 77 is detected by the attention region detection unit 71 and the highlighting setting information 81. From these color information, the color difference between the endoscopic image 75 and the figure 79 is calculated as an index value.
- the color information indicates information related to color such as hue, lightness, and saturation.
- the visibility determination unit 72 calculates the color information from the image information acquired from the endoscopic image 75, the inside of the highlighting including the lesion portion 77, specifically, endoscopy.
- the average value within the range 82 (see also FIG. 7) surrounded by the figure 79 in the mirror image 75 is calculated as the color information.
- the setting information 81 includes the position information of the figure 79
- the color information can be calculated by cutting out the range 82 surrounded by the figure 79 from the endoscope image 75 based on the position information. ..
- the average value of the figure 79 is calculated as the color information.
- the visibility determination unit 72 calculates the color difference between the endoscopic image 75 and the figure 79 from these color information.
- the color difference in the visibility determination unit 72 for example, it is obtained by the color difference formula by CIEDE2000 defined in JIS Z 8730 7.3. By using the calculation method standardized in this way, it is possible to obtain a color difference that matches the human visual characteristics.
- the color difference formula by CIEDE2000 is used, the color information for obtaining the color difference is from the L component indicating the brightness, the a component indicating the degree of red or green, and the b component indicating the degree of yellow or blue. Information on the configured CIELab color space is used.
- the calculation method for obtaining the color difference is not limited to the above method, and any calculation method that considers human vision may be used.
- the calculation method may be performed using the Euclidean distance (also referred to as CIE76) in the CIELab color space. good.
- the visibility notification control unit 73 notifies the user of the determination result determined by the visibility determination unit 72. As shown in FIG. 9, the visibility notification control unit 73 outputs the color difference as the determination result calculated as described above to the display control unit 58 as the notification information 83, and displays it on the display screen 84 of the monitor 18. In the present embodiment, when the color difference is equal to or less than the preset first threshold value, the visibility notification control unit 73 outputs the color difference information as the notification information 83 to the display control unit 58.
- a preset first threshold value is set to 2.0.
- the difference can be discriminated when two colors are compared side by side, and when the color difference is 2 to 3, the two colors are separated and compared. It is defined as a level where you can sometimes tell the difference. Since the visibility notification control unit 73 sets the first threshold value to 2.0 as described above based on the JIS regulations, it is possible to notify that the visibility has deteriorated.
- the flowchart shown in FIG. 10 and the process in which the image processing unit 56 and the display control unit 58 determine the visibility of highlighting and display the determination result on the display screen 84 of the monitor 18 in the attention area detection mode The description will be given with reference to the explanatory diagram shown in FIG.
- the doctor who is the user operates the mode switching unit 13c to switch to the attention area detection mode.
- the observation target is illuminated with the illumination light for the region of interest detection mode.
- the observation target illuminated by the illumination light for the region of interest detection mode is imaged by the image sensor 38 to acquire the endoscopic image 75.
- the display control unit 58 sequentially acquires the endoscope images 75 (S11) and displays them on the display screen 84 of the monitor 18 in real time.
- the attention area detection unit 71 performs the attention area detection process for detecting the attention area in the observation target on the acquired endoscope image 75.
- the detection information 76 by the region of interest 71 is output in association with the endoscopic image 75.
- the display control unit 58 sets the emphasized region 78 using the detection information 76 associated with the endoscopic image 75, particularly the position and size information of the lesion 77 (S13).
- the display control unit 58 After setting the highlighting area 78, as shown in FIG. 11B, the display control unit 58 superimposes and displays the graphic 79 as the highlighting on the position of the highlighting area 78 in the endoscopic image 75 (S14). , The setting information 81 of the figure 79 is output to the image processing unit 56.
- the lesion 77 is not detected in the observation target (N in S12)
- the highlighting is naturally not performed.
- the visibility determination unit 72 reads out the original endoscopic image 75.
- Color information is calculated from the image information acquired from the endoscope image 75 and the setting information 81 of the figure 79, and the visibility of highlighting is determined (S15).
- the color difference calculated from the color information is compared with the first threshold value.
- the graphic 79 when the graphic 79 as a highlighting display is displayed, the graphic 79 depends on the color of the subject in the endoscopic image 75, the presence or absence of an object existing in the subject, and the like. May be assimilated with the surroundings or become less noticeable with respect to the surrounding parts, resulting in reduced visibility. In such a case, in general, the value of the color difference between the endoscopic image 75 and the figure 79 also decreases.
- the color difference information is output to the display control unit 58 as the notification information 83.
- the display control unit 58 displays the notification information 83 on the display screen 84 to notify that the visibility is deteriorated (S17).
- the visibility determination unit 72 does not notify.
- the visibility of the highlighting in the endoscopic image 75 is determined, and when the visibility of the highlighting is reduced, the notification is given, so that the doctor who is the user reduces the visibility of the highlighting. It is possible to avoid a state in which the user does not notice the area of interest such as a lesion.
- the example of displaying the color difference information on the display screen as the judgment result of the visibility is given, but the present invention is not limited to this, and the identification information is displayed according to the index value as the judgment result. It may be.
- FIG. 13 shows an example in which the identification information 85A and 85B according to the color difference as the determination result are displayed on the display screen 84.
- color information is calculated from the image information acquired from the endoscope image 75 and the highlighting setting information 81, and the color difference between the endoscope image 75 and the figure 79 is calculated from these color information.
- the process of comparing the color difference with the first threshold value is the same as that of the first embodiment.
- the visibility determination unit 72 outputs the identification information 85A to the display control unit 58.
- the display control unit 58 displays the identification information 85A on the display screen 84 to notify that the visibility is deteriorated.
- the character information "low visibility" is displayed as the identification information 85A.
- the identification information is displayed not only when the color difference is equal to or less than the first threshold value but also when the color difference exceeds the first threshold value, that is, when the visibility is high. May be good.
- the visibility determination unit 72 outputs the identification information 85B to the display control unit 58.
- the display control unit 58 displays the identification information 85B on the display screen 84 to notify that the visibility is high.
- the character information "high visibility" is displayed as the identification information 85B.
- the setting for displaying the identification information according to the color difference is not limited to the two-step setting of the case where the identification information is displayed below the first threshold value or the case where the first threshold value is exceeded, and the setting may be set to three or more steps. good.
- the visibility determination unit 72 is preset with three-step numerical values such that the color difference is 2.0 or less, larger than 2.0 and 4.0 or less and greater than 4.0, and the determination is made based on this setting. conduct. When the color difference is 2.0 or less, the visibility is low, when the color difference is larger than 2.0 and 4.0 or less, the visibility is medium, and when the color difference is larger than 4.0, the visibility is high. Is displayed on the display screen 84 as identification information. In this case, as the identification information, it is preferable to display character information such as "low visibility", "medium visibility", and "high visibility" according to the stage of color difference.
- the example of displaying the identification information according to the color difference as the judgment result is given as the judgment result of the visibility, but the judgment result is not limited to this, as in the example shown in FIG.
- the identification figure corresponding to the color difference may be displayed.
- color information is calculated from the image information acquired from the endoscope image 75 and the highlighting setting information 81, and the color difference between the endoscope image 75 and the figure 79 is calculated from these color information.
- the process of comparing the color difference with the first threshold value is the same as that of the first and second embodiments.
- the information of the icon 86A is output to the display control unit 58.
- the display control unit 58 displays the icon 86A as an identification figure on the display screen 84 to notify that the visibility is deteriorated.
- a mark imitating a sign indicating that there is a danger is displayed as the icon 86A.
- the doctor who is the user can recognize the decrease in visibility of the highlighting.
- the identification figure is displayed not only when the color difference is equal to or less than the first threshold value but also when the color difference exceeds the first threshold value, that is, when the visibility is high. May be good.
- the visibility determination unit 72 outputs the information of the icon 86B to the display control unit 58.
- the display control unit 58 displays the icon 86B as an identification figure on the display screen 84 to notify that the visibility is high.
- a double circle mark is displayed as the icon 86B.
- the setting for displaying the identification information according to the color difference it may be set in three or more stages as in the second embodiment.
- the visibility determination unit 72 sets numerical values in three stages in advance as in the second embodiment, and makes a determination based on these settings.
- the color difference is 2.0 or less, the visibility is low, when the color difference is larger than 2.0 and 4.0 or less, the visibility is medium, and when the color difference is larger than 4.0, the visibility is high.
- This is displayed on the display screen 84 as an identification figure.
- the identification figure it is preferable to display icons having different shapes depending on the stage of color difference.
- an image in which highlighting is superimposed on the endoscopic image is displayed on one display screen, and notification information or the like is displayed in a non-display area of the endoscopic image.
- An image in which notification information or the like is superimposed may be displayed on a display screen different from the display screen for displaying the endoscope image.
- a normal endoscope image 87 is displayed as a main image having a large display area, and a sub image having a display area smaller than that of the main image is included.
- the image 88 in which the highlighting is superimposed on the endoscopic image is displayed.
- the icon 86A as the identification information is further superimposed and displayed on the image 88 on which the highlighting is superimposed.
- the normal endoscopic image 87 is the endoscopic image 75 itself acquired by the image processing unit 56 during the region of interest detection mode in each of the above embodiments, and is used as highlighting. It shows a state in which figures and the like are not superimposed.
- the image 88 on which the highlighting is superimposed as a sub-image is an endoscopic image 75 on which a graphic 79 or the like as a highlighting is superimposed and displayed as in each of the above embodiments.
- color information is calculated from the image information acquired from the endoscope image 75 and the highlighting setting information 81, and the endoscope image 75 and the figure 79 are calculated from these color information.
- the color difference between and is calculated and compared with the first threshold value.
- the information of the icon 86A is output to the display control unit 58.
- the display control unit 58 further superimposes and displays the icon 86A on the image 88 on which the highlighting is superimposed to notify that the visibility is deteriorated.
- the information to be displayed as the visibility determination result is not limited to the icon 86A, and the color difference information as an index value, the identification information according to the color difference, and the like may be displayed as in each of the above embodiments. , Different identification information or identification figures may be displayed depending on the color difference. Further, in the example shown in FIG. 15, two display screens are displayed side by side on one monitor 18, but the main image and the sub image may be displayed on separate monitors.
- the visibility determination unit 72 calculates the color information from the inside of the graphic 79.
- the present invention is not limited to this, and even if the average value of the portion outside the figure 79, specifically, in the endoscopic image 75, excluding the range 82 surrounded by the figure 79 and the figure 79, is calculated as color information. good.
- the setting information 81 includes the position information of the figure 79
- the color information can be calculated by cutting out the outside of the figure 79 from the endoscope image 75 based on the position information.
- the doctor who is the user can recognize the decrease in the visibility of the highlighting.
- the visibility determination unit 72 uses a preset first threshold value for determining the color difference between the endoscopic image 75 and the highlighting.
- the first threshold value used for this determination is not always the same value, and the first threshold value may be weighted according to the thickness of the highlighting line.
- the display control unit 58 makes it possible to change the line thickness of the graphic 79 as highlighting according to the size of the region of interest, or changes the line thickness of the graphic 79 according to the input operation of the user. Make it possible.
- the visibility determination unit 72 weights the first threshold value according to the line thickness of the figure 79 when the highlighting setting information is acquired. do.
- the first threshold value is set small in inverse proportion to the line thickness of the figure 79.
- the first threshold value is 2.0
- the first threshold value is set to 2.0
- the first threshold value is made larger than 2.0.
- the visibility of highlighting the thicker the line thickness of the figure 79, the higher the visibility. Therefore, even if the first threshold value is set small according to the line thickness of the figure 79, each of the above embodiments Similarly, the user doctor can recognize the reduced visibility of the highlighting.
- the visibility determination unit 72 uses the color difference calculated from the image information of the endoscope image 75 and the highlighting setting information as an index value for determining the visibility, but the present invention is not limited to this.
- the visibility may be determined by the presence or absence of a non-detection target object existing inside the highlighting.
- an object 89 or a phenomenon that is not to be detected includes, for example, water, halation, bubbles, pigment, or the like other than the lesion portion 77. It may enter the inside of the figure 79 as a highlight.
- the brightness value is used as the color information acquired from the endoscopic image.
- the luminance value may be used as the color information acquired from the endoscopic image.
- the visibility determination unit 72 sets the brightness of each pixel in the range 82 surrounded by the figure 79 as a second threshold value in order to detect that the object 89 or the phenomenon not to be detected exists inside the figure 79. Compare with.
- This second threshold value is set to a value having high brightness assuming water, halation, bubbles, and the like. Therefore, when the brightness is equal to or higher than the second threshold value, there is a high possibility that an object 89 or a phenomenon not to be detected is reflected.
- the visibility determination unit 72 further determines the area ratio of the portion whose brightness is equal to or greater than the second threshold value (the ratio of the area of the portion whose brightness is equal to or greater than the second threshold value to the area of the range 82) with respect to the range 82 surrounded by the figure 79. ) And the third threshold.
- the third threshold value assumes that there are many areas with high brightness with respect to the range 82, and is set to, for example, an area ratio of 50%.
- the visibility determination unit 72 is in a state where the object 89 or the phenomenon to be detected is inside the figure 79, that is, the visibility. Is determined to be low, and the information of the icon 86A is output to the display control unit 58.
- the display control unit 58 superimposes and displays the icon 86A on the endoscope image 75 to notify that the visibility is deteriorated.
- the information to be displayed as the visibility determination result is not limited to the icon 86A, and information on the area ratio of the portion whose brightness is equal to or higher than the second threshold value, identification information according to the area ratio, and the like may be displayed. , Different identification information or identification figures may be displayed depending on the area ratio.
- the figure as highlighting is a square frame shape, but the frame shape is not limited to this and can surround a region of interest such as a polygon other than a rectangle (quadrangle), a circle, or an ellipse. It should be.
- the figure as the highlighting is not limited to one frame shape surrounding the area of interest, and may be composed of a plurality of shapes.
- the display control unit 58 arranges four L-shaped figures 91A to 91D surrounding the lesion 77 on each corner of the highlight area 78 as highlighting.
- the two-dot chain line is shown for convenience of explaining the arrangement of the L-shaped figures 91A to 91D, and is not actually displayed.
- the visibility determination unit 72 calculates the average value of the four L-shaped figures 91A to 91D as color information, and obtains the color information of the endoscopic image 75 and the color information of the average value of the L-shaped figures 91A to 91D. It is preferable to calculate the color difference from the above and compare it with the first threshold value.
- the visibility determination unit 72 calculates the color information for each of the four L-shaped figures 91A to 91D, and the color information of the endoscope image 75 and the L-shaped figures 91A to 91D, respectively.
- a total of four color differences are calculated from the color information of the above and compared with the first threshold value. In this case, for example, if any one of the four color differences is equal to or less than the first threshold value, it is determined that the visibility is low. Then, when the color difference is equal to or less than the first threshold value, information such as the color difference notification information 83, the identification information, and the identification figure is output to the display control unit 58. After that, the notification is performed in the same manner as in each of the above embodiments.
- the image processing unit 56 determines that the visibility is low, performs notification in the same manner as in each of the above embodiments, and detects the region of interest.
- the speculum image may be saved automatically. As a result, it is possible to later confirm the endoscopic image whose visibility has deteriorated even though the region of interest has been detected, so that it is possible to reliably avoid a state in which the user does not notice the region of interest such as a lesion.
- the storage destination for storing the endoscopic image in which the region of interest is detected is stored in, for example, a storage device provided in the processor device 16 or a server such as a cloud. You may.
- the image processing unit 56 not only notifies the user, but also emits a sound, emits an indicator, or blinks a part of the screen. May be done.
- the visibility determination unit 72 determines the visibility based on the color difference information, the presence or absence of an object other than the detection target, and the like, but the present invention is not limited to this, and the highlighting line for the region of interest is used. It may be judged from the thickness.
- FIG. 18 is an example of a display screen when the visibility determination unit 72 determines the visibility of the highlighting from the thickness of the highlighting line with respect to the region of interest.
- the visibility determination unit 72 calculates the ratio of the line thickness T1 of the figure 79 to the maximum dimension LM of the lesion portion 77 detected from the endoscopic image 75 by the attention region detection unit 71 as an index value. do.
- the maximum dimension LM of the lesion portion 77 is, for example, the dimension of the largest portion of the lesion portion 77 in either the X-axis direction or the Y-axis direction of the endoscopic image 75, and is shown in FIG.
- the maximum dimension LM is defined as the dimension of the largest portion of the lesion portion 77 in the X-axis direction.
- the visibility determination unit 72 compares the ratio of the line thickness T1 of the figure 79 to the maximum dimension LM of the lesion portion 77 described above with the threshold value, and when the ratio is equal to or less than the threshold value, the visibility is improved. Judge as low. Then, as in each of the above embodiments, information such as notification information, identification information, and identification figure is output to the display control unit 58. After that, the notification is performed in the same manner as in each of the above embodiments. In the example shown in FIG. 18, the character information "low visibility" is displayed as the identification information 92. As the ratio of the line thickness of the figure to the area of interest becomes smaller, the visibility becomes lower, so that the doctor who is the user can recognize the decrease in the visibility of the highlighting as in each of the above embodiments.
- the visibility determination unit 72 determines the visibility from the thickness of the highlighting line with respect to the area of interest, but the present invention is not limited to this, and the frame-shaped figure surrounding the area of interest is used as the highlighting. May be displayed and the visibility may be determined from the similarity of the frame shape with respect to the region of interest.
- FIG. 19 is an example of a display screen when determining the visibility of highlighting from the similarity of the frame shape with respect to the region of interest.
- the visibility determination unit 72 analyzes the similarity between the shape of the lesion portion 77 detected from the endoscopic image 75 by the region of interest detection unit 71 and the frame-shaped figure 93 surrounding the lesion portion 77. ..
- the figure 93 is a circular frame-shaped figure that surrounds the lesion portion 77 and is in contact with a plurality of locations on the outer periphery of the lesion portion 77.
- the visibility determination unit 72 analyzes the similarity between the contour shape of the lesion portion 77 and the inner peripheral shape of the figure 93 by a method such as well-known template matching.
- This similarity is compared with the threshold value, and if the similarity is equal to or higher than the threshold value, it is determined that the visibility is low. Then, as in each of the above embodiments, information such as notification information, identification information, and identification figure is output to the display control unit 58. After that, the notification is performed in the same manner as in each of the above embodiments. In the example shown in FIG. 19, the character information "low visibility" is displayed as the identification information 92. The higher the similarity of the frame shape to the region of interest, the lower the visibility. Therefore, as in each of the above embodiments, the doctor who is the user can recognize the decrease in visibility of the highlighting.
- the display control unit 58 superimposes and displays the frame-shaped figure on the position of the highlighted area, but the present invention is not limited to this, and the color of the highlighted area may be changed as the highlighted area.
- the display control unit 58 when the lesion portion 77 as the region of interest is detected and the highlight region is set, the display control unit 58 often sets the highlight region as a highlight in a color different from the original color, for example, in the endoscopic image 75.
- the included colors may be extracted and the emphasized area may be changed to a color different from the other parts of the endoscopic image 75.
- the different colors referred to here are, for example, colors having different hues.
- the highlighting of the highlighted area is not limited to the above, as long as it is an image processing that can be visually distinguished from the surroundings, such as saturation change processing, contrast processing, negative / positive inversion processing, and filtering processing. good.
- highlighting by image processing of the highlighted area and highlighting by a graphic surrounding the lesion portion in each of the above embodiments may be combined.
- the four-color LEDs 20a to 20d are used to illuminate the observation target, but a laser light source and a phosphor may be used to illuminate the observation target. Further, in each of the above embodiments, the four-color LEDs 20a to 20d are used to illuminate the observation target, but a white light source such as a xenon lamp and a rotation filter may be used to illuminate the observation target. Further, instead of the color image sensor 38, a monochrome image sensor may be used to image the observation target.
- the medical image processing apparatus of the present invention is applied to an endoscopic system that acquires an endoscopic image as a medical image, but various endoscopy such as a capsule endoscope is applied. Needless to say, it is applicable to mirror systems, and as other medical images, X-ray images, CT images, MR images, ultrasonic images, pathological images, PET (Positron Emission Tomography) images, etc. are acquired.
- the medical image processing apparatus of the present invention can also be applied to various medical imaging devices.
- the hardware structure of the processing unit that executes various processes is various processors as shown below.
- Various processors include CPU (Central Processing Unit), GPU (Graphical Processing Unit), FPGA (Field Programmable Gate Array), which are general-purpose processors that execute software (programs) and function as various processing units.
- CPU Central Processing Unit
- GPU Graphics Processing Unit
- FPGA Field Programmable Gate Array
- PLD Programmable Logic Device
- PLD Programmable Logic Device
- a dedicated electric circuit which is a processor having a circuit configuration specially designed to execute various processes.
- One processing unit may be composed of one of these various processors, or a combination of two or more processors of the same type or different types (for example, a plurality of FPGAs, a combination of a CPU and an FPGA, or a CPU. And GPU, etc.). Further, a plurality of processing units may be configured by one processor. As an example of configuring a plurality of processing units with one processor, first, as represented by a computer such as a client or a server, one processor is configured by a combination of one or more CPUs and software. There is a form in which this processor functions as a plurality of processing units.
- SoC System On Chip
- a processor that realizes the functions of the entire system including a plurality of processing units with one IC (Integrated Circuit) chip is used.
- the various processing units are configured by using one or more of the above-mentioned various processors as a hardware-like structure.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Engineering & Computer Science (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Animal Behavior & Ethology (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Physics & Mathematics (AREA)
- Biophysics (AREA)
- Optics & Photonics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Quality & Reliability (AREA)
- Gynecology & Obstetrics (AREA)
- Endoscopes (AREA)
Abstract
Description
プロセッサは、医療画像を取得し、医療画像内から注目領域を検出し、検出した注目領域を強調するための強調表示を設定し、強調表示を医療画像に重畳してモニタに表示させ、注目領域が検出された医療画像から取得される画像情報と、設定した強調表示とから強調表示の視認性を判定し、視認性の判定結果をユーザに報知する。光源装置は、観察対象を照明するための照明光を発する。内視鏡は、照明光で照明された観察対象を撮像する撮像センサを有する。モニタは、撮像センサが出力する画像信号を信号処理した医療画像を表示する。
図1に示すように、内視鏡システム10は、内視鏡12と、光源装置14と、プロセッサ装置16と、モニタ18(表示部)と、コンソール19とを有する。内視鏡12は、光源装置14と光学的に接続し、かつ、プロセッサ装置16と電気的に接続する。内視鏡12は、被検体内に挿入する挿入部12aと、挿入部12aの基端部分に設けた操作部12bと、挿入部12aの先端側に設けた湾曲部12c及び先端部12dを有している。操作部12bのアングルノブ13aを操作することにより、湾曲部12cが湾曲動作する。この湾曲動作によって、先端部12dが所望の方向に向けられる。
上記第1実施形態では、視認性の判定結果として、色差の情報を表示画面に表示させる例を上げているが、これに限らず、判定結果としての指標値に応じて識別情報を表示させるようにしてもよい。図13においては、判定結果としての色差に応じた識別情報85A、85Bを表示画面84に表示させた一例を示している。
上記第2実施形態では、視認性の判定結果として、判定結果としての色差に応じた識別情報を表示させる例を上げているが、これに限らず、図14に示す例のように、判定結果としての色差に応じた識別図形を表示させてもよい。
上記各実施形態では、内視鏡画像に強調表示を重畳させた画像を1つの表示画面内に表示させ、内視鏡画像の非表示領域に報知情報等を表示させているが、通常の内視鏡画像を表示する表示画面とは異なる表示画面に、報知情報等を重畳させた画像を表示させてもよい。
上記各実施形態では、視認性判定部72は、視認性の判定として、内視鏡画像75の画像情報及び強調表示の設定情報から算出した色差を指標値として用いているが、これに限らず、強調表示の内側に存在する検出対象外の物体の有無で視認性の判定を行ってもよい。この場合、例えば、図16(A)に示すように、内視鏡画像75において、検出対象外の物体89又は現象として、例えば、水、ハレーション、泡、色素等、病変部77以外のものが強調表示としての図形79の内側に進入する場合がある。これらの検出対象外の物体89又は現象が内視鏡画像75に映り込んだ部分は、色情報を取得した場合、一般的に明度又は輝度が高い。このため、本実施形態では、内視鏡画像から取得する色情報として明度の値を用いる。なお、これに限らず、内視鏡画像から取得する色情報として輝度の値を用いてもよい。
上記各実施形態では、視認性判定部72は、色差の情報、又は検出対象外の物体の有無などで視認性の判定を行っているが、これに限らず、注目領域に対する強調表示の線の太さから判定してもよい。図18は、視認性判定部72が注目領域に対する強調表示の線の太さから強調表示の視認性を判断する場合の表示画面の一例である。
上記第6形態では、視認性判定部72は、注目領域に対する強調表示の線の太さから視認性の判定を行っているが、これに限らず、強調表示として注目領域を囲む枠形状の図形を表示し、注目領域に対する枠形状の類似度から視認性を判定してもよい。図19は、注目領域に対する枠形状の類似度から強調表示の視認性を判断する場合の表示画面の一例である。
12 内視鏡
12a 挿入部
12b 操作部
12c 湾曲部
12d 先端部
13a アングルノブ
13b 静止画像取得部
13c モード切替部
13d ズーム操作部
14 光源装置
16 プロセッサ装置
18 モニタ
19 コンソール
20 光源部
20a V-LED
20b B-LED
20c G-LED
20d R-LED
22 光源制御部
23 波長カットフィルタ
24 ライトガイド
30a 照明光学系
30b 撮像光学系
32 照明レンズ
34 対物レンズ
36 拡大光学系
36a ズームレンズ
36b レンズ駆動部
38 撮像センサ
40 CDS回路
42 AGC回路
44 A/D変換回路
50 画像信号取得部
52 DSP
54 ノイズ低減部
56 画像処理部
58 表示制御部
60 通常モード画像処理部
62 特殊モード画像処理部
64 注目領域検出モード画像処理部
70 検出用画像処理部
71 注目領域検出部
72 視認性判定部
73 視認性報知制御部
75 内視鏡画像
76 検出情報
77 病変部
78 強調領域
79 図形
81 設定情報
82 範囲
83 報知情報
84 表示画面
85A 識別情報
85B 識別情報
86A アイコン
86B アイコン
87 通常の内視鏡画像
88 強調表示を重畳させた画像
89 検出対象外の物体
91A~91D L字状図形
Claims (20)
- プロセッサを備える医療画像処理装置であって、
前記プロセッサは、
医療画像を取得し、
前記医療画像内から注目領域を検出し、
前記検出した前記注目領域を強調するための強調表示を設定し、前記強調表示を前記医療画像に重畳して表示させ、
前記注目領域が検出された前記医療画像から取得される画像情報と、前記設定した前記強調表示とから強調表示の視認性を判定し、
前記視認性の判定結果をユーザに報知する医療画像処理装置。 - 前記プロセッサは、前記医療画像における前記強調表示の内側から前記画像情報を取得する請求項1記載の医療画像処理装置。
- 前記プロセッサは、前記医療画像における前記強調表示の外側から前記画像情報を取得する請求項1の医療画像処理装置。
- 前記プロセッサは、前記画像情報から算出した色情報と、前記強調表示から算出した色情報とから、前記医療画像と前記強調表示との色差を取得し、前記色差から前記視認性を判定する請求項1ないし3のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記画像情報から算出した平均値を前記色情報として算出する請求項1ないし4のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記強調表示として、前記注目領域を囲む枠形状の図形を表示し、前記注目領域に対する前記枠形状の線の太さから前記視認性を判定する請求項1ないし3のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記強調表示として、前記注目領域を囲む枠形状の図形を表示し、前記注目領域に対する前記枠形状の類似度から前記視認性を判定する請求項1ないし3のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記判定結果を表示画面上に表示させる請求項1ないし7のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記視認性の判定結果として、数値化した指標値を算出し、表示画面上に表示させる請求項1ないし8のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記指標値が予め設定された閾値以下の場合、前記報知として前記指標値の表示を行う請求項9記載の医療画像処理装置。
- 前記プロセッサは、前記画像情報及び前記強調表示から算出した色差を前記指標値として用いる請求項9又は10記載の医療画像処理装置。
- 前記プロセッサは、前記視認性の判定結果を数値化した指標値を算出し、前記指標値に応じた識別情報又は識別図形を表示させる請求項1ないし11のいずれか1項記載の医療画像処理装置。
- 前記プロセッサは、前記強調表示の内側に存在する検出対象外の物体の有無で前記視認性の判定を行う請求項1記載の医療画像処理装置。
- 前記プロセッサは、前記医療画像における前記強調表示の内側の範囲に対して、前記強調表示の内側の明度又は輝度が第2の閾値以上である部分の面積比が第3の閾値以上の場合、前記検出対象外の物体が有ると判定する請求項12記載の医療画像処理装置。
- 前記プロセッサは、前記医療画像を表示させる表示画面とは異なる表示画面に前記判定結果を表示させる請求項8記載の医療画像処理装置。
- 前記プロセッサは、前記指標値が予め設定された第1の閾値以下の場合、前記注目領域が検出された前記医療画像を自動で保存する請求項9記載の医療画像処理装置。
- 前記プロセッサは、前記指標値が予め設定された第1の閾値以下の場合、ユーザに警告をする請求項9記載の医療画像処理装置。
- 観察対象を照明するための照明光を発する光源装置と、
前記照明光で照明された観察対象を撮像する撮像センサを有する内視鏡と、
プロセッサと、
前記撮像センサが出力する画像信号を信号処理した医療画像を表示するモニタとを備え、
前記プロセッサは、
前記医療画像を取得し、
前記医療画像内から注目領域を検出し、
前記検出した前記注目領域を強調するための強調表示を設定し、前記強調表示を前記医療画像に重畳して前記モニタに表示させ、
前記注目領域が検出された前記医療画像から取得される画像情報と、前記設定した前記強調表示とから強調表示の視認性を判定し、
前記視認性の判定結果をユーザに報知する内視鏡システム。 - 医療画像を取得するステップと、
前記取得した前記医療画像内から注目領域を検出するステップと、
前記検出した前記注目領域を強調するための強調表示を設定し、前記強調表示を前記医療画像に重畳して表示させるステップと、
前記注目領域が検出された前記医療画像から取得される画像情報と、前記設定した前記強調表示とから強調表示の視認性を判定するステップと、
前記視認性の判定結果をユーザに報知するステップとを含む医療画像処理装置の作動方法。 - 医療画像を取得するとともに、医療画像に画像処理を施す医療画像処理装置にインストールされる医療画像処理装置用プログラムにおいて、
コンピュータに、
前記医療画像を取得する機能と、
前記医療画像内から注目領域を検出する機能と、
前記検出した前記注目領域を強調するための強調表示を設定し、前記強調表示を前記医療画像に重畳して表示させる機能と、
前記注目領域が検出された前記医療画像から取得される画像情報と、前記設定した前記強調表示とから強調表示の視認性を判定する機能と、
前記視認性の判定結果をユーザに報知する機能とを実現させるための医療画像処理装置用プログラム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP21780377.4A EP4129152A4 (en) | 2020-04-03 | 2021-04-02 | MEDICAL IMAGE PROCESSING APPARATUS, ENDOSCOPE SYSTEM, OPERATING METHOD FOR MEDICAL IMAGE PROCESSING APPARATUS, AND PROGRAM FOR MEDICAL IMAGE PROCESSING APPARATUS |
JP2022511157A JPWO2021201272A1 (ja) | 2020-04-03 | 2021-04-02 | |
CN202180026610.7A CN115361898A (zh) | 2020-04-03 | 2021-04-02 | 医疗图像处理装置、内窥镜系统及医疗图像处理装置的工作方法、以及医疗图像处理装置用程序 |
US17/937,290 US20230027950A1 (en) | 2020-04-03 | 2022-09-30 | Medical image processing apparatus, endoscope system, method of operating medical image processing apparatus, and non-transitory computer readable medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020067362 | 2020-04-03 | ||
JP2020-067362 | 2020-04-03 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/937,290 Continuation US20230027950A1 (en) | 2020-04-03 | 2022-09-30 | Medical image processing apparatus, endoscope system, method of operating medical image processing apparatus, and non-transitory computer readable medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021201272A1 true WO2021201272A1 (ja) | 2021-10-07 |
Family
ID=77927224
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/014299 WO2021201272A1 (ja) | 2020-04-03 | 2021-04-02 | 医療画像処理装置、内視鏡システム及び医療画像処理装置の作動方法並びに医療画像処理装置用プログラム |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230027950A1 (ja) |
EP (1) | EP4129152A4 (ja) |
JP (1) | JPWO2021201272A1 (ja) |
CN (1) | CN115361898A (ja) |
WO (1) | WO2021201272A1 (ja) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7148625B2 (ja) * | 2018-09-18 | 2022-10-05 | 富士フイルム株式会社 | 医用画像処理装置、プロセッサ装置、医用画像処理装置の作動方法、及びプログラム |
CN117392449B (zh) * | 2023-10-24 | 2024-09-24 | 青岛美迪康数字工程有限公司 | 基于进镜图像特征的肠镜部位识别方法、装置和设备 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017081976A1 (ja) | 2015-11-10 | 2017-05-18 | オリンパス株式会社 | 内視鏡装置 |
JP2017213097A (ja) * | 2016-05-30 | 2017-12-07 | オリンパス株式会社 | 画像処理装置、画像処理方法およびプログラム |
WO2018198161A1 (ja) | 2017-04-24 | 2018-11-01 | オリンパス株式会社 | 内視鏡画像処理装置及び内視鏡画像処理方法 |
WO2019146077A1 (ja) * | 2018-01-26 | 2019-08-01 | オリンパス株式会社 | 内視鏡画像処理装置、内視鏡画像処理方法及び内視鏡画像処理プログラム |
WO2019244255A1 (ja) * | 2018-06-19 | 2019-12-26 | オリンパス株式会社 | 内視鏡画像処理装置および内視鏡画像処理方法 |
WO2020017212A1 (ja) * | 2018-07-20 | 2020-01-23 | 富士フイルム株式会社 | 内視鏡システム |
WO2020054541A1 (ja) * | 2018-09-11 | 2020-03-19 | 富士フイルム株式会社 | 医療画像処理装置、医療画像処理方法及びプログラム、内視鏡システム |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7282796B2 (ja) * | 2018-09-28 | 2023-05-29 | 富士フイルム株式会社 | 医用画像処理装置、医用画像処理装置の作動方法、及びプログラム、診断支援装置並びに内視鏡システム |
-
2021
- 2021-04-02 CN CN202180026610.7A patent/CN115361898A/zh active Pending
- 2021-04-02 WO PCT/JP2021/014299 patent/WO2021201272A1/ja active Application Filing
- 2021-04-02 JP JP2022511157A patent/JPWO2021201272A1/ja active Pending
- 2021-04-02 EP EP21780377.4A patent/EP4129152A4/en active Pending
-
2022
- 2022-09-30 US US17/937,290 patent/US20230027950A1/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017081976A1 (ja) | 2015-11-10 | 2017-05-18 | オリンパス株式会社 | 内視鏡装置 |
JP2017213097A (ja) * | 2016-05-30 | 2017-12-07 | オリンパス株式会社 | 画像処理装置、画像処理方法およびプログラム |
WO2018198161A1 (ja) | 2017-04-24 | 2018-11-01 | オリンパス株式会社 | 内視鏡画像処理装置及び内視鏡画像処理方法 |
WO2019146077A1 (ja) * | 2018-01-26 | 2019-08-01 | オリンパス株式会社 | 内視鏡画像処理装置、内視鏡画像処理方法及び内視鏡画像処理プログラム |
WO2019244255A1 (ja) * | 2018-06-19 | 2019-12-26 | オリンパス株式会社 | 内視鏡画像処理装置および内視鏡画像処理方法 |
WO2020017212A1 (ja) * | 2018-07-20 | 2020-01-23 | 富士フイルム株式会社 | 内視鏡システム |
WO2020054541A1 (ja) * | 2018-09-11 | 2020-03-19 | 富士フイルム株式会社 | 医療画像処理装置、医療画像処理方法及びプログラム、内視鏡システム |
Also Published As
Publication number | Publication date |
---|---|
EP4129152A4 (en) | 2023-09-20 |
EP4129152A1 (en) | 2023-02-08 |
JPWO2021201272A1 (ja) | 2021-10-07 |
CN115361898A (zh) | 2022-11-18 |
US20230027950A1 (en) | 2023-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6785948B2 (ja) | 医療用画像処理装置及び内視鏡システム並びに医療用画像処理装置の作動方法 | |
US11033175B2 (en) | Endoscope system and operation method therefor | |
US12029384B2 (en) | Medical image processing apparatus, endoscope system, and method for operating medical image processing apparatus | |
JP7335399B2 (ja) | 医用画像処理装置及び内視鏡システム並びに医用画像処理装置の作動方法 | |
US20230027950A1 (en) | Medical image processing apparatus, endoscope system, method of operating medical image processing apparatus, and non-transitory computer readable medium | |
WO2022014235A1 (ja) | 画像解析処理装置、内視鏡システム、画像解析処理装置の作動方法、及び画像解析処理装置用プログラム | |
JP2020065685A (ja) | 内視鏡システム | |
US20230029239A1 (en) | Medical image processing system and method for operating medical image processing system | |
JP7130043B2 (ja) | 医用画像処理装置及び内視鏡システム並びに医用画像処理装置の作動方法 | |
US11627864B2 (en) | Medical image processing apparatus, endoscope system, and method for emphasizing region of interest | |
US20190246874A1 (en) | Processor device, endoscope system, and method of operating processor device | |
US20230101620A1 (en) | Medical image processing apparatus, endoscope system, method of operating medical image processing apparatus, and non-transitory computer readable medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21780377 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022511157 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2021780377 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2021780377 Country of ref document: EP Effective date: 20221103 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |