US20190114738A1 - Image processing apparatus and image processing method - Google Patents
Image processing apparatus and image processing method Download PDFInfo
- Publication number
- US20190114738A1 US20190114738A1 US16/213,246 US201816213246A US2019114738A1 US 20190114738 A1 US20190114738 A1 US 20190114738A1 US 201816213246 A US201816213246 A US 201816213246A US 2019114738 A1 US2019114738 A1 US 2019114738A1
- Authority
- US
- United States
- Prior art keywords
- recorded
- area
- image
- display
- displayed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000012545 processing Methods 0.000 title claims abstract description 70
- 238000003672 processing method Methods 0.000 title claims description 13
- 238000001454 recorded image Methods 0.000 claims abstract description 114
- 238000000034 method Methods 0.000 claims abstract description 103
- 238000001514 detection method Methods 0.000 claims abstract description 101
- 230000002708 enhancing effect Effects 0.000 claims description 5
- 230000003902 lesion Effects 0.000 description 134
- 238000010586 diagram Methods 0.000 description 24
- 239000003550 marker Substances 0.000 description 21
- 238000005259 measurement Methods 0.000 description 20
- 208000037062 Polyps Diseases 0.000 description 6
- 238000004891 communication Methods 0.000 description 4
- 239000002131 composite material Substances 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000003321 amplification Effects 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
- A61B1/000094—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope extracting biological structures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4023—Scaling of whole images or parts thereof, e.g. expanding or contracting based on decimating pixels or lines of pixels; based on inserting pixels or lines of pixels
Definitions
- the present invention relates to an image processing apparatus and an image processing method.
- a configuration is conventionally known in which, for example, images obtained by performing image pickup of an object are simultaneously displayed in one relatively large display area on a display screen of a display apparatus and in another relatively small area on the display screen.
- Japanese Patent Application Laid-Open Publication No. H10-262923 discloses a configuration for causing images obtained by performing image pickup of an object by an electronic endoscope to be simultaneously displayed on a monitor as a parent screen and a child screen having mutually different sizes.
- An image processing apparatus of one aspect of the present invention includes a processor.
- the processor is configured to: perform a process for detecting an area of interest for each of a plurality of observation images that are sequentially inputted; record the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and perform a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased
- An image processing method of one aspect of the present invention includes: performing a process for detecting an area of interest for each of a plurality of observation images obtained by performing image pickup of an object; recording the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and performing a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased.
- FIG. 1 is a diagram showing a configuration of main parts of an endoscope system including an image processing apparatus according to an embodiment
- FIG. 2 is a block diagram for illustrating an example of a specific configuration of the image processing apparatus according to the embodiment
- FIG. 3 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment
- FIG. 4 is a diagram showing an example of a display image displayed on a display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 5 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 6 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 7 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 8 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment, which is different from the example of FIG. 3 ;
- FIG. 9 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- FIG. 10 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 11 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 12 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 13 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 14 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed;
- FIG. 15 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- FIG. 16 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- an endoscope system 1 is configured including a light source driving apparatus 11 , an endoscope 21 , a video processor 31 , an image processing apparatus 32 and a display apparatus 41 .
- the light source driving apparatus 11 is configured, for example, being provided with a drive circuit.
- the light source driving apparatus 11 is connected to the endoscope 21 and the video processor 31 .
- the light source driving apparatus 11 is configured to generate a light source driving signal for driving a light source portion 23 of the endoscope 21 based on a light source control signal from the video processor 31 and output the generated light source driving signal to the endoscope 21 .
- the endoscope 21 is connected to the light source driving apparatus 11 and the video processor 31 .
- the endoscope 21 is configured including an elongated-shaped insertion portion 22 insertable into a body cavity of an examinee A distal end portion of the insertion portion 22 is provided with the light source portion 23 and an image pickup portion 24 .
- the light source portion 23 is configured being provided with a light emitting device such as a white LED.
- the light source portion 23 is configured to generate illuminating light by emitting light in response to a light source driving signal outputted from the light source driving apparatus 11 and radiate the generated illuminating light to an object such as a body tissue.
- the image pickup portion 24 is configured including an image sensor, for example, a color CCD or a color CMOS.
- the image pickup portion 24 is configured to perform an operation corresponding to an image pickup control signal outputted from the video processor 31 .
- the image pickup portion 24 is configured to receive reflected light from an object illuminated by illuminating light from the light source portion 23 , pick up an image of the received reflected light to generate an image pickup signal, and output the generated image pickup signal to the video processor 31 .
- the video processor 31 is connected to the light source driving apparatus 11 and the endoscope 21 .
- the video processor 31 is configured to generate a light source control signal for controlling a light emission state of the light source portion 23 and output the light source control signal to the light source driving apparatus 11 .
- the video processor 31 is configured to generate and output an image pickup control signal for controlling an image pickup operation of the image pickup portion 24 .
- the video processor 31 is configured to generate an observation image G 1 of an object by performing predetermined processing of an image pickup signal outputted from the endoscope 21 and sequentially output the generated observation image G 1 to the image processing apparatus 32 frame by frame.
- the image processing apparatus 32 is configured being provided with an electronic circuit such as an image processing circuit.
- the image processing apparatus 32 is configured to perform an operation for generating a display image based on an observation image G 1 outputted from the video processor 31 and causing the generated display image to be displayed on the display apparatus 41 .
- the image processing apparatus 32 is configured including an area-of-interest detecting portion 34 , a continuous detection judging portion 35 and a display controlling portion 36 .
- FIG. 2 is a block diagram for illustrating an example of a specific configuration of the image processing apparatus according to the embodiment.
- the area-of-interest detecting portion 34 is configured to calculate predetermined feature values for each observation image G 1 sequentially outputted from the video processor 31 and further detect a lesion candidate area L, which is an area of interest included in the observation image G 1 , based on the calculated predetermined feature values. That is, the area-of-interest detecting portion 34 is configured such that a plurality of observation images G 1 obtained by performing image pickup of an object by the endoscope 21 are sequentially inputted and is configured to detect a process for detecting the lesion candidate area L for each of the plurality of observation images G 1 .
- the area-of-interest detecting portion 34 is configured including a feature value calculating portion 34 a and a lesion candidate detecting portion 34 b.
- the feature value calculating portion 34 a is connected to the video processor 31 and the lesion candidate detecting portion 34 b .
- the feature value calculating portion 34 a is configured to calculate the predetermined feature values for each observation image G 1 sequentially outputted from the video processor 31 and output the calculated predetermined feature values to the lesion candidate detecting portion 34 b.
- the feature value calculating portion 34 a calculates, for each of a plurality of small areas obtained by dividing an observation image G 1 in a predetermined size, an inclination value, which is a value indicating an amount of change in brightness or density between each pixel in the small area and each pixel in a small area next to the small area, as a feature value.
- the feature value calculating portion 34 a may calculate a value different from the inclination value described above as a feature value as long as the feature value calculating portion 34 a calculates a value capable of quantitatively evaluating an observation image G 1 .
- the lesion candidate detecting portion 34 b is connected to the continuous detection judging portion 35 and the display controlling portion 36 .
- the lesion candidate detecting portion 34 b is configured including a ROM 34 c in which one or more pieces of polyp model information are stored in advance.
- the polyp model information stored in the ROM 34 c is configured, for example, being provided with feature values obtained by quantifying common points and/or similarities in a lot of polyp images.
- the lesion candidate detecting portion 34 b is configured to detect a lesion candidate area L based on the predetermined feature values outputted from the feature value calculating portion 34 a and the plurality of pieces of polyp model information read from the ROM 34 c , acquire lesion candidate information IL, which is information showing the detected lesion candidate area L, and output the acquired lesion candidate information IL to each of the continuous detection judging portion 35 and the display controlling portion 36 .
- the lesion candidate detecting portion 34 b detects the one small area as a lesion candidate area L.
- the lesion candidate detecting portion 34 b acquires lesion candidate information IL that includes position information and size information about a lesion candidate area L detected by the method described above, and outputs the acquired lesion candidate information IL to each of the continuous detection judging portion 35 and the display controlling portion 36 .
- the position information about a lesion candidate area L is information showing a position of the lesion candidate area L in an observation image G 1 and is acquired, for example, as a pixel position of the lesion candidate area L existing in the observation image G 1 .
- the size information about the lesion candidate area L is information showing a size of the lesion candidate area L in the observation image G 1 and is acquired, for example, as the number of pixels of the lesion candidate area L existing in the observation image G 1 .
- the area-of-interest detecting portion 34 may not be configured including the feature value calculating portion 34 a or the lesion candidate detecting portion 34 b as long as the area-of-interest detecting portion 34 performs the process for detecting a lesion candidate area L from an observation image G 1 . More specifically, the area-of-interest detecting portion 34 may be configured to detect a lesion candidate area L from an observation image G 1 , for example, by performing a process of applying an image identifier, which has acquired a function of being capable of distinguishing a polyp image by a learning method such as deep learning, to the observation image G 1 .
- the continuous detection judging portion 35 is connected to the display controlling portion 36 .
- the continuous detection judging portion 35 is configured including a RAM 35 a capable of storing at least lesion candidate information IL one frame before, among respective pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b.
- the continuous detection judging portion 35 is configured to, for example, based on first lesion candidate information outputted from the lesion candidate detecting portion 34 b and second lesion candidate information stored in the RAM 35 a one frame before the first lesion candidate, determine whether a first lesion candidate area shown by the first lesion candidate information and a second lesion candidate area shown by the second lesion candidate information are the same lesion candidate area L or not.
- the continuous detection judging portion 35 is configured to, if the first and second lesion candidate areas described above are the same lesion candidate area L, acquire a judgment result that detection of a lesion candidate area L in an observation image G 1 continues, that is, a judgment result that a lesion candidate area L detected by the lesion candidate detecting portion 34 b continues to exist in an observation image G 1 and output the judgment result to the display controlling portion 36 .
- the continuous detection judging portion 35 is configured to, if the first and second lesion candidate areas described above are not the same lesion candidate area L, acquire a judgment result that detection of a lesion candidate area L in an observation image G 1 has been interrupted/ceased, that is, a judgment result that a lesion candidate area L detected by the lesion candidate detecting portion 34 b has moved outside from an inside of an observation image G 1 and output the judgment result to the display controlling portion 36 .
- the display controlling portion 36 is connected to the display apparatus 41 .
- the display controlling portion 36 is configured to, when lesion candidate information IL is inputted from the lesion candidate detecting portion 34 b , measure a continuous detection time period TL, which is a time period elapsed after detection of a lesion candidate area L in an observation images G 1 is started, based on a judgment result outputted from the continuous detection judging portion 35 .
- the display controlling portion 36 is configured to perform a process for generating a display image using observation images G 1 sequentially outputted from the video processor 31 and perform a process for causing the generated display image to be displayed on a display screen 41 A of the display apparatus 41 .
- the display controlling portion 36 is configured to, based on a judgment result outputted from the continuous detection judging portion 35 , perform a process described later if detection of a lesion candidate area L has been interrupted/ceased before the continuous detection time period TL reaches a predetermined time period TH (for example, 0.5 seconds).
- the display controlling portion 36 is configured to, based on a judgment result outputted from the continuous detection judging portion 35 , perform an enhancement process described later if detection of a lesion candidate area L continues at a timing when the continuous detection time period TL reaches the predetermined time period TH.
- the display controlling portion 36 is configured including an enhancement processing portion 36 a and a recording portion 36 b.
- the enhancement processing portion 36 a is configured to, if detection of a lesion candidate area L continues at the timing when the continuous detection time period TL reaches the predetermined time period TH, generate a marker image G 2 for enhancing a position of the lesion candidate area L based on lesion candidate information IL outputted from the lesion candidate detecting portion 34 b and start an enhancement process of adding the marker image G 2 to the observation image G 1 .
- the marker image G 2 added by the enhancement process of the enhancement processing portion 36 a may be in any form as long as the marker image G 2 is capable of presenting a position of a lesion candidate area L as visual information.
- the enhancement processing portion 36 a may perform the enhancement process using only position information included in lesion candidate information IL or may perform the enhancement process using both of position information and size information included in lesion candidate information IL as long as the enhancement processing portion 36 a generates a marker image G 2 for enhancing a position of a lesion candidate area L.
- the recording portion 36 b is configured to record observation images G 1 sequentially outputted from the video processor 31 during a period of measurement of the continuous detection time period TL, as recorded images R 1 . That is, the recording portion 36 b is configured to sequentially (in time-series order) record a plurality of observation images G 1 sequentially outputted from the video processor 31 during a period until detection of a lesion candidate area L by the lesion candidate detecting portion 34 b of the area-of-interest detecting portion 34 is interrupted/ceased after the detection is started, as a plurality of recorded images R 1 .
- the recording portion 36 b is configured to be capable of mutually associating and recording pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b during the period of measurement of the continuous detection time period TL and recorded images R 1 .
- the display apparatus 41 is provided with a monitor and the like and is configured to be capable of displaying a display image outputted from the image processing apparatus 32 .
- FIG. 3 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment.
- the endoscope 21 radiates illuminating light to an object, receives reflected light from the object, picks up an image of the received reflected light to generate an image pickup signal, and outputs the generated image pickup signal to the video processor 31 .
- the video processor 31 generates an observation image G 1 of an object by performing predetermined processing of an image pickup signal outputted from the endoscope 21 and sequentially outputs the generated observation image G 1 to the image processing apparatus 32 frame by frame.
- the display controlling portion 36 performs a process for causing a display image on which an observation image G 1 is disposed in a display area D 1 on the display screen 41 A to be displayed.
- a display image as shown in FIG. 4 is displayed on the display screen 41 A of the display apparatus 41 , for example, during a period before time Ta in FIG. 3 .
- the display area D 1 is set in advance, for example, as an area having a larger size than a display area D 2 to be described later on the display screen 41 A.
- FIG. 4 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the display controlling portion 36 performs a process for causing a display image on which an observation image G 1 including the lesion candidate area L 1 is disposed in the display area D 1 on the display screen 41 A to be displayed, at a timing when measurement of the continuous detection time period TL is started, that is, at a timing when detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is started.
- a display image as shown in FIG. 5 is displayed on the display screen 41 A of the display apparatus 41 at a timing of the time Ta in FIG. 3 .
- FIGS. 5 and 6 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the recording portion 36 b of the display controlling portion 36 starts a process for recording observation images G 1 sequentially outputted from the video processor 31 as recorded images R 1 at the timing when measurement of the continuous detection time period TL is started, as recorded images R 1 and starts a process for mutually associating and recording the recorded images R 1 and pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b .
- the process for mutually associating and recording the recorded images R 1 and the pieces of lesion candidate information IL is started at the timing of the time Ta in FIG. 3 .
- the recording portion 36 b of the display controlling portion 36 stops the process for recording observation images G 1 sequentially outputted from the video processor 31 as recorded images R 1 , at a timing when measurement of the continuous detection time period TL is stopped, that is, at a timing when detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is interrupted/ceased, and stops the process for mutually associating and recording the recorded images R 1 and pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b .
- the process for mutually associating and recording the recorded images R 1 and the pieces of lesion candidate information IL is stopped at the timing of the time Tb in FIG. 3 .
- observation images G 1 of N (N ⁇ 2) frames are sequentially recorded as recorded images R 1 .
- the display controlling portion 36 starts a process for displaying a display image on which an observation image G 1 is disposed in the display area D 1 on the display screen 41 A, and recorded images R 1 recorded by the recording portion 36 b during the period of measurement of the continuous detection time period TL are sequentially disposed in the display area D 2 on the display screen 41 A, at a timing when the predetermined time period TH elapses after detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is started.
- FIG. 7 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the display controlling portion 36 After the timing when the predetermined time period TH elapses after detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is started, the display controlling portion 36 performs a process for causing recorded images R 1 to be sequentially displayed in order opposite to order of recording of the recorded images R 1 by recording portion 36 b .
- recorded images R 1 of N frames sequentially recorded by the recording portion 36 b are displayed in the display area D 2 in order of N-th frame ⁇ (N ⁇ 1)-th frame ⁇ . . . ⁇ second frame ⁇ first frame.
- the continuous detection time period TL is shorter than the predetermined time period TH, a process for, while causing a plurality of observation images G 1 sequentially outputted from the video processor 31 to be sequentially displayed in the display area D 1 on the display screen 41 A, causing a plurality of recorded images R 1 recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D 2 on the display screen 41 A in order opposite to order of recording by the recording portion 36 b at a timing of the time Tc in FIG. 3 is started.
- the continuous detection time period TL is shorter than the predetermined time period TH, a situation may occur in which, although the lesion candidate area L 1 is detected by the lesion candidate detecting portion 34 b , the lesion candidate area L 1 moves outside from an inside of an observation image G 1 without a marker image G 2 being displayed. Therefore, it is conceivable that, if the continuous detection time period TL is shorter than the predetermined time period TH, oversight of the lesion candidate area L 1 by a user's visual examination easily occurs.
- FIG. 8 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment, which is different from the example of FIG. 3 . Note that, hereinafter, specific description of parts to which the processes and the like already described are applicable will be appropriately omitted for simplification.
- the display controlling portion 36 performs a process for causing a display image on which an observation image G 1 is disposed in a display area D 1 on the display screen 41 A to be displayed. According to such an operation of the display controlling portion 36 , the display image as shown in FIG. 4 is displayed on the display screen 41 A of the display apparatus 41 , for example, during a period before time Td in FIG. 8 .
- the display controlling portion 36 performs the process for causing a display image on which an observation image G 1 including the lesion candidate area L 1 is disposed in the display area D 1 on the display screen 41 A to be displayed, at the timing when measurement of the continuous detection time period TL is started, that is, at the timing when detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is started.
- the display image as shown in FIG. 5 is displayed on the display screen 41 A of the display apparatus 41 at a timing of the time Td in FIG. 8 .
- the recording portion 36 b of the display controlling portion 36 starts the process for recording observation images G 1 sequentially outputted from the video processor 31 as recorded images R 1 at the timing when measurement of the continuous detection time period TL is started, as recorded images R 1 and starts the process for mutually associating and recording the recorded images R 1 and pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b .
- the process for mutually associating and recording the recorded images R 1 and the pieces of lesion candidate information IL is started at the timing of the time Td in FIG. 8 .
- the enhancement processing portion 36 a of the display controlling portion 36 starts an enhancement process of adding a marker image G 2 for enhancing a position of the lesion candidate area L 1 detected by the lesion candidate detecting portion 34 b to an observation image G 1 at a timing when the predetermined time period TH elapses after measurement of the continuous detection time period TL is started.
- a display image as shown in FIG. 9 is displayed on the display screen 41 A of the display apparatus 41 at a timing of time Te in FIG.
- FIG. 8 which corresponds to the timing when the predetermined time period TH elapses after detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is started.
- a display image as shown in FIG. 10 is displayed on the display screen 41 A of the display apparatus 41 at a timing immediately before time Tf after the time Te in FIG. 8 is reached.
- FIGS. 9 and 10 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the recording portion 36 b of the display controlling portion 36 stops the process for recording observation images G 1 sequentially outputted from the video processor 31 as recorded images R 1 , at the timing when measurement of the continuous detection time period TL is stopped, that is, at the timing when detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is interrupted/ceased, and stops the process for mutually associating and recording the recorded images R 1 and pieces of lesion candidate information IL outputted from the lesion candidate detecting portion 34 b .
- the process for mutually associating and recording the recorded images R 1 and the pieces of lesion candidate information IL is stopped at a timing of the time Tf in FIG. 8 .
- observation images G 1 of P (P ⁇ 2) frames are sequentially recorded as recorded images R 1 .
- the display controlling portion 36 starts the process for displaying a display image on which an observation image G 1 is disposed in the display area D 1 on the display screen 41 A, and recorded images R 1 recorded by the recording portion 36 b during the period of measurement of the continuous detection time period TL are sequentially disposed in the display area D 2 on the display screen 41 A, at the timing when detection of the lesion candidate area L 1 is interrupted/ceased.
- FIG. 11 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the display controlling portion 36 After the timing when detection of the lesion candidate area L 1 by the lesion candidate detecting portion 34 b is interrupted/ceased, the display controlling portion 36 performs the process for causing recorded images R 1 to be sequentially displayed in order opposite to order of recording of the recorded images R 1 by the recording portion 36 b .
- recorded images R 1 of P frames sequentially recorded by the recording portion 36 b are displayed in the display area D 2 in order of P-th frame ⁇ (P ⁇ 1)-th frame ⁇ . . . ⁇ second frame ⁇ first frame.
- the process for, while causing a plurality of observation images G 1 sequentially outputted from the video processor 31 to be sequentially displayed in the display area D 1 on the display screen 41 A, causing a plurality of recorded images R 1 recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D 2 on the display screen 41 A in order opposite to order of recording by the recording portion 36 b at the timing of the time Tf in FIG. 8 is started.
- a difference time period ⁇ T between the continuous detection time period TL and the predetermined time period TH which corresponds to a period from the time Te to the time Tf in FIG. 8 , is extremely short, a situation may occur in which, without the user visually confirming a marker image G 2 instantaneously displayed in the display area D 1 , the lesion candidate area L 1 enhanced by the marker image G 2 moves outside from an inside of an observation image G 1 . Therefore, it is also conceivable that, if the difference time period ⁇ T is extremely short, oversight of the lesion candidate area L 1 by the user's visual examination easily occurs.
- the process for displaying recorded images R 1 in the display area D 2 in order opposite to order of recording by the recording portion 36 b is performed by the display controlling portion 36 , but the process is not limited to this.
- a process for displaying recorded images R 1 in the display area D 2 in the same order as order of recording by the recording portion 36 b may be performed by the display controlling portion 36 .
- the process for causing recorded images R 1 of respective frames recorded during the period of measurement of the continuous detection time period TL to be sequentially displayed in the display area D 2 is performed by the display controlling portion 36 , but the process is not limited to this.
- a process for causing recorded images R 1 of a part of frames among the recorded images R 1 of the respective frames to be displayed in the display area D 2 may be performed by the display controlling portion 36 .
- a process for causing only a recorded image R 1 corresponding to one frame recorded last among the recorded images R 1 of the respective frames recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D 2 may be performed by the display controlling portion 36 .
- a process for, while decimating the plurality of recorded images R 1 recorded during the period of measurement of the continuous detection time period TL at predetermined intervals, causing the recorded images R 1 to be displayed in the display area D 2 may be performed by the display controlling portion 36 .
- the recording portion 36 b may record respective observation images G 1 obtained by decimating a plurality of observation images G 1 sequentially outputted from the video processor 31 during the period of measurement of the continuous detection time period TL at predetermined intervals, as recorded images R 1 .
- the recording portion 36 b is not limited to such a configuration that sequentially records observation images G 1 of a plurality of frames as recorded images R 1 but may, for example, record only an observation image of one frame as a recorded image R 1 . More specifically, the recording portion 36 b may, for example, record only an observation image G 1 inputted to the display controlling portion 36 at a timing immediately before measurement of the continuous detection time period TL is stopped (the observation images G 1 shown in FIGS. 6 and 10 ) as a recorded image R 1 .
- the display controlling portion 36 when the display controlling portion 36 causes a recorded image R 1 of each frame recorded in the recording portion 36 b to be displayed in the display area D 2 , the display controlling portion 36 may cause the recorded image R 1 to be equal-speed displayed at the same frame rate as a frame rate at the time of recording by the recording portion 36 b , may cause the recorded image R 1 to be double-speed displayed at a frame rate higher than the frame rate at the time of recording by the recording portion 36 b , or may cause the recorded image R 1 to be slow-displayed at a frame rate lower than the frame rate at the time of recording by the recording portion 36 b.
- the display controlling portion 36 when the display controlling portion 36 causes the recorded image R 1 recorded in the recording portion 36 b to be displayed in the display area D 2 , the display controlling portion 36 may cause the recorded image R 1 to be displayed in the same color as color at the time of recording by the recording portion 36 b , may cause the recorded image R 1 to be displayed in subtractive color obtained from the color at the time of recording by the recording portion 36 b , or may cause the recorded image R 1 to be displayed only in predetermined one color.
- recording of recorded images R 1 may be started at a desired timing before the timing when detection of the lesion candidate area L 1 is started as long as recording of recorded images R 1 is performed during the period of measurement of the continuous detection time period TL.
- an enhancement process for generating a marker image G 2 based on lesion candidate information IL recorded in a state of being associated with the recorded image R 1 and adding the generated marker image G 2 to the recorded image R 1 may be performed by the enhancement processing portion 36 a .
- the display controlling portion 36 for example, it is possible to, at the timing of the time Tc in FIG. 3 or at the timing of the time Tf in FIG. 8 , cause a display image on which a position of the lesion candidate area L 1 included in the recorded image R 1 in the display area D 2 is enhanced as shown in FIG. 12 to be displayed on the display screen 41 A.
- FIG. 12 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- a process for causing a display image on which a recorded image R 1 recorded in the recording portion 36 b is disposed in the display area D 1 to be temporarily displayed instead of an observation image G 1 , and causing each of such recorded images R 1 displayed in the display area D 1 instead of an observation image G 1 to be sequentially redisplayed in the display area D 2 may be performed by the display controlling portion 36 .
- the display controlling portion 36 for example, it is possible to, at the timing of the time Tc in FIG. 3 or at the timing of the time Tf in FIG. 8 , cause a display image on which a position of the lesion candidate area L 1 included in the recorded image R 1 in the display area D 1 is enhanced as shown in FIG.
- FIGS. 13 and 14 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- a process for causing a recording image R 1 and/or a marker image G 2 to be displayed in the display area D 1 may be performed by the display controlling portion 36 .
- a process for generating a composite image GR 1 by combining an observation image G 1 sequentially outputted from the video processor 31 and a recorded image R 1 recorded in the recording portion 36 b , generating a marker image G 2 based on lesion candidate information IL associated with the recorded image R 1 , and adding the marker image G 2 to the composite image GR 1 to cause the marker image G 2 and the composite image GR 1 to be displayed in the display area D 1 may be performed by the display controlling portion 36 .
- the display controlling portion 36 for example, it is possible to, at the timing of the time Tc in FIG. 3 or at the timing of the time Tf in FIG.
- FIG. 15 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- a process for generating a marker image G 2 based on lesion candidate information IL recorded in the recording portion 36 b , and adding the generated marker image G 2 to an observation image G 1 sequentially outputted from the video processor 31 to cause the marker image G 2 and the observation image G 1 to be displayed in the display area D 1 may be performed by the display controlling portion 36 .
- the display controlling portion 36 for example, it is possible to cause a display image as shown in FIG. 16 to be displayed on the display screen 41 A of the display apparatus 41 at the timing of the time Tc in FIG. 3 or at the timing of the time Tf in FIG. 8 .
- FIG. 16 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed.
- the image processing apparatus and the like may include a processor and a storage (e.g., a memory).
- the functions of individual units in the processor may be implemented by respective pieces of hardware or may be implemented by an integrated piece of hardware, for example.
- the processor may include hardware, and the hardware may include at least one of a circuit for processing digital signals and a circuit for processing analog signals, for example.
- the processor may include one or a plurality of circuit devices (e.g., an IC) or one or a plurality of circuit elements (e.g., a resistor, a capacitor) on a circuit board, for example.
- the processor may be a CPU (Central Processing Unit), for example, but this should not be construed in a limiting sense, and various types of processors including a GPU (Graphics Processing Unit) and a DSP (Digital Signal Processor) may be used.
- the processor may be a hardware circuit with an ASIC or an FPGA.
- the processor may include an amplification circuit, a filter circuit, or the like for processing analog signals.
- the memory may be a semiconductor memory such as an SRAM and a DRAM; a register; a magnetic storage device such as a hard disk device; and an optical storage device such as an optical disk device.
- the memory stores computer-readable instructions, for example. When the instructions are executed by the processor, the functions of each unit of the image processing device and the like are implemented.
- the instructions may be a set of instructions constituting a program or an instruction for causing an operation on the hardware circuit of the processor.
- the units in the image processing apparatus and the like and the display device according to the present embodiment may be connected with each other via any types of digital data communication such as a communication network or via communication media.
- the communication network may include a LAN (Local Area Network), a WAN (Wide Area Network), and computers and networks which form the internee, for example.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Surgery (AREA)
- Physics & Mathematics (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- Optics & Photonics (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Signal Processing (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Endoscopes (AREA)
- Instruments For Viewing The Inside Of Hollow Bodies (AREA)
- Image Analysis (AREA)
Abstract
An image processing apparatus includes a processor. The processor is configured to: perform a process for detecting an area of interest for each of a plurality of observation images that are sequentially inputted; record the plurality of observation images as one or more recorded images during a period until detection of the area of interest is interrupted/ceased after the detection is started; and perform a process for, while causing the plurality of observation images to be sequentially displayed, causing at least one recorded image to be displayed at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased.
Description
- This application is a continuation application of PCT/JP2016/067924 filed on Jun. 16, 2016, the entire contents of which are incorporated herein by this reference.
- The present invention relates to an image processing apparatus and an image processing method.
- In a medical field, a configuration is conventionally known in which, for example, images obtained by performing image pickup of an object are simultaneously displayed in one relatively large display area on a display screen of a display apparatus and in another relatively small area on the display screen.
- More specifically, for example, Japanese Patent Application Laid-Open Publication No. H10-262923 discloses a configuration for causing images obtained by performing image pickup of an object by an electronic endoscope to be simultaneously displayed on a monitor as a parent screen and a child screen having mutually different sizes.
- An image processing apparatus of one aspect of the present invention includes a processor. The processor is configured to: perform a process for detecting an area of interest for each of a plurality of observation images that are sequentially inputted; record the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and perform a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased
- An image processing method of one aspect of the present invention includes: performing a process for detecting an area of interest for each of a plurality of observation images obtained by performing image pickup of an object; recording the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and performing a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased.
-
FIG. 1 is a diagram showing a configuration of main parts of an endoscope system including an image processing apparatus according to an embodiment; -
FIG. 2 is a block diagram for illustrating an example of a specific configuration of the image processing apparatus according to the embodiment; -
FIG. 3 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment; -
FIG. 4 is a diagram showing an example of a display image displayed on a display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 5 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 6 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 7 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 8 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment, which is different from the example ofFIG. 3 ; -
FIG. 9 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 10 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 11 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 12 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 13 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 14 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; -
FIG. 15 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed; and -
FIG. 16 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - An embodiment of the present invention will be described below with reference to drawings.
- As shown in
FIG. 1 , anendoscope system 1 is configured including a lightsource driving apparatus 11, anendoscope 21, avideo processor 31, animage processing apparatus 32 and adisplay apparatus 41. - The light
source driving apparatus 11 is configured, for example, being provided with a drive circuit. The lightsource driving apparatus 11 is connected to theendoscope 21 and thevideo processor 31. The lightsource driving apparatus 11 is configured to generate a light source driving signal for driving alight source portion 23 of theendoscope 21 based on a light source control signal from thevideo processor 31 and output the generated light source driving signal to theendoscope 21. - The
endoscope 21 is connected to the lightsource driving apparatus 11 and thevideo processor 31. Theendoscope 21 is configured including an elongated-shaped insertion portion 22 insertable into a body cavity of an examinee A distal end portion of theinsertion portion 22 is provided with thelight source portion 23 and animage pickup portion 24. - The
light source portion 23 is configured being provided with a light emitting device such as a white LED. Thelight source portion 23 is configured to generate illuminating light by emitting light in response to a light source driving signal outputted from the lightsource driving apparatus 11 and radiate the generated illuminating light to an object such as a body tissue. - The
image pickup portion 24 is configured including an image sensor, for example, a color CCD or a color CMOS. Theimage pickup portion 24 is configured to perform an operation corresponding to an image pickup control signal outputted from thevideo processor 31. Theimage pickup portion 24 is configured to receive reflected light from an object illuminated by illuminating light from thelight source portion 23, pick up an image of the received reflected light to generate an image pickup signal, and output the generated image pickup signal to thevideo processor 31. - The
video processor 31 is connected to the lightsource driving apparatus 11 and theendoscope 21. Thevideo processor 31 is configured to generate a light source control signal for controlling a light emission state of thelight source portion 23 and output the light source control signal to the lightsource driving apparatus 11. Thevideo processor 31 is configured to generate and output an image pickup control signal for controlling an image pickup operation of theimage pickup portion 24. Thevideo processor 31 is configured to generate an observation image G1 of an object by performing predetermined processing of an image pickup signal outputted from theendoscope 21 and sequentially output the generated observation image G1 to theimage processing apparatus 32 frame by frame. - The
image processing apparatus 32 is configured being provided with an electronic circuit such as an image processing circuit. Theimage processing apparatus 32 is configured to perform an operation for generating a display image based on an observation image G1 outputted from thevideo processor 31 and causing the generated display image to be displayed on thedisplay apparatus 41. As shown inFIG. 2 , theimage processing apparatus 32 is configured including an area-of-interest detecting portion 34, a continuousdetection judging portion 35 and adisplay controlling portion 36.FIG. 2 is a block diagram for illustrating an example of a specific configuration of the image processing apparatus according to the embodiment. - The area-of-
interest detecting portion 34 is configured to calculate predetermined feature values for each observation image G1 sequentially outputted from thevideo processor 31 and further detect a lesion candidate area L, which is an area of interest included in the observation image G1, based on the calculated predetermined feature values. That is, the area-of-interest detecting portion 34 is configured such that a plurality of observation images G1 obtained by performing image pickup of an object by theendoscope 21 are sequentially inputted and is configured to detect a process for detecting the lesion candidate area L for each of the plurality of observation images G1. The area-of-interest detecting portion 34 is configured including a featurevalue calculating portion 34 a and a lesioncandidate detecting portion 34 b. - The feature
value calculating portion 34 a is connected to thevideo processor 31 and the lesioncandidate detecting portion 34 b. The featurevalue calculating portion 34 a is configured to calculate the predetermined feature values for each observation image G1 sequentially outputted from thevideo processor 31 and output the calculated predetermined feature values to the lesioncandidate detecting portion 34 b. - More specifically, the feature
value calculating portion 34 a calculates, for each of a plurality of small areas obtained by dividing an observation image G1 in a predetermined size, an inclination value, which is a value indicating an amount of change in brightness or density between each pixel in the small area and each pixel in a small area next to the small area, as a feature value. Note that the featurevalue calculating portion 34 a may calculate a value different from the inclination value described above as a feature value as long as the featurevalue calculating portion 34 a calculates a value capable of quantitatively evaluating an observation image G1. - The lesion
candidate detecting portion 34 b is connected to the continuousdetection judging portion 35 and thedisplay controlling portion 36. The lesioncandidate detecting portion 34 b is configured including aROM 34 c in which one or more pieces of polyp model information are stored in advance. - More specifically, the polyp model information stored in the
ROM 34 c is configured, for example, being provided with feature values obtained by quantifying common points and/or similarities in a lot of polyp images. - The lesion
candidate detecting portion 34 b is configured to detect a lesion candidate area L based on the predetermined feature values outputted from the featurevalue calculating portion 34 a and the plurality of pieces of polyp model information read from theROM 34 c, acquire lesion candidate information IL, which is information showing the detected lesion candidate area L, and output the acquired lesion candidate information IL to each of the continuousdetection judging portion 35 and thedisplay controlling portion 36. - More specifically, for example, if a feature value of one small area outputted from the feature
value calculating portion 34 a corresponds to at least one feature value included in the plurality of pieces of polyp model information read from theROM 34 c, the lesioncandidate detecting portion 34 b detects the one small area as a lesion candidate area L. The lesioncandidate detecting portion 34 b acquires lesion candidate information IL that includes position information and size information about a lesion candidate area L detected by the method described above, and outputs the acquired lesion candidate information IL to each of the continuousdetection judging portion 35 and thedisplay controlling portion 36. - Note that the position information about a lesion candidate area L is information showing a position of the lesion candidate area L in an observation image G1 and is acquired, for example, as a pixel position of the lesion candidate area L existing in the observation image G1. The size information about the lesion candidate area L is information showing a size of the lesion candidate area L in the observation image G1 and is acquired, for example, as the number of pixels of the lesion candidate area L existing in the observation image G1.
- Note that the area-of-
interest detecting portion 34 may not be configured including the featurevalue calculating portion 34 a or the lesioncandidate detecting portion 34 b as long as the area-of-interest detecting portion 34 performs the process for detecting a lesion candidate area L from an observation image G1. More specifically, the area-of-interest detecting portion 34 may be configured to detect a lesion candidate area L from an observation image G1, for example, by performing a process of applying an image identifier, which has acquired a function of being capable of distinguishing a polyp image by a learning method such as deep learning, to the observation image G1. - The continuous
detection judging portion 35 is connected to thedisplay controlling portion 36. The continuousdetection judging portion 35 is configured including aRAM 35 a capable of storing at least lesion candidate information IL one frame before, among respective pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b. - The continuous
detection judging portion 35 is configured to, for example, based on first lesion candidate information outputted from the lesioncandidate detecting portion 34 b and second lesion candidate information stored in theRAM 35 a one frame before the first lesion candidate, determine whether a first lesion candidate area shown by the first lesion candidate information and a second lesion candidate area shown by the second lesion candidate information are the same lesion candidate area L or not. The continuousdetection judging portion 35 is configured to, if the first and second lesion candidate areas described above are the same lesion candidate area L, acquire a judgment result that detection of a lesion candidate area L in an observation image G1 continues, that is, a judgment result that a lesion candidate area L detected by the lesioncandidate detecting portion 34 b continues to exist in an observation image G1 and output the judgment result to thedisplay controlling portion 36. The continuousdetection judging portion 35 is configured to, if the first and second lesion candidate areas described above are not the same lesion candidate area L, acquire a judgment result that detection of a lesion candidate area L in an observation image G1 has been interrupted/ceased, that is, a judgment result that a lesion candidate area L detected by the lesioncandidate detecting portion 34 b has moved outside from an inside of an observation image G1 and output the judgment result to thedisplay controlling portion 36. - The
display controlling portion 36 is connected to thedisplay apparatus 41. Thedisplay controlling portion 36 is configured to, when lesion candidate information IL is inputted from the lesioncandidate detecting portion 34 b, measure a continuous detection time period TL, which is a time period elapsed after detection of a lesion candidate area L in an observation images G1 is started, based on a judgment result outputted from the continuousdetection judging portion 35. Thedisplay controlling portion 36 is configured to perform a process for generating a display image using observation images G1 sequentially outputted from thevideo processor 31 and perform a process for causing the generated display image to be displayed on adisplay screen 41A of thedisplay apparatus 41. - The
display controlling portion 36 is configured to, based on a judgment result outputted from the continuousdetection judging portion 35, perform a process described later if detection of a lesion candidate area L has been interrupted/ceased before the continuous detection time period TL reaches a predetermined time period TH (for example, 0.5 seconds). Thedisplay controlling portion 36 is configured to, based on a judgment result outputted from the continuousdetection judging portion 35, perform an enhancement process described later if detection of a lesion candidate area L continues at a timing when the continuous detection time period TL reaches the predetermined time period TH. Thedisplay controlling portion 36 is configured including anenhancement processing portion 36 a and arecording portion 36 b. - The
enhancement processing portion 36 a is configured to, if detection of a lesion candidate area L continues at the timing when the continuous detection time period TL reaches the predetermined time period TH, generate a marker image G2 for enhancing a position of the lesion candidate area L based on lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b and start an enhancement process of adding the marker image G2 to the observation image G1. - Note that the marker image G2 added by the enhancement process of the
enhancement processing portion 36 a may be in any form as long as the marker image G2 is capable of presenting a position of a lesion candidate area L as visual information. In other words, theenhancement processing portion 36 a may perform the enhancement process using only position information included in lesion candidate information IL or may perform the enhancement process using both of position information and size information included in lesion candidate information IL as long as theenhancement processing portion 36 a generates a marker image G2 for enhancing a position of a lesion candidate area L. - The
recording portion 36 b is configured to record observation images G1 sequentially outputted from thevideo processor 31 during a period of measurement of the continuous detection time period TL, as recorded images R1. That is, therecording portion 36 b is configured to sequentially (in time-series order) record a plurality of observation images G1 sequentially outputted from thevideo processor 31 during a period until detection of a lesion candidate area L by the lesioncandidate detecting portion 34 b of the area-of-interest detecting portion 34 is interrupted/ceased after the detection is started, as a plurality of recorded images R1. Therecording portion 36 b is configured to be capable of mutually associating and recording pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b during the period of measurement of the continuous detection time period TL and recorded images R1. - The
display apparatus 41 is provided with a monitor and the like and is configured to be capable of displaying a display image outputted from theimage processing apparatus 32. - Next, operation of the present embodiment will be described. Note that, in the description below, a case where the lesion
candidate detecting portion 34 b detects one lesion candidate area L1 will be described as an example for simplification. In the description below, a case where the continuous detection time period TL of the lesion candidate area L1 is shorter than the predetermined time period TH, that is, a case where a process corresponding to a time chart ofFIG. 3 is performed by thedisplay controlling portion 36 will be described as an example.FIG. 3 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment. - For example, when the light
source driving apparatus 11 and thevideo processor 31 are powered on, theendoscope 21 radiates illuminating light to an object, receives reflected light from the object, picks up an image of the received reflected light to generate an image pickup signal, and outputs the generated image pickup signal to thevideo processor 31. - The
video processor 31 generates an observation image G1 of an object by performing predetermined processing of an image pickup signal outputted from theendoscope 21 and sequentially outputs the generated observation image G1 to theimage processing apparatus 32 frame by frame. - During a period while the lesion candidate area L1 is not detected by the area-of-
interest detecting portion 34, thedisplay controlling portion 36 performs a process for causing a display image on which an observation image G1 is disposed in a display area D1 on thedisplay screen 41A to be displayed. According to such an operation of thedisplay controlling portion 36, a display image as shown inFIG. 4 is displayed on thedisplay screen 41A of thedisplay apparatus 41, for example, during a period before time Ta inFIG. 3 . Note that it is assumed that the display area D1 is set in advance, for example, as an area having a larger size than a display area D2 to be described later on thedisplay screen 41A.FIG. 4 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - The
display controlling portion 36 performs a process for causing a display image on which an observation image G1 including the lesion candidate area L1 is disposed in the display area D1 on thedisplay screen 41A to be displayed, at a timing when measurement of the continuous detection time period TL is started, that is, at a timing when detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is started. According to such an operation of thedisplay controlling portion 36, for example, a display image as shown inFIG. 5 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing of the time Ta inFIG. 3 . According to the operation of thedisplay controlling portion 36 as described above, for example, a display image as shown inFIG. 6 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing immediately before time Tb after the time Ta inFIG. 3 is reached.FIGS. 5 and 6 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - The
recording portion 36 b of thedisplay controlling portion 36 starts a process for recording observation images G1 sequentially outputted from thevideo processor 31 as recorded images R1 at the timing when measurement of the continuous detection time period TL is started, as recorded images R1 and starts a process for mutually associating and recording the recorded images R1 and pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b. According to such an operation of therecording portion 36 b, the process for mutually associating and recording the recorded images R1 and the pieces of lesion candidate information IL is started at the timing of the time Ta inFIG. 3 . - The
recording portion 36 b of thedisplay controlling portion 36 stops the process for recording observation images G1 sequentially outputted from thevideo processor 31 as recorded images R1, at a timing when measurement of the continuous detection time period TL is stopped, that is, at a timing when detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is interrupted/ceased, and stops the process for mutually associating and recording the recorded images R1 and pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b. According to such an operation of therecording portion 36 b, the process for mutually associating and recording the recorded images R1 and the pieces of lesion candidate information IL is stopped at the timing of the time Tb inFIG. 3 . According to the operation of therecording portion 36 b as described above, for example, during the continuous detection time period TL which corresponds to a period from the time Ta to the time Tb inFIG. 3 , observation images G1 of N (N≥2) frames, including at least the observation images G1 as shown inFIGS. 5 and 6 , are sequentially recorded as recorded images R1. - If detection of the lesion candidate area L1 is interrupted/ceased before the continuous detection time period TL reaches the predetermined time period TH, based on a judgment result outputted from the continuous
detection judging portion 35, thedisplay controlling portion 36 starts a process for displaying a display image on which an observation image G1 is disposed in the display area D1 on thedisplay screen 41A, and recorded images R1 recorded by therecording portion 36 b during the period of measurement of the continuous detection time period TL are sequentially disposed in the display area D2 on thedisplay screen 41A, at a timing when the predetermined time period TH elapses after detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is started. According to such an operation of thedisplay controlling portion 36, for example, a display image on which a recorded image R1 corresponding to an observation image G1 inFIG. 6 is disposed in the display area D2 as shown inFIG. 7 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing of time Tc after the time Tb inFIG. 3 , which corresponds to a timing when the predetermine time period TH elapses after detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is started. Note that it is assumed that the display area D2 is set in advance, for example, as an area having a smaller size than the display area D1 described above on thedisplay screen 41A.FIG. 7 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - After the timing when the predetermined time period TH elapses after detection of the lesion candidate area L1 by the lesion
candidate detecting portion 34 b is started, thedisplay controlling portion 36 performs a process for causing recorded images R1 to be sequentially displayed in order opposite to order of recording of the recorded images R1 by recordingportion 36 b. According to such an operation of thedisplay controlling portion 36, for example, recorded images R1 of N frames sequentially recorded by therecording portion 36 b are displayed in the display area D2 in order of N-th frame→(N−1)-th frame→ . . . →second frame→first frame. - That is, according to the operation of the
display controlling portion 36 as described above, if the continuous detection time period TL is shorter than the predetermined time period TH, a process for, while causing a plurality of observation images G1 sequentially outputted from thevideo processor 31 to be sequentially displayed in the display area D1 on thedisplay screen 41A, causing a plurality of recorded images R1 recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D2 on thedisplay screen 41A in order opposite to order of recording by therecording portion 36 b at a timing of the time Tc inFIG. 3 is started. - Here, for example, as shown in
FIG. 3 , if the continuous detection time period TL is shorter than the predetermined time period TH, a situation may occur in which, although the lesion candidate area L1 is detected by the lesioncandidate detecting portion 34 b, the lesion candidate area L1 moves outside from an inside of an observation image G1 without a marker image G2 being displayed. Therefore, it is conceivable that, if the continuous detection time period TL is shorter than the predetermined time period TH, oversight of the lesion candidate area L1 by a user's visual examination easily occurs. - In comparison, according to the operation of the
display controlling portion 36 as described above, it is possible to, for example, as shown inFIGS. 3 and 7 , cause a recorded image R1 including the lesion candidate area L1 already detected by the lesioncandidate detecting portion 34 b to be displayed in the display area D2 on thedisplay screen 41A when the predetermined time period TH elapses after detection of the lesion candidate area L1 is started. Therefore, according to the operation of thedisplay controlling portion 36 as described above, it is possible to reduce oversight of a lesion that may occur in endoscopic observation. According to the operation of thedisplay controlling portion 36 as described above, since recorded images R1 are displayed in the display area D2 in order opposite to order of recording by therecording portion 36 b, it is possible to inform the user, for example, of a position of the lesion candidate area L1 that has moved outside from an inside of an observation image G1 without a marker image G2 being given. - Note that, according to the present embodiment, it is possible to enable recording of recorded images R1 to be also continued, for example, during a period from the time Tb to the time Tc in
FIG. 3 as long as recording of recorded images R1 is performed during the period of measurement of the continuous detection time period TL. - According to the present embodiment, for example, if the continuous detection time period TL of the lesion candidate area L1 is equal to or longer than the predetermined time period TH, the process corresponding to the time chart of
FIG. 8 may be performed by thedisplay controlling portion 36. Details of such a process will be described below.FIG. 8 is a time chart for illustrating an example of a process performed in the image processing apparatus according to the embodiment, which is different from the example ofFIG. 3 . Note that, hereinafter, specific description of parts to which the processes and the like already described are applicable will be appropriately omitted for simplification. - During a period while the lesion candidate area L1 is not detected by the area-of-
interest detecting portion 34, thedisplay controlling portion 36 performs a process for causing a display image on which an observation image G1 is disposed in a display area D1 on thedisplay screen 41A to be displayed. According to such an operation of thedisplay controlling portion 36, the display image as shown inFIG. 4 is displayed on thedisplay screen 41A of thedisplay apparatus 41, for example, during a period before time Td inFIG. 8 . - The
display controlling portion 36 performs the process for causing a display image on which an observation image G1 including the lesion candidate area L1 is disposed in the display area D1 on thedisplay screen 41A to be displayed, at the timing when measurement of the continuous detection time period TL is started, that is, at the timing when detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is started. According to such an operation of thedisplay controlling portion 36, for example, the display image as shown inFIG. 5 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing of the time Td inFIG. 8 . - The
recording portion 36 b of thedisplay controlling portion 36 starts the process for recording observation images G1 sequentially outputted from thevideo processor 31 as recorded images R1 at the timing when measurement of the continuous detection time period TL is started, as recorded images R1 and starts the process for mutually associating and recording the recorded images R1 and pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b. According to such an operation of therecording portion 36 b, the process for mutually associating and recording the recorded images R1 and the pieces of lesion candidate information IL is started at the timing of the time Td inFIG. 8 . - Based on a judgment result outputted from the continuous
detection judging portion 35, theenhancement processing portion 36 a of thedisplay controlling portion 36 starts an enhancement process of adding a marker image G2 for enhancing a position of the lesion candidate area L1 detected by the lesioncandidate detecting portion 34 b to an observation image G1 at a timing when the predetermined time period TH elapses after measurement of the continuous detection time period TL is started. According to such an operation of theenhancement processing portion 36 a, for example, a display image as shown inFIG. 9 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing of time Te inFIG. 8 , which corresponds to the timing when the predetermined time period TH elapses after detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is started. According to the operation of theenhancement processing portion 36 a as described above, a display image as shown inFIG. 10 is displayed on thedisplay screen 41A of thedisplay apparatus 41 at a timing immediately before time Tf after the time Te inFIG. 8 is reached. Note that, hereinafter, description will be advanced, with a case where a rectangular frame surrounding the lesion candidate area L1 as shown inFIGS. 9 and 10 is added as a marker image G2 by the enhancement process of theenhancement processing portion 36 a given as an example.FIGS. 9 and 10 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - The
recording portion 36 b of thedisplay controlling portion 36 stops the process for recording observation images G1 sequentially outputted from thevideo processor 31 as recorded images R1, at the timing when measurement of the continuous detection time period TL is stopped, that is, at the timing when detection of the lesion candidate area L1 by the lesioncandidate detecting portion 34 b is interrupted/ceased, and stops the process for mutually associating and recording the recorded images R1 and pieces of lesion candidate information IL outputted from the lesioncandidate detecting portion 34 b. According to such an operation of therecording portion 36 b, the process for mutually associating and recording the recorded images R1 and the pieces of lesion candidate information IL is stopped at a timing of the time Tf inFIG. 8 . According to the operation of therecording portion 36 b as described above, for example, during the continuous detection time period TL which corresponds to a period from the time Td to the time Tf inFIG. 8 , observation images G1 of P (P≥2) frames, including at least observation images G1 as shown inFIGS. 9 and 10 , are sequentially recorded as recorded images R1. - If detection of the lesion candidate area L1 is interrupted/ceased after the continuous detection time period TL becomes equal to or longer than the predetermined time period TH, based on a judgment result outputted from the continuous
detection judging portion 35, thedisplay controlling portion 36 starts the process for displaying a display image on which an observation image G1 is disposed in the display area D1 on thedisplay screen 41A, and recorded images R1 recorded by therecording portion 36 b during the period of measurement of the continuous detection time period TL are sequentially disposed in the display area D2 on thedisplay screen 41A, at the timing when detection of the lesion candidate area L1 is interrupted/ceased. According to such an operation of thedisplay controlling portion 36, for example, a display image on which a recorded image R1 corresponding to the observation image G1 inFIG. 10 is displayed in the display area D2, as shown inFIG. 11 , is displayed on thedisplay screen 41A of thedisplay apparatus 41 at the timing of the time Tf inFIG. 8 .FIG. 11 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - After the timing when detection of the lesion candidate area L1 by the lesion
candidate detecting portion 34 b is interrupted/ceased, thedisplay controlling portion 36 performs the process for causing recorded images R1 to be sequentially displayed in order opposite to order of recording of the recorded images R1 by therecording portion 36 b. According to such an operation of thedisplay controlling portion 36, for example, recorded images R1 of P frames sequentially recorded by therecording portion 36 b are displayed in the display area D2 in order of P-th frame→(P−1)-th frame→ . . . →second frame→first frame. - That is, according to the operation of the
display controlling portion 36 as described above, if the continuous detection time period TL is equal or longer than the predetermined time period TH, the process for, while causing a plurality of observation images G1 sequentially outputted from thevideo processor 31 to be sequentially displayed in the display area D1 on thedisplay screen 41A, causing a plurality of recorded images R1 recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D2 on thedisplay screen 41A in order opposite to order of recording by therecording portion 36 b at the timing of the time Tf inFIG. 8 is started. - Here, for example, if a difference time period ΔT between the continuous detection time period TL and the predetermined time period TH, which corresponds to a period from the time Te to the time Tf in
FIG. 8 , is extremely short, a situation may occur in which, without the user visually confirming a marker image G2 instantaneously displayed in the display area D1, the lesion candidate area L1 enhanced by the marker image G2 moves outside from an inside of an observation image G1. Therefore, it is also conceivable that, if the difference time period ΔT is extremely short, oversight of the lesion candidate area L1 by the user's visual examination easily occurs. - In comparison, according to the operation of the
display controlling portion 36 as described above, it is possible to, for example, as shown inFIGS. 8 and 11 , cause a recorded image R1 including the lesion candidate area L1 already detected by the lesioncandidate detecting portion 34 b to be displayed in the display area D2 on thedisplay screen 41A when detection of the lesion candidate area L1 is interrupted/ceased after the predetermined time TH elapses. Therefore, according to the operation of thedisplay controlling portion 36 as described above, it is possible to reduce oversight of a lesion that may occur in endoscopic observation. According to the operation of thedisplay controlling portion 36 as described above, since recorded images R1 are displayed in the display area D2 in order opposite to order of recording by therecording portion 36 b, it is possible to inform the user, for example, of a position of the lesion candidate area L1 that has moved outside from an inside of an observation image G1 after a marker image G2 is instantaneously given. - Note that the operation of the
display controlling portion 36 as described above is also applied in the case of recording and displaying a recording image R1 including a plurality of lesion candidate areas almost similarly. - According to the present embodiment, the process for displaying recorded images R1 in the display area D2 in order opposite to order of recording by the
recording portion 36 b is performed by thedisplay controlling portion 36, but the process is not limited to this. For example, a process for displaying recorded images R1 in the display area D2 in the same order as order of recording by therecording portion 36 b may be performed by thedisplay controlling portion 36. - According to the present embodiment, the process for causing recorded images R1 of respective frames recorded during the period of measurement of the continuous detection time period TL to be sequentially displayed in the display area D2 is performed by the
display controlling portion 36, but the process is not limited to this. A process for causing recorded images R1 of a part of frames among the recorded images R1 of the respective frames to be displayed in the display area D2 may be performed by thedisplay controlling portion 36. More specifically, for example, a process for causing only a recorded image R1 corresponding to one frame recorded last among the recorded images R1 of the respective frames recorded during the period of measurement of the continuous detection time period TL to be displayed in the display area D2 may be performed by thedisplay controlling portion 36. Or, for example, a process for, while decimating the plurality of recorded images R1 recorded during the period of measurement of the continuous detection time period TL at predetermined intervals, causing the recorded images R1 to be displayed in the display area D2 may be performed by thedisplay controlling portion 36. - According to the present embodiment, for example, the
recording portion 36 b may record respective observation images G1 obtained by decimating a plurality of observation images G1 sequentially outputted from thevideo processor 31 during the period of measurement of the continuous detection time period TL at predetermined intervals, as recorded images R1. - According to the present embodiment, the
recording portion 36 b is not limited to such a configuration that sequentially records observation images G1 of a plurality of frames as recorded images R1 but may, for example, record only an observation image of one frame as a recorded image R1. More specifically, therecording portion 36 b may, for example, record only an observation image G1 inputted to thedisplay controlling portion 36 at a timing immediately before measurement of the continuous detection time period TL is stopped (the observation images G1 shown inFIGS. 6 and 10 ) as a recorded image R1. - According to the present embodiment, when the
display controlling portion 36 causes a recorded image R1 of each frame recorded in therecording portion 36 b to be displayed in the display area D2, thedisplay controlling portion 36 may cause the recorded image R1 to be equal-speed displayed at the same frame rate as a frame rate at the time of recording by therecording portion 36 b, may cause the recorded image R1 to be double-speed displayed at a frame rate higher than the frame rate at the time of recording by therecording portion 36 b, or may cause the recorded image R1 to be slow-displayed at a frame rate lower than the frame rate at the time of recording by therecording portion 36 b. - According to the present embodiment, when the
display controlling portion 36 causes the recorded image R1 recorded in therecording portion 36 b to be displayed in the display area D2, thedisplay controlling portion 36 may cause the recorded image R1 to be displayed in the same color as color at the time of recording by therecording portion 36 b, may cause the recorded image R1 to be displayed in subtractive color obtained from the color at the time of recording by therecording portion 36 b, or may cause the recorded image R1 to be displayed only in predetermined one color. - According to the present embodiment, for example, recording of recorded images R1 may be started at a desired timing before the timing when detection of the lesion candidate area L1 is started as long as recording of recorded images R1 is performed during the period of measurement of the continuous detection time period TL.
- According to the present embodiment, when the
display controlling portion 36 causes a recorded image R1 recorded in therecording portion 36 b to be displayed in the display area D2, an enhancement process for generating a marker image G2 based on lesion candidate information IL recorded in a state of being associated with the recorded image R1 and adding the generated marker image G2 to the recorded image R1 may be performed by theenhancement processing portion 36 a. According to such an operation of thedisplay controlling portion 36, for example, it is possible to, at the timing of the time Tc inFIG. 3 or at the timing of the time Tf inFIG. 8 , cause a display image on which a position of the lesion candidate area L1 included in the recorded image R1 in the display area D2 is enhanced as shown inFIG. 12 to be displayed on thedisplay screen 41A.FIG. 12 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - According to the present embodiment, a process for causing a display image on which a recorded image R1 recorded in the
recording portion 36 b is disposed in the display area D1 to be temporarily displayed instead of an observation image G1, and causing each of such recorded images R1 displayed in the display area D1 instead of an observation image G1 to be sequentially redisplayed in the display area D2 may be performed by thedisplay controlling portion 36. According to such an operation of thedisplay controlling portion 36, for example, it is possible to, at the timing of the time Tc inFIG. 3 or at the timing of the time Tf inFIG. 8 , cause a display image on which a position of the lesion candidate area L1 included in the recorded image R1 in the display area D1 is enhanced as shown inFIG. 13 to be displayed on thedisplay screen 41A. According to the operation of thedisplay controlling portion 36 as described above, for example, it is possible to cause a display image as shown inFIG. 14 to be displayed on thedisplay screen 41A after display of each recorded image R1 in the display area D1 is completed.FIGS. 13 and 14 are diagrams showing examples of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - According to the present embodiment, for example, if the display area D2 is not set on the
display screen 41A, a process for causing a recording image R1 and/or a marker image G2 to be displayed in the display area D1 may be performed by thedisplay controlling portion 36. - More specifically, for example, a process for generating a composite image GR1 by combining an observation image G1 sequentially outputted from the
video processor 31 and a recorded image R1 recorded in therecording portion 36 b, generating a marker image G2 based on lesion candidate information IL associated with the recorded image R1, and adding the marker image G2 to the composite image GR1 to cause the marker image G2 and the composite image GR1 to be displayed in the display area D1 may be performed by thedisplay controlling portion 36. According to such an operation of thedisplay controlling portion 36, for example, it is possible to, at the timing of the time Tc inFIG. 3 or at the timing of the time Tf inFIG. 8 , cause a display image on which a position of the lesion candidate area L1 included in the composite image GR1 in the display area D1 is enhanced as shown inFIG. 15 to be displayed on thedisplay screen 41A.FIG. 15 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - Otherwise, for example, a process for generating a marker image G2 based on lesion candidate information IL recorded in the
recording portion 36 b, and adding the generated marker image G2 to an observation image G1 sequentially outputted from thevideo processor 31 to cause the marker image G2 and the observation image G1 to be displayed in the display area D1 may be performed by thedisplay controlling portion 36. According to such an operation of thedisplay controlling portion 36, for example, it is possible to cause a display image as shown inFIG. 16 to be displayed on thedisplay screen 41A of thedisplay apparatus 41 at the timing of the time Tc inFIG. 3 or at the timing of the time Tf inFIG. 8 .FIG. 16 is a diagram showing an example of a display image displayed on the display apparatus after a process of the image processing apparatus according to the embodiment is performed. - The image processing apparatus and the like according to the present embodiment may include a processor and a storage (e.g., a memory). The functions of individual units in the processor may be implemented by respective pieces of hardware or may be implemented by an integrated piece of hardware, for example. The processor may include hardware, and the hardware may include at least one of a circuit for processing digital signals and a circuit for processing analog signals, for example. The processor may include one or a plurality of circuit devices (e.g., an IC) or one or a plurality of circuit elements (e.g., a resistor, a capacitor) on a circuit board, for example. The processor may be a CPU (Central Processing Unit), for example, but this should not be construed in a limiting sense, and various types of processors including a GPU (Graphics Processing Unit) and a DSP (Digital Signal Processor) may be used. The processor may be a hardware circuit with an ASIC or an FPGA. The processor may include an amplification circuit, a filter circuit, or the like for processing analog signals. The memory may be a semiconductor memory such as an SRAM and a DRAM; a register; a magnetic storage device such as a hard disk device; and an optical storage device such as an optical disk device. The memory stores computer-readable instructions, for example. When the instructions are executed by the processor, the functions of each unit of the image processing device and the like are implemented. The instructions may be a set of instructions constituting a program or an instruction for causing an operation on the hardware circuit of the processor.
- The units in the image processing apparatus and the like and the display device according to the present embodiment may be connected with each other via any types of digital data communication such as a communication network or via communication media. The communication network may include a LAN (Local Area Network), a WAN (Wide Area Network), and computers and networks which form the internee, for example.
Claims (20)
1. An image processing apparatus comprising a processor, the processor being configured to:
perform a process for detecting an area of interest for each of a plurality of observation images that are sequentially inputted;
record the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and
perform a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased.
2. The image processing apparatus according to claim 1 , wherein, if the time period until the detection of the area of interest is interrupted/ceased after the detection is started is shorter than the predetermined time period, the processor performs the process for, while causing the plurality of observation images to be sequentially displayed on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed on the display screen at the first timing.
3. The image processing apparatus according to claim 1 , wherein, if the time period until the detection of the area of interest is interrupted/ceased after the detection is started is equal to or longer than the predetermined time period, the processor performs the process for, while causing the plurality of observation images to be sequentially displayed on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed on the display screen at the second timing.
4. The image processing apparatus according to claim 1 , wherein
the processor mutually associates and records position information showing a position of the area of interest and the one or more recorded images during the time period; and
the processor performs an enhancement process for enhancing the position of the area of interest included in the at least one recorded image caused to be displayed on the display screen, based on the position information, either at the first timing or at the second timing.
5. The image processing apparatus according to claim 1 , wherein the processor performs a process for, while causing the plurality of observation images to be sequentially displayed in a first display area on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed in a second display area on the display screen.
6. The image processing apparatus according to claim 5 , wherein the second display area is set as an area having a smaller size than the first display area on the display screen.
7. The image processing apparatus according to claim 1 , wherein the processor performs a process for causing the one or more recorded images to be sequentially displayed on the display screen in order opposite to order of recording the one or more recorded images.
8. The image processing apparatus according to claim 1 , wherein the processor performs a process for causing a recorded image recorded last, among the one or more recorded images, to be displayed on the display screen.
9. The image processing apparatus according to claim 1 , wherein the processor performs a process for causing the one or more recorded images to be displayed on the display screen while decimating the one or more recorded images at predetermined intervals.
10. The image processing apparatus according to claim 1 , wherein the processor records each of observation images obtained by decimating the plurality of observation images at predetermined intervals as a recorded image during the time period.
11. An image processing method comprising:
performing a process for detecting an area of interest for each of a plurality of observation images obtained by performing image pickup of an object;
recording the plurality of observation images as one or more recorded images during a time period until detection of the area of interest is interrupted/ceased after the detection is started; and
performing a process for, while causing the plurality of observation images to be sequentially displayed on a display screen of a display apparatus, causing at least one recorded image among the one or more recorded images to be displayed on the display screen at a first timing when a predetermined time period elapses after the detection of the area of interest is started or at a second timing when the detection of the area of interest is interrupted/ceased.
12. The image processing method according to claim 11 , wherein, if a time period until the detection of the area of interest is interrupted/ceased after the detection is started is shorter than the predetermined time period, the process for, while causing the plurality of observation images to be sequentially displayed on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed on the display screen is performed at the first timing.
13. The image processing method according to claim 11 , wherein, if the time period until the detection of the area of interest is interrupted/ceased after the detection is started is equal to or longer than the predetermined time period, the process for, while causing the plurality of observation images to be sequentially displayed on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed on the display screen is performed at the second timing.
14. The image processing method according to claim 11 , wherein
position information showing a position of the area of interest and the one or more recorded images are mutually associated and recorded during the time period; and
an enhancement process for enhancing the position of the area of interest included in the at least one recorded image caused to be displayed on the display screen, based on the position information is performed either at the first timing or at the second timing.
15. The image processing method according to claim 11 , wherein a process for, while causing the plurality of observation images to be sequentially displayed in a first display area on the display screen, causing the at least one recorded image among the one or more recorded images to be displayed in a second display area on the display screen is performed.
16. The image processing method according to claim 15 , wherein the second display area is set as an area having a smaller size than the first display area on the display screen.
17. The image processing method according to claim 11 , wherein a process for causing the one or more recorded images to be sequentially displayed on the display screen in order opposite to order of recording of the one or more recorded images is performed.
18. The image processing method according to claim 11 , wherein a process for causing a recorded image recorded last among the one or more recorded images to be displayed on the display screen is performed.
19. The image processing method according to claim 11 , wherein a process for causing the one or more recorded images to be displayed on the display screen while decimating the one or more recorded images at predetermined intervals is performed.
20. The image processing method according to claim 11 , wherein each of observation images obtained by decimating the plurality of observation images at predetermined intervals is recorded as a recorded image during the time period.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2016/067924 WO2017216922A1 (en) | 2016-06-16 | 2016-06-16 | Image processing device and image processing method |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/067924 Continuation WO2017216922A1 (en) | 2016-06-16 | 2016-06-16 | Image processing device and image processing method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190114738A1 true US20190114738A1 (en) | 2019-04-18 |
Family
ID=60663971
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/213,246 Abandoned US20190114738A1 (en) | 2016-06-16 | 2018-12-07 | Image processing apparatus and image processing method |
Country Status (3)
Country | Link |
---|---|
US (1) | US20190114738A1 (en) |
JP (1) | JPWO2017216922A1 (en) |
WO (1) | WO2017216922A1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210149182A1 (en) * | 2018-07-06 | 2021-05-20 | Olympus Corporation | Image processing apparatus for endoscope, image processing method for endoscope, and recording medium |
US20210244260A1 (en) * | 2018-09-11 | 2021-08-12 | Sony Corporation | Medical observation system, medical observation apparatus and medical observation method |
EP3841958A4 (en) * | 2018-08-20 | 2021-10-13 | FUJIFILM Corporation | Endoscopic system and medical image processing system |
US20220133214A1 (en) * | 2019-02-14 | 2022-05-05 | Nec Corporation | Lesion area dividing device, medical image diagnostic system, lesion area dividing method, and non-transitory computer-readable medium storing program |
US11436726B2 (en) * | 2018-08-20 | 2022-09-06 | Fujifilm Corporation | Medical image processing system |
US11950760B2 (en) | 2018-05-17 | 2024-04-09 | Fujifilm Corporation | Endoscope apparatus, endoscope operation method, and program |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113365545A (en) * | 2019-02-13 | 2021-09-07 | 奥林巴斯株式会社 | Image recording apparatus, image recording method, and image recording program |
WO2021044910A1 (en) | 2019-09-03 | 2021-03-11 | 富士フイルム株式会社 | Medical image processing device, endoscope system, medical image processing method, and program |
WO2021199294A1 (en) * | 2020-03-31 | 2021-10-07 | 日本電気株式会社 | Information processing device, display method, and non-transitory computer-readable medium having program stored therein |
JP2023178526A (en) * | 2020-09-15 | 2023-12-18 | 富士フイルム株式会社 | Image processing device, endoscope system, operation method of image processing device, and image processing device program |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060155174A1 (en) * | 2002-12-16 | 2006-07-13 | Arkady Glukhovsky | Device, system and method for selective activation of in vivo sensors |
US20100182412A1 (en) * | 2007-07-12 | 2010-07-22 | Olympus Medical Systems Corp. | Image processing apparatus, method of operating image processing apparatus, and medium storing its program |
US20150031954A1 (en) * | 2012-06-08 | 2015-01-29 | Olympus Medical Systems Corp. | Capsule endoscope apparatus and receiving apparatus |
US20170032578A1 (en) * | 2011-06-28 | 2017-02-02 | Kyocera Corporation | Display device |
US20190231167A1 (en) * | 2016-03-14 | 2019-08-01 | Endochoice, Inc. | System and method for guiding and tracking a region of interest using an endoscope |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05115432A (en) * | 1991-10-24 | 1993-05-14 | Olympus Optical Co Ltd | Endoscopic system |
JP2001238205A (en) * | 2000-02-24 | 2001-08-31 | Olympus Optical Co Ltd | Endoscope system |
JP6053673B2 (en) * | 2011-04-28 | 2016-12-27 | オリンパス株式会社 | Fluorescence observation apparatus and image display method thereof |
JP6137921B2 (en) * | 2013-04-16 | 2017-05-31 | オリンパス株式会社 | Image processing apparatus, image processing method, and program |
JP6323183B2 (en) * | 2014-06-04 | 2018-05-16 | ソニー株式会社 | Image processing apparatus and image processing method |
-
2016
- 2016-06-16 JP JP2018523117A patent/JPWO2017216922A1/en active Pending
- 2016-06-16 WO PCT/JP2016/067924 patent/WO2017216922A1/en active Application Filing
-
2018
- 2018-12-07 US US16/213,246 patent/US20190114738A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060155174A1 (en) * | 2002-12-16 | 2006-07-13 | Arkady Glukhovsky | Device, system and method for selective activation of in vivo sensors |
US20100182412A1 (en) * | 2007-07-12 | 2010-07-22 | Olympus Medical Systems Corp. | Image processing apparatus, method of operating image processing apparatus, and medium storing its program |
US20170032578A1 (en) * | 2011-06-28 | 2017-02-02 | Kyocera Corporation | Display device |
US20150031954A1 (en) * | 2012-06-08 | 2015-01-29 | Olympus Medical Systems Corp. | Capsule endoscope apparatus and receiving apparatus |
US20190231167A1 (en) * | 2016-03-14 | 2019-08-01 | Endochoice, Inc. | System and method for guiding and tracking a region of interest using an endoscope |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11950760B2 (en) | 2018-05-17 | 2024-04-09 | Fujifilm Corporation | Endoscope apparatus, endoscope operation method, and program |
US20210149182A1 (en) * | 2018-07-06 | 2021-05-20 | Olympus Corporation | Image processing apparatus for endoscope, image processing method for endoscope, and recording medium |
US11656451B2 (en) * | 2018-07-06 | 2023-05-23 | Olympus Corporation | Image processing apparatus for endoscope, image processing method for endoscope, and recording medium |
EP3841958A4 (en) * | 2018-08-20 | 2021-10-13 | FUJIFILM Corporation | Endoscopic system and medical image processing system |
US11436726B2 (en) * | 2018-08-20 | 2022-09-06 | Fujifilm Corporation | Medical image processing system |
US11867896B2 (en) | 2018-08-20 | 2024-01-09 | Fujifilm Corporation | Endoscope system and medical image processing system |
US20210244260A1 (en) * | 2018-09-11 | 2021-08-12 | Sony Corporation | Medical observation system, medical observation apparatus and medical observation method |
US11969144B2 (en) * | 2018-09-11 | 2024-04-30 | Sony Corporation | Medical observation system, medical observation apparatus and medical observation method |
US20220133214A1 (en) * | 2019-02-14 | 2022-05-05 | Nec Corporation | Lesion area dividing device, medical image diagnostic system, lesion area dividing method, and non-transitory computer-readable medium storing program |
Also Published As
Publication number | Publication date |
---|---|
WO2017216922A1 (en) | 2017-12-21 |
JPWO2017216922A1 (en) | 2019-04-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190114738A1 (en) | Image processing apparatus and image processing method | |
US20200065970A1 (en) | Image processing apparatus and storage medium | |
US10893792B2 (en) | Endoscope image processing apparatus and endoscope image processing method | |
US11176665B2 (en) | Endoscopic image processing device and endoscopic image processing method | |
US20190069757A1 (en) | Endoscopic image processing apparatus | |
US11871903B2 (en) | Endoscopic image processing apparatus, endoscopic image processing method, and non-transitory computer readable recording medium | |
US11100645B2 (en) | Computer-aided diagnosis apparatus and computer-aided diagnosis method | |
US9898664B2 (en) | Image processing device, image processing method, and information storage device | |
US7965876B2 (en) | Systems and methods for image segmentation with a multi-stage classifier | |
US11341637B2 (en) | Endoscope image processing device and endoscope image processing method | |
US20080303898A1 (en) | Endoscopic image processing apparatus | |
ATE441360T1 (en) | INTRAVASCULAR IMAGING | |
US10758206B2 (en) | Method and system for enhanced visualization of lung sliding by automatically detecting and highlighting lung sliding in images of an ultrasound scan | |
EP2366327A3 (en) | An electronic endoscope system, an electronic endoscope processor, and a method of acquiring blood vessel information | |
US20210338042A1 (en) | Image processing apparatus, diagnosis supporting method, and recording medium recording image processing program | |
US20140313320A1 (en) | Image pickup apparatus | |
JP2020531099A5 (en) | ||
US20140378836A1 (en) | Ultrasound system and method of providing reference image corresponding to ultrasound image | |
US11992177B2 (en) | Image processing device for endoscope, image processing method for endoscope, and recording medium | |
JP4855912B2 (en) | Endoscope insertion shape analysis system | |
JP2011244884A (en) | Endoscopic system | |
JP2002000546A (en) | Endoscopic system | |
JP2018027401A5 (en) | Diagnostic system and information processing apparatus | |
WO2020110310A1 (en) | Display device and endoscope system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: OLYMPUS CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONODA, YASUKO;REEL/FRAME:047708/0308 Effective date: 20181003 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |