AU2007221810B2 - Image display unit, image display method and image display program - Google Patents
Image display unit, image display method and image display program Download PDFInfo
- Publication number
- AU2007221810B2 AU2007221810B2 AU2007221810A AU2007221810A AU2007221810B2 AU 2007221810 B2 AU2007221810 B2 AU 2007221810B2 AU 2007221810 A AU2007221810 A AU 2007221810A AU 2007221810 A AU2007221810 A AU 2007221810A AU 2007221810 B2 AU2007221810 B2 AU 2007221810B2
- Authority
- AU
- Australia
- Prior art keywords
- image
- images
- image display
- display apparatus
- displayed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Landscapes
- Endoscopes (AREA)
Description
Australian Patents Act 1990 Regulation 3.2 ORIGINAL COMPLETE SPECIFICATION STANDARD PATENT Invention Title ""Image display unit, image display method and image display program."" The following statement is a full description of this invention, including the best method of performing it known to me/us:- P/00/011 Q.\OPER\SEW\2OO7\OctoberUO327436 lodge repon diviionaidoc 3/10107
DESCRIPTION
0 Image display apparatus, image display method, and image display program TECHNICAL FIELD 00 The present invention relates to an image display apparatus, an image display method, and an image display program.
S BACKGROUND ART Recently, swallowable capsule endoscopes have been produced as a type of endoscopes. The capsule endoscopes are provided with an imaging capability and a radio capability. A capsule endoscope is configured to sequentially take images of organs such as the stomach and the small intestine within an observation period from the time it has been swallowed through the mouth of a patient for observation (examination) to its natural excretion from the human body (see Japanese Patent Application Laid-open No. H11-225996 Publication).
During the observation period, image data taken in a body by the capsule endoscope is sequentially transmitted outside through radio communication and is stored in a memory. Since a patient carries around a receiver having a radio communication capability and a memory capability, the patient can freely perform normal actions during the observation period from swallowing of the capsule endoscope to its excretion. After observation, a doctor or a nurse can display the images of organs on a display based on the image data stored in the memory and use it to make a diagnosis.
As the above type of capsule endoscope, "M2A
O
(registered trademark)" by Given Imaging Ltd. of Israel, o and "NORIKA (registered trademark)" by RF SYSTEM lab. of SJapan are presently available, and they have already come to practical applications.
However, unlike an ordinary endoscope, the capsule endoscope described above takes images of each organ within a period from the time a subject swallows to its natural excretion, meaning an extended period of observation (examination), for example, not less than ten hours.
Therefore, the number of images to be taken in time sequence is correspondingly huge.
At the stage of diagnosis or the like, no particular consideration is given to improving the ability to retrieve a desired image from the vast amount of images taken over a long period of time, or providing a display screen allowing easy recognition of what time in the overall imaging period the displayed image was taken, of which organ is being shown, and the like.
The object of the present invention is to provide an image display apparatus, image display method, and image display program which can improve the ability to retrieve taken internal images of a body and ensure easy recognition of the organ depicted in each image.
DISCLOSURE OF INVENTION To solve the above problems and to realize the object, an image display apparatus according to the present invention includes an input unit that inputs image data taken in time sequence by an in-vivo imaging device, a scale display control unit that controls to display a scale indicating an overall imaging period of input image data taken in time sequence and input by the input unit, a color information detecting unit that detects color information of a screen of the image data input by the input unit, a o color display control unit that controls to display a color corresponding to the color information detected by the color information detecting unit at a time-corresponding position on the scale, an image display control unit that controls to display an image corresponding to the image data input by the input unit, an image designation unit that designates the image subjected to be displayed by the image display control unit, and an index display control unit that controls to display, on the scale, an index indicating a position corresponding to an imaging time of the image designated by the image designation unit.
In the present invention, the color information detecting unit may include an average color detecting unit that detects color information on average color from color information of a screen of the image data input by the input unit.
The present invention further includes a designated image display control unit that controls to display the image designated by the image designation unit. The index display control unit and the designated image display control unit are configured to make correlation display indicating a correlation between the index displayed on the scale and the displayed designated image.
The present invention further includes an organ discriminating unit that discriminates an organ based on the color information detected by the color information detecting unit, and an organ name display control unit that controls to display a name of the organ discriminated by the organ discriminating unit in association with the scale.
In the present invention, the organ discriminating unit may include a discoloration edge detecting unit that detects a discoloration edge from the color information detected by the color information detecting unit.
C.)
o The present invention includes an inputting step for inputting image data taken in time sequence by an in-vivo imaging device, a scale display control step for controlling to display a scale indicating an overall imaging period of input image data taken in time sequence 0and input in the inputting step, a color information detecting step for detecting color information of a screen of the image data input in the inputting step, a color display control step for controlling to display a color corresponding to the color information detected by the color information detecting unit at a time-corresponding position on the scale, an image display control step for controlling to display an image corresponding to the image data input in the inputting step, an image designating step for designating the image subjected to be displayed in the image display control step, and an index display control step for controlling to display, on the scale, an index indicating a position corresponding to an imaging time of the image designated in the image designating step.
In the present invention, the color information detecting step may include an average color detecting step for detecting color information on average color from color information of a screen of the image data input in the inputting step.
The present invention further includes a designated image display controlling step for controlling to display the image designated in the image designating step. The index display control step and the image designating step are configured to make correlation display indicating a correlation between the index displayed on the scale and the displayed designated image.
The present invention further includes an organ discriminating step for discriminating an organ based on
C)
o the color information detected in the color information detecting step, and an organ-name display control step for controlling to display a name of the organ discriminated in the organ discriminating step in association with the scale.
In the present invention, the organ discriminating step may include detecting a discoloration edge from the color information detected in the color information detecting step.
BRIEF DESCRIPTION OF DRAWINGS Fig. 1 is a schematic diagram of the internal structure of a capsule endoscope according to an embodiment of the present invention; Fig. 2 is a schematic diagram of a capsule endoscope system according to the embodiment; Fig. 3 is a block diagram showing a structural example of the capsule endoscope system according to the embodiment; Fig. 4 shows an example of screen transition associated with the observation procedures according to the embodiment; Fig. 5 shows an example of screen transition associated with the observation procedures according to the embodiment; Fig. 6 shows an example of screen transition associated with the observation procedures according to the embodiment; Fig. 7 shows an example of screen transition associated with the diagnosis procedures according to the embodiment; Fig. 8 shows an example of screen transition associated with the diagnosis procedures according to the
O
embodiment; o Fig. 9 is a flowchart of the operation for average Scolor bar display according to the embodiment; Fig. 10 shows an example of a display screen associated with a diagnosis process according to a modification of the embodiment; 00 Fig. 11 illustrates the principle of automatic discrimination of organ names according to the modification Sof the embodiment; Fig. 12 is a flowchart of the procedures of discriminating the organ names according to the modification of the embodiment; Fig. 13 shows an example of application of the modification shown in Fig. 11; Fig. 14 shows an example of screen transition associated with the diagnosis procedures according to the embodiment; and Fig. 15 is a flowchart of an operation for displaying the imaging time of a designated image according to the embodiment.
BEST MODE(S) FOR CARRYING OUT THE INVENTION Exemplary embodiments of the present invention are described below with reference to the drawings.
The overall structure of a capsule endoscope according to one embodiment of the present invention is described with reference to Fig. 1. Fig. 1 is a schematic diagram of the internal structure of the capsule endoscope according to this embodiment of the invention. As shown in Fig. i, a capsule endoscope 10 includes an imaging unit 111 which can take the internal image of a celom, illumination units 112a and 112b which illuminate the interior of the celom, a power supply unit 13 which supplies them with power, and a capsule housing 14 which has at least the imaging unit 111,
C.)
o the illumination units 112 and the power supply unit 13 M disposed inside.
The capsule housing 14 according to this embodiment includes a distal-end cover 120 which covers the imaging unit 111 and the illumination units 112a, 112b, and a 00 capsule body 122 which is provided in a water-proof state with respect to the distal-end cover 120 via a seal member 121 and has the imaging unit 111, etc. disposed therein. A 1C 0 rear-end cover 123 may be provided as separate from the capsule body 122 as needed. Although the rear-end cover 123 is provided integrally with the capsule body and has a flat shape in this embodiment, the shape is not limited and may be, for example, a dome shape.
The distal-end cover 120 may clearly separate an illumination window 120a, which transmits illumination light L from the illumination unit 112a, 112b, and an imaging window 120b, which performs imaging in the illumination range, from each other. In this embodiment, the entire distal-end cover 120 is transparent and the areas of the illumination window 120a and the imaging window 120b partly overlap each other.
The imaging unit 111 is provided on an imaging board 124 with a solid-state imaging device 125 formed of, for example, a CCD, which performs imaging in the range that is illuminated with the illumination light L from the illumination unit 112a, 112b, and an image forming lens 126 which includes a fixed lens 126a and a movable lens 126b, and forms the image of a subject to the solid-state imaging device 125, and executes sharp image forming with a focus adjusting unit 128 with a fixed frame 128a which secures the fixed lens 126a and a movable frame 128b, which secures the movable lens 126b. In the present invention, the
O
imaging unit 111 is not limited to the CCD, but an imaging
C.)
o unit such as CMOS, may be used.
The illumination units 112a, 112b are provided on an illumination board 130 and are comprised of, for example, a light-emitting diode (LED), and a plurality of illumination 00 units 112a, 112b (four in this embodiment as one example) are laid out around the image forming lens 126 which constitutes the imaging unit 111. In the present invention, the illumination units 112a, 112b are not limited to the LED but other illumination units may be used as well.
The power supply unit 13 is provided on a power supply board 132 provided with an internal switch 131 and uses, for example, a button type battery as a power supply 133.
While a silver oxide cell, for example, is used as the battery in the present invention, the invention is not limited to it and may use a chargeable battery, a dynamo type battery or the like.
Although one which can perform an ON operation by, for example, the oppositional action of magnets is used as the internal switch 131, the present invention is not limited to this type and other switch units can be also exemplified.
In this embodiment, besides the individual units described above, a radio unit 142 comprising an antenna or the like for radio communication with outside is provided on a radio board 141 and communication with outside is carried out as needed.
A signal processing/control unit 143 for processing or controlling the individual units is provided on an imaging board 124 and executes various processes in the capsule endoscope The signal processing/control unit 143 includes a partial function of video signal processing, a transmission signal generating function which performs mixing of a video C(1 signal and a sync signal, affixing of an error correction o code, etc., a modulation function which performs conversion to, for example, PSK, MSK, GMSK, QMSK, ASK, AM, or FM system, a power supply control function which controls power supply according to ON-OFF of a switch, driver circuits such as an LED driver circuit, a timing generator F- (TG) function which controls the number of imaging shots C and a memory function which stores various data, such as Sparameters for setting the number of imaging shots, and the C 10 like, and executes various signal processes/controls.
The video signal processing function includes processes, such as image data correction white balance (WB) correction, y correction, color processing, and AGC), and correlation double sampling or analog-digital conversion (ADC) and an auto exposure function (AE) if necessary.
Besides the radio unit 142, for example, information collecting units, such as various sensors, a chemical releasing unit which releases chemicals, a tissue collecting unit which cuts tissues in a celom and collects them, etc. may be disposed in the capsule endoscope 10 as needed.
A capsule endoscope system according to this embodiment is described with reference to Fig. 2. Fig. 2 is a schematic diagram of a capsule endoscope system according to this embodiment. At the time of performing examination using the capsule endoscope 10, the capsule endoscope system as shown in Fig. 2 is used.
The capsule endoscope system according to this embodiment comprises the capsule endoscope 10 and its package 50, a jacket 3 which a patient or a subject 2 wears, a receiver 4 attachable to/detachable from the jacket 3, a work station 5, a CF (compact flash (registered trademark))
O
memory reader/writer 6, a label printer 7, a database 8, o and a network 9, as shown in Fig. 2, for example.
The jacket 3 is provided with antennas 31, 32, 33, and 34 which catch radio waves of taken images to be sent from the radio unit 142 of the capsule endoscope 10 so that the O jacket 3 can communicate with the receiver 4 wirelessly or by a cable. The number of antennas is not particularly limited to four but should be plural, so that radio waves according to positions of the capsule endoscope 10 moved can be received properly.
The receiver 4 is provided with an antenna 41 which is used when directly receiving taken images through radio waves, a display unit 42 which displays information necessary for observation (examination) and an input unit 43 which inputs information necessary for observation (examination). A CF memory 44 which stores received taken image data can be detachably attached to the receiver 4.
Further, the receiver 4 is provided with a power supply unit 45 capable of supplying power even at the time of portable usage and a signal processing/control unit 46 which performs processes needed for observation (examination). As the power supply unit 45, for example, a dry cell, Li ion secondary battery, and Ni hydrogen battery can be exemplified and a chargeable type may also be used.
The work station 5 has a processing function for performing a diagnosis based on images of organs or the like in a patient, taken by the capsule endoscope 10 by a doctor or a nurse. This work station 5 has interfaces, though not shown, which connect to the receiver 4, the CF memory reader/writer 6, and the label printer 7 in a communicable manner and executes read/write of the CF memory 44, chart printing, etc.
The work station 5 has a communication function for O connecting to the network 9 and stores doctor results of a O patient into the database 8 via the network 9. Further, M the work station 5 has a display unit 51, and receives 0 taken image data of inside a patient from the receiver 4 and displays the images of organs or the like on the display unit 51.
00 As shown in Fig. 2, as the capsule endoscope 10 is 1 taken out of the package 50 and is swallowed by the subject 2 through the mouth, prior to initiation examination, it 10 passes through the esophagus, moves inside the celom by peristalsis of the digestive tracts and takes images inside the celom one after another.
The radio waves of taken images are output via the radio unit 142 as needed or for the imaging results and are caught by the antennas 31, 32, 33, and 34 of the jacket 3.
A signal from the antenna the intensity of whose received radio waves is high is sent to the receiver 4 outside.
In the receiver 4, taken image data received one after another is stored in the CF memory 44. The receiver 4 is not synchronized with the start of imaging of the capsule endoscope 10 and the initiation of reception and end of reception are controlled by manipulation of the input unit 43. The taken image data may be still picture data taken by plural frames per second for dynamic display or ordinary moving picture data.
When observation (examination) of the subject 2 by the capsule endoscope 10 is finished, the taken image data stored in the CF memory 44 is transferred to the work station 5 via a cable. The work station 5 memorizes the transferred taken image data in association with individual patients.
The taken image data inside the celom taken by the capsule endoscope 10 and stored in the receiver 4 in this manner is displayed by the display unit 51 of the work o station 5. Accordingly, acquisition of effective data for 0 physiological study and diagnosis of lesion can be carried out over the entire digestive tracts of a human body including the deep body portion (small intestine, etc.) which cannot be reached by an ultrasonic probe, endoscope, etc.
The processing system of the capsule endoscope system is described with reference to Fig. 3. Fig. 3 is a block diagram showing a structural example of the interior of the capsule endoscope system according to this embodiment. The description is given on only the essential structures of the individual units.
As already explained using Fig. 1, the capsule endoscope 10 has the structure to take the image of an internal target (organs, etc.) with the imaging unit 111 from reflection of light illuminated from the illumination units 112a and 112b and send the taken image from the radio unit 142 in the form of a radio signal.
The jacket 3 has a structure such that a selector is connected to the four antennas 31, 32, 33, 34, and an I/F 36 to which a cable to connect to the receiver 4 is connected to the selector 35. The jacket 3 receives radio signals sent from the capsule endoscope 10 at the four antennas 31, 32, 33, and 34, select a received signal according to the radio wave intensity by the selector and is transferred to the receiver 4 via the I/F 36. The jacket 3 is not provided with a large-capacity memory and taken images received via the antennas 31, 32, 33, and 34 are transferred one after another to the receiver 4 at the subsequent stage.
The receiver 4 has, as the internal structure, an I/F for communication to the I/F 36 of the jacket 3 via a I I 13 cable, a CPU 46 which controls the entire receiver 4 O according to a program prepared beforehand, a CF memory I/F 47 which performs data communication with the attached
CF
memory 44, and an I/F 48 which performs communication with the work station 5 by a cable.
To secure the state of being capable of receiving taken images from the jacket 3 at any time, the receiver 4 is always attached to the subject 2 during observation of inside a body by the capsule endoscope 10. During 1C 0 observation, therefore, taken images are received one after another from the jacket 3 and the received images are stored in the CF memory 44 via the CF memory I/F 47 one after another. During observation, the receiver 4 is not connected to the work station 5 and the subject 2 is not restricted in a hospital or the like and can move freely.
The CF memory reader/writer 6 has, as the internal structure, a CPU 61 which controls the entire reader/writer according to a program prepared beforehand, a CF memory I/F 62 which performs data communication with the attached CF memory 44, and an I/F 63 which performs communication with the work station 5 by a cable.
The CF memory reader/writer 6 is attached with the CF memory 44 and is connected to the work station 5 via the I/F 63, performs formatting of taken information for diagnosis according to this embodiment with respect to the CF memory 44 or reads stored taken image data from the CF memory 44 and transfers the data to the work station The taken image data here is in the form of JPEG or the like.
According to this embodiment, as apparent from the above, it is possible to arbitrarily select direct transfer of taken image data to the work station 5 from the receiver 4 or moving the CF memory 44 to the CF memory reader/writer 14 C1 6 to transfer taken image data to the work station O The work station 5 has the display unit 51 which 0 displays images of organs, etc. according to this embodiment, an I/F 52 which manages communication with the I/F 48 of the receiver 4 via a cable and the I/F 63 of the CF memory reader/writer 6 via a cable, a large-capacity 00 memory 53 which stores data to be handled in various C processes, a CPU 54 which controls the entire work station according to a program prepared beforehand, an input unit C 10 55 which inputs various kinds of operations and an output unit 56 which is connected to the label printer 7 or the database 8 or other printers over the network 9 for performing various kinds of output processes.
When the observation period ends and the receiver 4 is connected to the work station 5 in a communicable manner, taken image data stored in the CF memory 44 is transferred from the receiver 4 to the work station 5 and stored in the memory 53. In the work station 5, taken images from the capsule endoscope 10 according to this embodiment, the display of an average color slider to be discussed later, the locus of the capsule endoscope 10, etc. are displayed at the time of a diagnosis. The diagnosis results are output as a chart from the printer and stored in the database 8 patient by patient.
Next, specific procedures according to this embodiment are explained. Figs. 4, 5, and 6 are show one example of screen transition associated with the observation procedures according to this embodiment, Figs. 7 and 8 show one example of screen transition associated with the diagnosis procedures according to this embodiment, and Fig.
9 is a flowchart of the operation for average color bar display according to this embodiment. A program for displaying an average color slider is directly installed
I
from a recording medium such as CD-ROM or is downloaded o from outside such as a network, then installed and stored in the memory 53 of the work station 5 as its storage scheme.
First, a doctor (or a nurse) formats the CF memory 44 using the work station 5 and the CF memory reader/writer 6.
In this case, as procedures prior to observation, the CF memory 44 is inserted into the CF memory reader/writer 6 and a guidance screen prompting connection of the CF memory reader/writer 6 to the work station 5 is displayed on the display unit 51 of the work station 5 (Fig. When the doctor performs a menu operation for "NEXT", the process proceeds to the next guidance screen display. It is assumed that the doctor has prepared according to the guidance at this time. If the preparation is inadequate and the menu operation for "NEXT" is done in that state, a message of non-insertion of the CF memory, non-connection of the CF memory reader/writer or the like may be displayed.
The next guidance screen displays a guidance screen prompting entry of diagnosis information and patient information (Fig. As the diagnosis information, there are input items of, for example, a hospital name, the name of capsule-administering doctor (nurse), the date/time of capsule administration, a capsule serial number and a receiver serial number. As the patient information, there are input items of, for example, a patient ID, the name of a patient, gender of the patient, the age of the patient and the birth date of the patient. When the input operation for various input items is completed and the menu operation for "NEXT" is done, a confirmation screen for the entered items is displayed (Fig. The screen may go back to the previous screen through a menu operation for
"BACK".
0 As the next guidance screen (Fig. shows a o confirmation of the items entered on the previous screen and the doctor further performs the menu operation for "NEXT", it is considered that nothing is wrong about the input information and the display screen goes to the next screen (Fig. At this time, information on the input 00 items is written in the CF memory 44. When the menu operation for "BACK" is done, the items entered previously Scan be corrected.
The next guidance screen (Fig. shows a message of an instruction to remove the CF memory 44, an instruction to put labels having necessary ID information printed according to the input items confirmation of the items entered on the previous screen to the receiver 4 and the CF memory 44, and an instruction to insert the CF memory 44 into the receiver 4. When the doctor performs a menu operation for "COMPLETED", preparation before administration of the capsule endoscope 10 into the subject is completed.
Then, the administration of the capsule endoscope into the subject 2 is completed, observation of the interior of the body is started and storage of taken image data into the CF memory 44 is started by the operation of the receiver 4. When the observation period ends and storage into the CF memory 44 is finished, the doctor receives guidance from the work station 5 again.
First, the CF memory 44 is removed from the receiver 4 and a guidance screen prompting insertion of the CF memory reader/writer 6 is displayed (Fig. After preparation takes places according to the message, when the doctor performs the menu operation for "NEXT", the display screen goes to the next (Fig. In the next guidance screen (Fig. the diagnosis I I I 17
O
O
information and patient information recorded in the CF o memory 44 are read from the memory and displayed. The o information of the displayed contents, information (taken image data, etc.) acquired through observation is acquired by the work station When the doctor performs the menu operation for "NEXT" upon completion of acquisition of the information in that manner, a process of acquiring data from the CF memory 44 is carried out. When the data acquisition process is finished, a guidance screen prompting completion of data acquisition from the CF memory 44, removal of the CF memory 44 from the CF memory reader/writer 6 and instruction for initiation of diagnosis is displayed (Fig. When the doctor performs the menu operation for "COMPLETED", a sequence of guidance associated with the observation procedures is completed.
In the transition of a series of screens, there are icons of CANCEL and HELP which the doctor can arbitrarily select and operate. When the CANCEL is operated, the inputs so far are initialized.
At the stage of the diagnosis process, first, a list of diagnosis information and patient information of individual patients saved in the memory 53 of the work station 5 is displayed (Fig. Accordingly, the doctor can select on which patient diagnosis is to be done with, for example, a cursor. The selected state has only to be given in inverted display. When a menu operation for "OBSERVATION" is done with the cursor selecting state, a patient to be diagnosed is decided. With regard to diagnosed patients, affixing "DONE" on the displayed list as shown in Fig. 7 can ensure an easy confirmation of whether a diagnosis has been made.
As a patient to be diagnosed is decided in this manner, 18 a diagnosis procedure screen is displayed as shown in Fig.
o 8. This diagnosis procedure screen shows information necessary for diagnosis. 501 and 502 are respectively patient information and diagnosis information of the associated patient, and 503 is an image display field illustrating one of taken images. 504A shows a checkedimage display field giving a list of taken images of interest which have been arbitrarily checked (selected) by a doctor by operating a software-based check button CHK.
505 shows a 3D (three dimensional) position display field showing an imaging position (position inside a body) of the taken image, displayed in the image display field 503, in a 3D manner, 506 shows a playback operation field 506 for performing a playback operation for a taken image to be displayed in the image display field 503, and 507 shows an average color bar colored in time sequence with average colors according to the organs for taken images from the start point of reception by the receiver to the end point of reception. The average color bar 507 serves as a scale indicating the passing time during the observation period. The display screen further displays individual menus for "HELP", "BACK", "CANCEL", and "END DIAGNOSIS/PRINT CHART" The average color bar 507 is average colors acquired from the individual frames of a taken image and colored in time sequence using the characteristics of colors different from one organ to another. In the average color bar 507, therefore, the average color of a taken image when the capsule endoscope 10 is moving according to regions of each organ becomes nearly uniform. Even if an image taken while movement in the same organ contains noise, nearly a uniform color for each organ can be acquired by obtaining the average color of a single screen frame by frame.
O
In the average color bar 507, a slider S is shown o movable in the direction of the time axis. The slider S 0 serves as an index to indicate the position of a taken image to be displayed in the image display field 503, at a position on the average color bar 507. Therefore, moving/display control of the slider S is carried out 00 according to the operation of the playback operation field 506.
The movement of the slider S on the average color bar 1 0 507 and changing of the taken image to be displayed in the image display field 503 are synchronized. That is, a software-based FRAME PLAYBACK button, PLAYBACK button, and FAST PLAYBACK (FP) button for operations in the forward playback direction along the time-sequential direction and a software-based REVERSE FRAME PLAYBACK button, REVERSE PLAYBACK button, and FAST REVERSE PLAYBACK (FR) button for operations in the reverse playback direction along the time-sequential direction are displayed and controlled.
Further, a STOP button is displayed and controlled in the playback operation field 506.
When a doctor clicks the PLAYBACK button with a mouse (not shown) by operating the input unit 55, an image based on taken image data is displayed in the image display field 503 in time sequence in the forward playback direction.
When the FRAME PLAYBACK button is clicked, a next image in the forward playback direction is displayed, and when the FAST PLAYBACK button is clicked, images are reproduced and displayed faster than the playback done by the PLAYBACK button in the forward playback direction. When the STOP button is clicked during playback or during fast playback, changing of the displayed image is stopped while an image at the time the clicking was made is displayed.
When the doctor clicks the REVERSE PLAYBACK button r with the mouse (not shown) by operating the input unit
C.)
o an image based on taken image data is displayed in the M image display field 503 in the reverse playback direction with respect to the time-sequential direction. When the REVERSE FRAME PLAYBACK button is clicked, an image previous by one in the forward playback direction is displayed, and when the FAST REVERSE PLAYBACK button is clicked, images are reproduced and displayed faster than the playback done by the REVERSE PLAYBACK button in the reverse playback direction. When the STOP button is clicked during reverse playback or during fast reverse playback, changing of the displayed image is stopped while an image at the time the clicking was made is displayed.
When a diseased part like a bleeding part is found, or the like at the time of image playback or reverse playback in the image display field 503, a checked image distinguished from other images can be extracted at the doctor's discretion. When such checking is desired, the doctor operates the check button CHK. The checked image is additionally displayed as a thumbnail image in the checkedimage display field 504A. Due to the restriction of the display area, the checked-image display field 504A can display up to a predetermined number of images. In this embodiment, as shown in Fig. 8, for example, up to five images can be displayed and for other checked images, display images are switched by scrolling.
As the average color bar 507 is segmented by the average colors according to the types of the organs, the doctor can intuitively and quickly move the display image to the position of the taken image associated with the desired organ referring to the average color bar 507. At this time, the slider S of the average color bar 507 is moved by using the mouse (not shown). As the slider S is
O
O
operated to move on the average color bar 507, a process of o sequentially changing the image to the one at the position o indicated by the slider S following the movement is executed in the image display field 503.
In this embodiment, when the doctor finds a bleeding O part from the display image, a flag as a bleeding part can g be affixed to each taken image. In this case, though not p shown, a sub menu is displayed with the current state Sdisplayed in the image display field 503 to manually set the flag of the bleeding part. Accordingly, display can be made in association with the positions on the average color bar 507, such as bleeding parts Vl, V2, as shown in Fig. 8, for example.
A bleeding part can be automatically extracted through image processing, in which case an AUTO-RETRIEVE BLEEDING PART button as indicated by 508 is operated. The operation of the AUTO-RETRIEVE BLEEDING PART button 508 may be done for the image currently displayed in the image display field 503 or for all the images. When it is found in automatic retrieval, a flag is put in association with each image as done in the case of manual operation.
The diagnosis by a doctor can be terminated by a menu operation for "END DIAGNOSIS/PRINT CHART". The diagnosis results are made into a chart and printed through a printer (not shown) from the work station 5 or via the database 8.
In the display of the average color bar 507, a process is executed as shown in Fig. 9. That is, when a patient to be diagnosed is decided from a list shown in Fig. 7, a file of imaging information corresponding to that patient is designed. Then, one frame of image files is read from the memory 53 and opened (step Sl), and the average color of the taken images frame by frame is measured (step S2).
When the average color is measured and average color 22 data is acquired, the average color data for the first U frame is stored in the memory 53 (step S3). Then, a Mprocessed image file is closed and an image file located next in time sequence is read out and opened, and a similar process is repeatedly executed thereafter (NO route of step 00 When the average colors for all the imaging information of the patient to be diagnosed are obtained (step S5), the average color bar 507 is displayed and controlled as shown in Fig. 8 using the average color data stored in the memory 53 (step S6). In this manner, the display of the average color bar 6 is completed. At this time, the initial position of the slider S is the left end (start position) of the average color bar 507 but is not restrictive.
Because the amount of the imaging information including taken image data is huge, it is unnecessary to open all the image files and acquire the average colors for all the frames, and the average color may be acquired while efficiently thinning several frames. Although the acquired average color itself is displayed on the average color bar 507 in this embodiment, it is not restrictive and a color corresponding to this average color has only to be displayed on the average color bar 507.
According to this embodiment, as described above, a scale indicating the overall imaging period of input image data taken in time sequence by the capsule endoscope (internal imaging device) is displayed, a movable slider is shown on the scale, an image at the imaging time corresponding to the position of the slider is displayed in response to the movement of the slider on the scale, and a color corresponding to average color information for one screen of input image data is displayed at the time- 23 associated position on the scale, so that distinguishing o coloring is carried out according to the taken part and an Corgan in the body can easily be determined from the distinguished colors. Accordingly, the ability to retrieve the image is improved and it is possible to easily recognize the organ depicted in each image.
Although the position of an organ is identified using the average colors arranged on the average color bar as an index in the embodiment described above, the present invention is not limited to this type and an additional function of displaying the name of an organ in association with the average color may be provided as in a modification to be discussed below. As the modification to be discussed below is the same in the structure and functions described above, only what is added is discussed.
Fig. 10 is a diagram of one example of a display screen associated with a diagnosis process according to one modification of this embodiment, Fig. 11 illustrates the principle of automatic discrimination of organ names according to one modification of this embodiment, and Fig.
12 is a flowchart of the procedures of discriminating the organ names according to one modification of this embodiment.
In Fig. 10, the organ names are displayed in association with each average color on the average color bar 507. Average colors are lined on the average color bar 507 in the order of the esophagus, the stomach, the small intestine, and the large intestine in the order of imaging done in a body by the capsule endoscope 10 in time sequence.
Therefore, the average color bar 507 shows organ names 509 in the order of the esophagus, the stomach, the small intestine, and the large intestine in association with the average colors of the individual organs.
At the time of automatic discrimination of organ names, O it is the automatic discrimination in the ranges of organs.
0 The level of red and the level of blue for individual taken images at elapsed times have the characteristics as shown in Fig. 11. As an actual image contains a noise component, it is subjected to a low-pass filter (LPF) process in the 00 direction of the time axis with respect to the levels of red and blue that have the characteristics to remove noises.
Then, edge portions (discoloration edges) the levels of red 1 0 and blue in the direction of the time axis after the LPF process commonly have are extracted.
In the example in Fig. 11, there are three discoloration edges, and extracted in the above manner. Therefore, automatic discrimination is done such that from the positions of the discoloration edges (1) and in the direction of the time axis, the first discoloration edge is a transitional portion from the esophagus to the stomach, is a transitional portion from the stomach to the small intestine and is a transitional portion from the small intestine to the large intestine. At this time, the order of the organ names is based on the layout of the organs to be taken by the capsule endoscope 10 in the direction of the time axis.
As the processing based on the principle described above, first, the red level and blue level are computed (step S21), the LPF process in the direction of the time axis is performed on the red level and blue level (step S22) and the discoloration edges and are detected (step S23). Then, automatic discrimination of the ranges of the organs is carried out from the timeassociated positions of the discoloration edges (2) and and the organ names are displayed in association with the individual average colors on the average color bar 507 (step S24).
U In the above manner, a scale indicating the overall imaging period of input image data taken in time sequence by the capsule endoscope is displayed, a movable slider is shown on the scale, an image at the imaging time corresponding to the position of the slider is displayed in response to the movement of the slider on the scale, and organs are discriminated based on color information for one screen of input image data and organ names are displayed in association with the scale, so that organs in the body can easily be determined from the displayed organ names. This also improves the ability to retrieve images and makes it possible to easily recognize the organ depicted in each image.
Although the ranges of the organs on the average color bar are automatically discriminated from the discoloration edges in the modification described above, the present invention is not limited to this type and a pH sensor may be provided in the capsule endoscope 10 so that the ranges of the organs are specified more accurately using the measured pH values. In this case, the pH values are measured by the pH sensor during the observation period and like taken images, the pH values are measured in time sequence and are stored in the receiver 4. At that time, the taken images and pH values are recorded in association with each other, such as coexisting in each frame (image file).
Fig. 13 shows an example of application of the modification in Fig. 11. In the automatic discrimination with pH values added, as shown in Fig. 13, using the fact that the stomach is in an acidic state, an acidic part is compared with the discoloration edges and to discriminate the stomach part, thereby further increasing the discrimination precision.
o Chart creation according to the above embodiment is explained next. Fig. 14 shows one example of screen transition associated with the diagnosis procedures according to this embodiment, and Fig. 15 is a flowchart of an operation for displaying the imaging time of a designated image according to this embodiment. While a diagnosis by a doctor can be terminated through the menu operation for "END DIAGNOSIS/PRINT CHART", further transition to the chart creating procedures can be made.
When the process is shifted from the display screen in Fig. 8 to the display screen in Fig. 14, comments of a doctor are entered and a mark indicating to which elapsed time on the average color bar 507 each checked image corresponds is displayed.
That is, in Fig. 14, 504B indicates a checked-image display field, set larger than the checked-image display field 504A and provided at the lower portion of the screen.
As a difference from the checked-image display field 504A, numbers to (10) are given to individual taken images and displayed. The checked-image display field 504B has the same function as the checked-image display field 504A.
510 is a comment input field where opinions (comments) of a doctor are input and displayed. The results of a diagnosis by a doctor are input as comments in the comment input field 510. 511 indicates an imaging time display mark that is displayed, as a mark on the average color bar 507, indicating which taken image at which elapsed time each checked image to be displayed in the checked-image display field 504B is. As the imaging time display mark, a downward arrow as an index indicating the imaging time for a checked image and the aforementioned number given to a checked image as relative display indicating the correlation with the checked image to show the correlation O with the checked image are displayed on the average color bar 507.
Fig. 14 exemplifies ten checked images. In this example, average colors are distinguished on the average color bar 507 in the order of the esophagus, the stomach, the small intestine, and the large intestine. As apparent C from the ranges of the organs of the organ names 509, therefore, a mark for a checked image is present in the 1C 0 range of the esophagus, and marks and for a checked image are present in the range of the stomach.
Further, marks and (10) for checked images are present in the range of the small intestine.
Therefore, the presence of images checked by a doctor are identified in the esophagus, the stomach, and the small intestine from the example in Fig. 14, and marks are displayed in association with the times at which the individual checked images have been taken, so that the doctor can easily confirm at which parts of the organs the checked images have been taken. Although the imaging time display mark is displayed on the average color bar 505 showing the organ names in Fig. 14, it may be displayed on the average color bar that does not show the organ names as in Fig. 8. Although a correlation indication (number) indicating the correlation with a checked image is displayed as the imaging time display mark in Fig. 14, it may be an index (downward arrow) indicating the position of the imaging time.
The process for the above mark display is described with reference to Fig. 15. In the imaging time display of a checked image or a designated image, first, the date/time of creating a file of the designated image is acquired from the memory 53 (step S31), and the time elapsed since the o date/time of the initiation of imaging is computed (step S32). Then, a mark display as shown in Fig. 4 is controlled on the scale of the average color bar 507 at the position corresponding to the elapsed time on the average color bar 507 (step S33). Thereafter, when chart printing 00 is manipulated, outputting for the chart printing is executed.
According to this embodiment, as described above, a scale indicating the overall imaging period of input image data taken in time sequence by the capsule endoscope (internal imaging device) is displayed, a color corresponding to average color information for one screen of input image data is displayed at a time-associated position on the scale, an image corresponding to the input image data is displayed, and an index indicating a position corresponding to an imaging time of a designated image is displayed, so that it is possible to visually and easily recognize how many and in which time band designated images are present. As organs can easily be determined from the colors distinguished from one taken part from another one, it is possible to easily recognize which part of which organ has more designated images.
Further, a scale indicating the overall imaging period of input image data taken in time sequence by the capsule endoscope is displayed, organs are discriminated based on color information of one screen of input image data, the names of the discriminated organ are displayed in association with the scale, images corresponding to the input image data are displayed and an index indicating the position corresponding to the imaging time of the designated image is displayed on the scale, so that organs in the body can easily be determined from the displayed organ names. This also makes it possible to easily o recognize which part of which organ has more designated images.
As explained above, this invention is constructed in such a way that a scale indicating the overall imaging period of input image data taken in time sequence by an internal imaging device is displayed, a color corresponding to average color information for one screen of input image data is displayed at a time-associated position on the scale, an image corresponding to the input image data is displayed, and an index indicating a position corresponding to an imaging time of a designated image is displayed, so that it is possible to visually and easily recognize how many and in which time band designated images are present and easily determine organs from the colors distinguished from one taken part from another one, thus making it possible to easily recognize which part of which organ has more designated images.
The invention also is constructed in such a way that a scale indicating the overall imaging period of input image data taken in time sequence by an internal imaging device is displayed, organs are discriminated based on color information of one screen of input image data, names of the discriminated organ are displayed in association with the scale, images corresponding to the input image data are displayed and an index indicating the position corresponding to the imaging time of the designated image is displayed on the scale, so that organs in the body can easily be determined from the displayed organ names, whereby it is possible to easily recognize which part of which organ has more designated images.
INDUSTRIAL APPLICABILITY P OPERXSEW\20D7XlctobcA30 3 274 pqeldac.II fl 2 7 c o O As explained above, the image display apparatus, the image display method, and the image display program according to the present invention are suitable for a system to retrieve a desired image from a vast amount of in-vivo 00 5 images taken in time sequence by using a capsule-type C endoscope.
Throughout this specification and the claims which follow, unless the context requires otherwise, the word "comprise", and variations such as "comprises" and "comprising", will be understood to imply the inclusion of a stated integer or step or group of integers or steps but not the exclusion of any other integer or step or group of integers or steps.
The reference in this specification to any prior publication (or information derived from it), or to any matter which is known, is not, and should not be taken as an acknowledgment or admission or any form of suggestion that that prior publication (or information derived from it) or known matter forms part of the common general knowledge in the field of endeavour to which this specification relates.
Claims (23)
- 2. The method according to claim 1, wherein the graphical representation comprises a summary of a color representation of at least some of the images.
- 3. The method according to claim 2, wherein the graphical representation has a bar shape adapted to indicate a magnitude of the image data.
- 4. The method according to claim 3, wherein the magnitude includes an imaging period of the image data obtained by an in-vivo imaging device.
- 5. The method according to claim 1, wherein the predetermined characteristic comprises an indication of the presence of blood in the image.
- 6. The method according to claim 5, wherein the graphical representation comprises a summary of a color representation of at least some of the images. P\OPER\SEW\2oo7\OobU\)3327416 d clims pgen dm3II/21 32 c 0 O 7. The method according to claim 5, wherein the graphical Srepresentation has a bar shape adapted to indicate a magnitude of the images. 00 5 8. The method according to claim 7, wherein the magnitude C includes an imaging period of the image data obtained by an r in-vivo imaging device.
- 9. The method according to claim 1, wherein the images are obtained by a swallowable in-vivo imaging device which moves inside digestive tracts by motility of the digestive tracts. The method according to claim 9, wherein the graphical representation has a bar shape adapted to indicate a magnitude of the images, and wherein the displaying of the graphical representation comprises displaying the bar shape in a substantially horizontal direction with respect to a display area of a display.
- 11. The method according to claim 10, further comprising displaying at least some of the series of images in a first area of the display; wherein the displaying of the graphical representation comprises displaying the bar shape in a second area located below the first area.
- 12. An image display apparatus comprising: an input unit adapted to input image data obtained by an in-vivo imaging device; and a signal processor adapted to execute: displaying a graphical representation with respect to at least some of images of the image data; P kOPERZSEW\2(X)7Oetobcr327436 div clams pages doc.311012XY7 33 c o O detecting from the image data an image having a Spredetermined characteristic; and displaying an indicator in a position along the graphical representation, wherein the indicator is adapted 00 5 to indicate the image having the predetermined C characteristic.
- 13. The image display apparatus according to claim 12, wherein the graphical representation comprises a summary of a color representation of at least some of the images.
- 14. The image display apparatus according to claim 13, wherein the graphical representation has a bar shape adapted to indicate a magnitude of the image data. The image display apparatus according to claim 14, wherein the magnitude includes an imaging period of the image data obtained by the in-vivo imaging device.
- 16. The image display apparatus according to claim 12, wherein the predetermined characteristic comprises an indication of the presence of blood in the image.
- 17. The image display apparatus according to claim 16, wherein the graphical representation comprises a summary of a color representation of at least some of the images.
- 18. The image display apparatus according to claim 16, wherein the graphical representation has a bar shape adapted to indicate a magnitude of the images.
- 19. The image display apparatus according to claim 18, P \PEMEW 0 ooObc 27436 dv calms pags dm.3/IO200 34 O wherein the magnitude includes an imaging period of the Simage data obtained by the in-vivo imaging device. The image display apparatus according to claim 12, 00 5 wherein the images are obtained by a swallowable in-vivo N' imaging device which moves inside digestive tracts by r motility of the digestive tracts.
- 21. The image display apparatus according to claim wherein the graphical representation has a bar shape adapted to indicate a magnitude of the images, and wherein the signal processor displays the bar shape in a substantially horizontal direction with respect to a display area of a display.
- 22. The image display apparatus according to claim 21, wherein the signal processor displays at least some of the series of images in a first area of the display, and displays the bar shape in a second area located below the first area.
- 23. An image display apparatus comprising: an obtaining unit that obtains a plurality of images taken by an in-vivo imaging device; a bar display unit that displays a bar illustrating a stream of images configured by the plurality of the images; a specific image detecting unit that detects an image including a specific image area from the images included in the stream, a specific event being indicated on the specific image area; and an indicator display unit that displays an indicator along the bar and in association with the detected image so P nOPER,.EW\2007k liobaA3321436di clims page. doc.311012(E7 c 0 O as to indicate a position of the detected image in the M stream, when the specific image detecting unit detects the image including the specific image area. 00 5 24. An image display apparatus according to claim 23, Cl wherein the specific event is a bleeding. An image display apparatus according to claim 23 or 24, further comprising: a detecting unit that detects summaries of contents of at least some of the images contained in the stream, wherein the summaries detected by the detecting unit is graphically displayed on the bar in association with positions in the stream, the positions containing the summary-detected images.
- 26. An image display apparatus according to claim wherein the detecting unit detects color information of the images to be detected, and detects the summaries of the images based on the color information.
- 27. An image display apparatus according to any one of claims 23 to 26, wherein the bar and the indicator are displayed on a report generation display.
- 28. An image display method comprising: obtaining a plurality of images taken by an in-vivo imaging device; displaying a bar illustrating a stream of images configured by the plurality of the images; detecting an image including a specific image area from the images included in the stream, a specific event being P \OPER\SEW\2007\Oc obCr\30 6 dicans Iaspig"do-3/IO2OO7 36 c o O indicated on the specific image area; and Sdisplaying an indicator along the bar in association with the detected image so as to indicate a position of the detected image in the stream, when the specific image 00 5 detecting unit detects the image including the specific ri image area.
- 29. The image display method according to claim 28, wherein the specific event is a bleeding. The image display method according to claim 28 or 29, further comprising: detecting summaries of contents of at least some of the images contained in the stream, wherein the summaries detected by the detecting of the summaries is graphically displayed on the bar in association with positions in the stream, the positions containing the summary-detected images.
- 31. The image display method according to claim 30, wherein the detecting of the summaries detects color information of the images to be detected, and detecting the summaries of the images based on the color information.
- 32. The image display method according to any one of claims 28 to 31, wherein the bar and the indicator are displayed on a report generation display.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2007221810A AU2007221810C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2003-122805 | 2003-04-25 | ||
AU2004233673A AU2004233673C1 (en) | 2003-04-25 | 2004-04-21 | Image display unit, image display method and image display program |
AU2007221810A AU2007221810C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2004233673A Division AU2004233673C1 (en) | 2003-04-25 | 2004-04-21 | Image display unit, image display method and image display program |
Publications (3)
Publication Number | Publication Date |
---|---|
AU2007221810A1 AU2007221810A1 (en) | 2007-10-25 |
AU2007221810B2 true AU2007221810B2 (en) | 2008-12-18 |
AU2007221810C1 AU2007221810C1 (en) | 2009-08-13 |
Family
ID=38659193
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2007221809A Ceased AU2007221809C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
AU2007221808A Ceased AU2007221808C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
AU2007221810A Ceased AU2007221810C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2007221809A Ceased AU2007221809C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
AU2007221808A Ceased AU2007221808C1 (en) | 2003-04-25 | 2007-10-03 | Image display unit, image display method and image display program |
Country Status (1)
Country | Link |
---|---|
AU (3) | AU2007221809C1 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11225996A (en) * | 1998-02-19 | 1999-08-24 | Olympus Optical Co Ltd | Capsule type in vivo information detector |
JP2002290783A (en) * | 2001-03-28 | 2002-10-04 | Fuji Photo Optical Co Ltd | Electronic endoscope device |
-
2007
- 2007-10-03 AU AU2007221809A patent/AU2007221809C1/en not_active Ceased
- 2007-10-03 AU AU2007221808A patent/AU2007221808C1/en not_active Ceased
- 2007-10-03 AU AU2007221810A patent/AU2007221810C1/en not_active Ceased
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11225996A (en) * | 1998-02-19 | 1999-08-24 | Olympus Optical Co Ltd | Capsule type in vivo information detector |
JP2002290783A (en) * | 2001-03-28 | 2002-10-04 | Fuji Photo Optical Co Ltd | Electronic endoscope device |
Also Published As
Publication number | Publication date |
---|---|
AU2007221810A1 (en) | 2007-10-25 |
AU2007221810C1 (en) | 2009-08-13 |
AU2007221808B2 (en) | 2008-12-18 |
AU2007221808A1 (en) | 2007-10-25 |
AU2007221809B2 (en) | 2008-12-18 |
AU2007221809C1 (en) | 2009-08-13 |
AU2007221809A1 (en) | 2007-10-25 |
AU2007221808C1 (en) | 2009-08-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2004233673B9 (en) | Image display unit, image display method and image display program | |
AU2004233674B2 (en) | Image display apparatus, image display method and image display program | |
US20040225223A1 (en) | Image display apparatus, image display method, and computer program | |
JP4554647B2 (en) | Image display device, image display method, and image display program | |
JP4547401B2 (en) | Image display device, image display method, and image display program | |
JP4547402B2 (en) | Image display device, image display method, and image display program | |
AU2008200088B2 (en) | Image Display Apparatus, Image Display Method and Image Display Program | |
AU2007221810B2 (en) | Image display unit, image display method and image display program | |
CA2614635C (en) | Image display apparatus, image display method, and image display program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
DA2 | Applications for amendment section 104 |
Free format text: THE NATURE OF THE AMENDMENT IS AS SHOWN IN THE STATEMENT(S) FILED 31 MAR 2009. |
|
DA3 | Amendments made section 104 |
Free format text: THE NATURE OF THE AMENDMENT IS AS SHOWN IN THE STATEMENT(S) FILED 31 MAR 2009 |
|
FGA | Letters patent sealed or granted (standard patent) | ||
MK14 | Patent ceased section 143(a) (annual fees not paid) or expired |