WO2012063623A1 - 画像表示装置及びカプセル型内視鏡システム - Google Patents
画像表示装置及びカプセル型内視鏡システム Download PDFInfo
- Publication number
- WO2012063623A1 WO2012063623A1 PCT/JP2011/074345 JP2011074345W WO2012063623A1 WO 2012063623 A1 WO2012063623 A1 WO 2012063623A1 JP 2011074345 W JP2011074345 W JP 2011074345W WO 2012063623 A1 WO2012063623 A1 WO 2012063623A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- unit
- vivo
- display
- image data
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/041—Capsule endoscopes for imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
- A61B1/000094—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope extracting biological structures
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00011—Operational features of endoscopes characterised by signal transmission
- A61B1/00016—Operational features of endoscopes characterised by signal transmission using wireless means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/0002—Operational features of endoscopes provided with data storages
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00043—Operational features of endoscopes provided with output arrangements
- A61B1/00045—Display arrangement
Definitions
- the present invention relates to an image display apparatus and a capsule endoscope system for displaying an in-vivo image acquired by a capsule endoscope introduced into a subject.
- the number of in-vivo images taken by the capsule endoscope is about 60,000 (about 8 hours), for example. For this reason, a certain amount of time is required until the in-vivo image can be displayed after the image data is taken into the image display device. In addition, when position estimation processing is executed for those in-vivo images, a huge amount of data processing must be executed. As a result, there is a problem that it takes time until image interpretation is enabled after image data is acquired, and the start of interpretation work is delayed.
- the present invention has been made in view of the above, and provides an image display device and a capsule endoscope system that can display an interpretation screen at an early stage after obtaining image data captured by a capsule endoscope.
- the purpose is to provide.
- an image display device receives a wireless communication with a capsule endoscope from a capsule endoscope that captures an in-vivo image of a subject.
- An image display device that displays an in-vivo image group based on in-vivo image data acquired via the device, the in-vivo image data and the capsule endoscope in the subject associated with the in-vivo image data
- the position information of the capsule endoscope at the time of capturing each in-vivo image is acquired by executing a position estimation process based on the storage unit that stores information related to the position of the image and the information related to the position
- a position information acquisition unit that performs determination, a determination unit that determines whether each in-vivo image satisfies a predetermined condition, and an in-vivo image that is determined to satisfy the predetermined condition by the determination unit Characterized in that it comprises a priority processing controller for controlling the position information acquisition unit to perform said position estimation processing.
- the image display device further includes a detection image processing unit that performs a detection image process for detecting a predetermined part of the in-vivo image data stored in the storage unit, and the predetermined condition is the detection image In the processing, the predetermined part is detected.
- the predetermined condition includes marking information added according to a signal input from the outside to the receiving device while the capsule endoscope passes through the subject.
- the image display device further includes a detection image processing unit that performs a detection image process for detecting a predetermined part on the in-vivo image data stored in the storage unit, and the predetermined condition is within the capsule type While the endoscope passes through the subject, it has marking information added according to a signal input to the receiving device from the outside, and the predetermined image is detected by the detection image processing unit. It is characterized by that.
- the detection image processing unit preferentially performs the detection image processing on the in-vivo image having the marking information.
- the image display device further includes a detection image processing unit that performs a detection image process for detecting a predetermined part on the in-vivo image data stored in the storage unit, and the predetermined condition is within the capsule type While the endoscope passes through the subject, it has marking information added according to a signal input from the outside to the receiving device, and the predetermined part is detected in the detection image processing unit It is characterized by satisfying at least one of the above.
- the image display device includes: a display image processing unit that performs display image processing on in-vivo image data stored in the storage unit; and the in-vivo image data that has been subjected to the display image processing.
- a display control unit to be displayed on the unit.
- the image display device further includes an input signal receiving unit that receives a signal input to the image display device from the outside, and the priority processing control unit receives a selection signal for selecting a predetermined in-vivo image by the input signal receiving unit. When received, the position estimation process is preferentially executed for the predetermined in-vivo image.
- a capsule endoscope system is introduced into a body of a subject, performs imaging, and generates in-vivo image data representing the in-vivo image of the subject, and the capsule endoscope A receiving device that receives in-vivo image data generated by a mirror by wireless communication and the image display device are provided.
- the position estimation process is preferentially executed for in-vivo images that satisfy a predetermined condition, position information of in-vivo images necessary for diagnosis can be acquired earlier. Therefore, it is possible to display the image interpretation screen at an early stage based on the in-vivo image data subjected to the display image processing and the position information acquired preferentially.
- FIG. 1 is a schematic diagram showing a schematic configuration of a capsule endoscope system according to Embodiment 1 of the present invention.
- FIG. 2 is a schematic diagram showing a schematic configuration of the capsule endoscope shown in FIG.
- FIG. 3 is a block diagram illustrating a configuration of the capsule endoscope and the receiving device illustrated in FIG. 1.
- FIG. 4 is a block diagram showing a configuration of the image display apparatus according to Embodiment 1 of the present invention.
- FIG. 5 is a flowchart showing the operation of the image display apparatus shown in FIG.
- FIG. 6 is a schematic diagram illustrating a first display example of an interpretation screen.
- FIG. 7 is a schematic diagram illustrating a second display example of the interpretation screen.
- FIG. 1 is a schematic diagram showing a schematic configuration of a capsule endoscope system according to Embodiment 1 of the present invention.
- FIG. 2 is a schematic diagram showing a schematic configuration of the capsule endoscope shown in FIG.
- FIG. 3 is a
- FIG. 8 is a block diagram showing a configuration of an image display apparatus according to Embodiment 2 of the present invention.
- FIG. 9 is a flowchart showing the operation of the image display apparatus shown in FIG.
- FIG. 10 is a block diagram showing a configuration of an image display apparatus according to Embodiment 3 of the present invention.
- FIG. 11 is a flowchart showing the operation of the image display apparatus shown in FIG.
- FIG. 12 is a block diagram showing a configuration of an image display apparatus according to Embodiment 4 of the present invention.
- FIG. 13 is a flowchart showing the operation of the image display apparatus shown in FIG.
- FIG. 14 is a block diagram showing a configuration of an image display apparatus according to Embodiment 5 of the present invention.
- FIG. 15 is a flowchart showing the operation of the image display apparatus shown in FIG.
- FIG. 16 is a block diagram illustrating a display example of an interpretation screen displayed on the image display apparatus illustrated in FIG.
- FIG. 17 is a block diagram illustrating a display example of an interpretation screen displayed in the image display apparatus illustrated in FIG.
- FIG. 18 is a block diagram showing a configuration of an image display apparatus according to Embodiment 6 of the present invention.
- FIG. 19 is a schematic diagram illustrating a display example of an interpretation screen displayed on the image display apparatus illustrated in FIG.
- a capsule endoscope system according to an embodiment of the present invention will be described with reference to the drawings.
- a system including a capsule endoscope that captures an in-vivo image introduced into the body of a subject is illustrated, but the present invention is not limited to this embodiment. .
- FIG. 1 is a schematic diagram showing a schematic configuration of a capsule endoscope system according to Embodiment 1 of the present invention.
- the capsule endoscope system 1 is introduced into the body of a subject 10 to perform imaging, and a capsule endoscope 2 that wirelessly transmits image data of the in-vivo image to the receiving device 3, and the capsule endoscope 2 Receiving device 3 that receives the in-vivo image data wirelessly transmitted from, and image display device 5 that displays the in-vivo image based on the in-vivo image data received by receiving device 3.
- the capsule endoscope 2 is swallowed from the mouth of the subject 10 and then moves inside the organ of the subject 10 by a peristaltic motion of the organ 10 while moving through the body of the subject 10 at a predetermined time interval (for example, 0.
- In-vivo image data is generated by performing predetermined signal processing on the imaging signals obtained by sequentially imaging at intervals of 5 seconds.
- the capsule endoscope 2 sequentially wirelessly transmits the generated in-vivo image data to the external receiving device 3 every time an in-vivo image of the subject 10 is captured.
- Identification information (for example, a serial number) for identifying an individual capsule endoscope is assigned to the capsule endoscope 2, and this identification information is also wirelessly transmitted together with the in-vivo image data.
- the receiving device 3 includes an antenna unit 4 having a plurality of receiving antennas 41a to 41h.
- Each of the receiving antennas 41a to 41h is realized by using, for example, a loop antenna, and corresponds to a predetermined position on the external surface of the subject 10 (for example, each organ in the subject 10 that is a passage path of the capsule endoscope 2). Arranged).
- the arrangement of the receiving antennas 41a to 41h may be arbitrarily changed according to the purpose of inspection or diagnosis. Further, the number of antennas provided in the antenna unit 4 is not limited to eight shown as the receiving antennas 41a to 41h, and may be smaller or larger than eight.
- the receiving device 3 is carried by the subject 10 while the capsule endoscope 2 is imaging (for example, until it is introduced from the mouth of the subject 10 and passes through the digestive tract to be discharged).
- the in-vivo image data wirelessly transmitted from the capsule endoscope 2 is received via the antenna unit 4 and stored in a built-in memory.
- the receiving device 3 is removed from the subject 10 and connected to the image display device 5 for transferring (downloading) information such as in-vivo image data.
- the image display device 5 is realized by a workstation or a personal computer having a display unit such as a CRT display or a liquid crystal display, and performs predetermined processing on in-vivo image data and information related to the position acquired via the receiving device 3.
- the in-vivo image is displayed on the display unit.
- the image display device 5 is connected to an operation input device 5b such as a keyboard and a mouse.
- an operation input device 5b such as a keyboard and a mouse.
- a touch panel provided to be superimposed on the display unit may be used.
- the user interprets the in-vivo images of the subject 10 sequentially displayed on the image display device 5 while operating these operation input devices 5b, thereby causing a living body part (for example, esophagus, The stomach 10, small intestine, large intestine, etc.) are observed (examined), and the subject 10 is diagnosed based on the observation results.
- a living body part for example, esophagus, The stomach 10, small intestine, large intestine, etc.
- the image display device 5 includes an interface such as a USB (Universal Serial Bus) port.
- the cradle 5a is connected to the image display device 5 through this USB port.
- the cradle 5 a is a reading device that reads in-vivo image data from the memory of the receiving device 3.
- the receiving device 3 is electrically connected to the image display device 5, and in-vivo image data stored in the memory of the receiving device 3 and related information (reception intensity information and time information). And the identification information of the capsule endoscope 2) are transferred to the image display device 5.
- the image display device 5 acquires a series of in-vivo image data and related information regarding the subject 10 in this way, and further displays the in-vivo image on the screen by executing processing to be described later.
- an output device such as a printer may be connected to the image display device 5 and a hard copy of the in-vivo image may be output to the output device.
- FIG. 2 is a schematic diagram illustrating a configuration example of the capsule endoscope 2.
- FIG. 3 is a block diagram showing the configuration of the capsule endoscope 2 and the receiving device 3.
- the capsule endoscope 2 includes a capsule container 20 including a housing portion 20a and a hemispherical optical dome 20b.
- the capsule container 20 is, for example, large enough to be swallowed by the subject 10.
- the accommodating part 20a has a substantially cylindrical shape in which one end is a hemispherical dome shape and the other end is open.
- the optical dome 20b is formed of a transparent material.
- a circuit board 23 provided with an imaging unit 21, an illumination unit 22, and a drive circuit for driving the imaging unit 21 and the illumination unit 22, respectively, inside the capsule container 20.
- a signal processing unit 24, a memory 25, a transmission unit 26, an antenna 27, and a battery 28 are accommodated.
- the imaging unit 21 includes, for example, an imaging element 21a such as a CCD or CMOS that generates image data of an in-subject image from an optical image formed on the light receiving surface, and an objective disposed on the light receiving surface side of the imaging element 21a. And an optical system 21b such as a lens.
- the illumination unit 22 is realized by an LED (Light Emitting Diode) or the like.
- the image pickup device 21a, the optical system 21b, and the illumination unit 22 are mounted on the circuit board 23.
- the drive circuit of the imaging unit 21 operates under the control of a signal processing unit 24 to be described later, and generates an imaging signal representing an image in the subject 10 periodically (for example, two frames per second) to perform signal processing. Input to the unit 24.
- the imaging unit 21 and the illumination unit 22 will be described as including respective drive circuits.
- the circuit board 23 on which the imaging unit 21 and the illumination unit 22 are mounted is attached to the opening of the housing unit 20a so that the light receiving surface of the imaging element 21a and the light emission direction of the illumination unit 22 face the optical dome 20b side. Therefore, the imaging direction of the imaging unit 21 and the illumination direction of the illumination unit 22 face the outside of the capsule endoscope 2 through the optical dome 20b as shown in FIG. Thereby, the inside of the subject 10 can be imaged by the imaging unit 21 while the inside of the subject 10 is illuminated by the illumination unit 22.
- the signal processing unit 24 controls the operation of each unit in the capsule endoscope 2 and A / D converts the imaging signal output from the imaging unit 21 to generate digital in-vivo image data. Apply signal processing.
- the memory 25 temporarily stores various operations executed by the signal processing unit 24 and in-vivo image data subjected to signal processing in the signal processing unit 24.
- the transmission unit 26 superimposes the in-vivo image data stored in the memory 25 on the radio signal together with the identification information of the capsule endoscope 2 and transmits it to the outside via the antenna 27.
- the battery 28 supplies power to each part in the capsule endoscope 2.
- the battery 28 includes a power supply circuit that boosts the power supplied from a primary battery such as a button battery or a secondary battery.
- the receiving device 3 includes a receiving unit 31, a signal processing unit 32, a memory 33, an interface (I / F) unit 34, an operation unit 35, a display unit 36, and a battery 37.
- the receiving unit 31 receives in-vivo image data wirelessly transmitted from the capsule endoscope 2 via the receiving antennas 41a to 41h.
- the signal processing unit 32 controls the operation of each unit in the receiving device 3 and performs predetermined signal processing on the in-vivo image data received by the receiving unit 31.
- the memory 33 stores various operations executed by the signal processing unit 32, in-vivo image data subjected to signal processing in the signal processing unit 32, and related information (reception intensity information, time information, etc.).
- the interface unit 34 transmits the in-vivo image data and related information stored in the memory 33 to the image display device 5 via the cradle 5a.
- the operation unit 35 allows the user to input various operation instructions and settings to the receiving device 3.
- the display unit 36 notifies or displays various information to the user.
- the battery 37 supplies power to each unit in the receiving device 3.
- the in-vivo image data received by the receiving device 3 can be transferred to the image display device 5 by various methods other than the above description.
- a memory removable from the receiving device 3 such as a USB memory or a compact flash (registered trademark) may be used instead of the built-in memory.
- the image display device 5 may be provided with a communication function with an external device, and the in-vivo image data may be transmitted from the receiving device 3 to the image display device 5 by wired or wireless communication.
- FIG. 4 is a block diagram showing a configuration of the image display device 5.
- the image display device 5 includes an interface (I / F) unit 51, a temporary storage unit 52, a display image processing unit 53, an examination information creation unit 54, a storage unit 55, and a detection.
- the image processing unit 56 includes a position information acquisition unit 57, a display control unit 58, a display unit 59, a priority image determination unit 61, and a priority processing control unit 62.
- the interface unit 51 receives in-vivo image data input through the cradle 5a and related information, and signals representing various commands and information input through the operation input device 5b.
- the temporary storage unit 52 is realized by a volatile memory such as DRAM or SRAM, and temporarily stores in-vivo image data input from the receiving device 3 via the interface unit 51 and related information thereof.
- a recording medium such as HDD, MO, CD-R, and DVD-R and a driving device for driving the recording medium are provided, and the in-vivo image data input from the interface unit 51 is recorded. You may make it store once in a medium.
- the display image processing unit 53 performs white balance processing, demosaicing, color conversion, density conversion (gamma conversion, etc.), smoothing (noise removal, etc.), sharpness on the in-vivo image data stored in the temporary storage unit 52.
- the inspection information creation unit 54 creates information related to the inspection based on the information input via the operation input device 5b. Specifically, patient information (ID, name, gender, age, date of birth, etc.) for identifying the subject 10 as a patient, and examination information (hospital name) for identifying the examination content for the subject 10 , Capsule administration doctor (nurse) name, capsule administration date / time, data acquisition date / time, serial number of the capsule endoscope 2, serial number of the receiving device 3, and the like.
- the examination information may be created in advance before the in-vivo image data is transferred from the receiving device 3, or may be created after the in-vivo image data is transferred.
- the storage unit 55 includes various processing programs executed in the image display device 5, in-vivo image data subjected to image processing by the display image processing unit 53, examination information created by the examination information creation unit 54, which will be described later
- the detected image information generated by the detection image processing unit 56 and the position information acquired by the position information acquiring unit 57 are stored.
- the storage unit 55 is realized by, for example, a semiconductor memory such as a flash memory, a RAM, and a ROM, a recording medium such as an HDD, an MO, a CD-R, and a DVD-R, and a drive device that drives the recording medium.
- the detection image processing unit 56 executes detection image processing for detecting a boundary of a lesioned part or an organ from an in-vivo image subjected to display image processing.
- the image processing for detection includes image recognition processing for detecting a lesion, such as tumor, mucosal, and vascularity, or a feature image region of a lesion related to bleeding, organ identification, bleeding This includes an average color calculation process for detecting the location.
- the position information acquisition unit 57 performs position estimation processing based on the reception intensity information and time information stored in the temporary storage unit 52, thereby obtaining the position coordinates of the capsule endoscope 2 at the time of capturing each in-vivo image.
- Information to be represented is acquired.
- the position information acquisition unit 57 acquires the reception intensity of each of the reception antennas 41a to 41h associated with the received in-vivo image data from the temporary storage unit 52, and focuses on each of the reception antennas 41a to 41h. Then, a spherical area having a radius corresponding to the reception intensity is extracted. Note that this radius increases as the reception strength decreases.
- the position where these regions intersect is estimated as the position of the capsule endoscope 2 at that time, that is, the position in the subject 10 that is projected in the in-vivo image.
- the position information acquisition unit 57 stores the acquired position information in the storage unit 55 in association with the corresponding in-vivo image data.
- the display control unit 58 controls the display unit 59 to display the in-vivo image, its position information, and various other information in a predetermined format.
- the display unit 59 is realized by a CRT display or a liquid crystal display, and displays an interpretation screen on which in-vivo images of the subject 10 are sequentially displayed under control of the display control unit 58, various information, a notification message to the user, and the like. To do.
- the priority image determination unit 61 determines an in-vivo image (hereinafter referred to as “priority image”) on which position estimation processing is preferentially executed based on the result of the detection image processing.
- a priority image is an in-vivo image in which a lesion such as neoplastic, mucosal, or vascularity or a lesion related to bleeding is detected, or an in-vivo image in which an organ boundary is captured.
- the priority image determination unit 61 determines whether or not a lesion or organ boundary is extracted from the detection-processed in-vivo image, and prioritizes the in-vivo image data from which the lesion or organ boundary is extracted. A flag indicating that the image is displayed is attached.
- the priority processing control unit 62 is configured to preferentially execute the position estimation process for the in-vivo image data determined as the priority image by the priority image determination unit 61 (that is, the in-vivo image data to which the flag is attached). 57 is controlled.
- the image display device 5 when the receiving device 3 is mounted on the cradle 5 a, the in-vivo image data and related information stored in the memory of the receiving device 3 are transferred to the image display device 5.
- the transferred in-vivo image data and the like are stored in the temporary storage unit 52.
- the image display device 5 starts data processing on the data stored in the temporary storage unit 52.
- FIG. 5 is a flowchart showing the operation of the image display device 5.
- the display image processing unit 53 performs display image processing on the in-vivo image data stored in the temporary storage unit 52.
- the in-vivo image data subjected to the display image processing is stored in the storage unit 55.
- step S12 the detection image processing unit 56 sequentially captures the in-vivo image data stored in the storage unit 55 (for example, the transfer order of the image data) and performs detection image processing. Thereby, when a lesioned part, an organ boundary, or the like is detected from the in-vivo image, detected image information is generated. This detected image information is stored in the storage unit 55 in association with the in-vivo image data.
- the priority image determination unit 61 prioritizes the in-vivo image data (that is, the in-vivo image data having the detected image information) in which the lesion part, the organ boundary, and the like are detected by the detection image processing unit 56. A flag indicating that the image is displayed is attached.
- step S ⁇ b> 14 the position information acquisition unit 57 sequentially takes in the relevant information of the in-vivo image data with the flag indicating that it is a priority image from the storage unit 55 under the control of the priority processing control unit 62, and performs position estimation processing. Execute. The position information generated thereby is stored in the storage unit 55 in association with the in-vivo image data together with the reception time.
- the image display device 5 can display the in-vivo image and the minimum necessary position information on the display unit 59 when the position estimation process for the priority image is completed. Therefore, in step S15, the display control unit 58 causes the display unit 59 to display a message notifying that the interpretation screen can be displayed. At this time, the display control unit 58 may display an interpretation screen on the display unit 59 when a predetermined time has elapsed after the display of the message, or whether or not to actually display the interpretation screen together with the message. You may display the screen which makes a user select. In the latter case, the display control unit 58 causes the display unit 59 to display the interpretation screen when a signal for selecting the interpretation screen display is input by a user operation.
- step S16 the position information acquisition unit 57 sequentially fetches the related information of the in-vivo image data other than the priority image from the storage unit 55 (for example, the transfer order of the image data), and executes the position estimation process.
- the position estimation processing for all in-vivo images is completed, the data processing in the image display device 5 is completed.
- FIG. 6 is a schematic diagram showing a display example of an interpretation screen displayed by the display unit 59 after step S15. After the position estimation processing for the priority image is completed, the display control unit 58 creates an interpretation screen 110 as shown in FIG.
- the image interpretation screen 110 reproduces a series of in-vivo images, a patient information area 111 for displaying identification information of a subject 10 as a patient, a diagnosis information area 112 for displaying identification information of a diagnosis performed on the subject 10, and so on.
- It includes a thumbnail area 116, a time bar 117 indicating the time when the in-vivo image currently displayed in the main display area 113 is acquired, and a position display area 118 in which the imaging position of each in-vivo image is displayed.
- a reduced image 115 and a point on the time bar 117 representing the time when the reduced image 115 is acquired are displayed in a connected manner.
- a humanoid image 120 simulating the subject 10 is displayed.
- the imaging position of the in-vivo image is represented by a point 121 based on the position information acquired by the position information acquisition unit 57.
- the display control unit 58 displays the in-vivo image corresponding to the selected point 121 as a main image.
- the display unit 59 may be controlled to display in the display area 113.
- the imaging position of the priority image (that is, the in-vivo image in which the lesion site, the boundary of the organ, etc. are reflected) is displayed in the position display area 118. Is displayed. Thereafter, as the position estimation process for the in-vivo images other than the priority image is executed, the points 121 displayed in the position display area 118 increase. At this time, for example, the points 121 corresponding to the priority images and the points 121 corresponding to the in-vivo images other than the priority images may be displayed with different colors, marks, or luminances, for example. .
- FIG. 7 is a schematic diagram showing another display example of the interpretation screen on the display unit 59.
- the interpretation screen 130 shown in FIG. 7 includes a position display area 131 instead of the position display area 118 of FIG.
- a humanoid image 132 simulating the subject 10 and a plurality of areas (divided areas) 133 divided into a 6 ⁇ 6 matrix, for example, are displayed.
- the imaging position of the in-vivo image is represented by filling the corresponding divided area 133 on the humanoid image 132 with a predetermined color or pattern based on the position information acquired by the position information acquisition unit 57.
- FIG. 7 shows a plurality of filled divided areas 134.
- the display color and brightness of the divided area 134 may be changed according to the number of in-vivo images having the divided area 134 as an imaging position. Specifically, it is conceivable to increase the brightness of the divided areas where the imaging positions are dense. Further, when a signal for selecting one of the divided areas 134 is input to the image display device 5 by performing a pointer operation or the like by the user, an in-vivo image corresponding to the divided area is displayed in the main display area 113. You may do it.
- the display control unit 58 sets the divided region 133 in more detail (for example, a 9 ⁇ 9 matrix) in accordance with the progress of the position estimation process for in-vivo images other than the priority image after the position estimation process for the priority image is completed. Etc.) It may be divided and displayed.
- the position estimation process is preferentially performed on the in-vivo image in which the lesion part or organ boundary is detected by the detection image process, and the preferential position estimation process is completed. Since the interpretation screen can be displayed in stages, the user can start interpretation early without waiting for the end of the position estimation processing for all in-vivo images. In addition, since the position information of important in-vivo images that should be noted in diagnosis has already been acquired at the time of the interpretation disclosure, the user can grasp the imaging position of such in-vivo images at an early stage and efficiently interpret the images. Can be done.
- the detection image processing unit 56 starts the detection image processing after the display image processing for all the in-vivo images is completed, but the in-vivo image data that has undergone display image processing is stored in the storage unit. After starting to be stored in 55, the image processing for detection may be executed in parallel.
- the position information acquisition unit 57 starts the position estimation process for the priority image after the detection image processing for all the in-vivo images is completed. May be executed in parallel.
- the priority processing control unit 62 controls the position information acquisition unit 57 to interrupt the priority image and execute the position estimation process.
- FIG. 8 is a block diagram showing a configuration of the image display apparatus according to the second embodiment.
- the image display device 5-2 illustrated in FIG. 8 includes a priority image determination unit 63 provided at a subsequent stage of the display image processing unit 53, and a priority processing control unit 64 that controls the position information acquisition unit 57.
- Other configurations are the same as those shown in FIG.
- an in-vivo image is given to the user (examiner such as a doctor). Is observed in real time, and a desired in-vivo image is marked. Specifically, the in-vivo image is displayed on the display unit 36 of the receiving device 3 shown in FIG. When this happens, the marking information is input to the receiving device. This marking information is taken into the image display device 5 as related information of in-vivo image data. Note that a separate viewer may be connected to the receiving device 3, and the in-vivo image may be displayed on the viewer instead of the display unit 36 to allow the user to observe.
- the priority image determination unit 63 determines whether or not the marking information is included in the related information of the in-vivo image data subjected to the display image processing, and is a priority image with respect to the in-vivo image data including the marking information. Add a flag to that effect.
- the priority processing control unit 64 controls the position information acquisition unit 57 to preferentially execute the position estimation process for the in-vivo image data to which the flag is attached in the priority image determination unit 63.
- FIG. 9 is a flowchart showing the operation of the image display device 5-2.
- the display image processing unit 53 performs display image processing on the in-vivo image data stored in the temporary storage unit 52.
- the priority image determination unit 63 attaches a flag indicating that the image is the priority image to the in-vivo image data having the marking information.
- the in-vivo image data subjected to the display image processing is stored in the storage unit 55.
- step S ⁇ b> 23 the position information acquisition unit 57 sequentially takes in the relevant information of the in-vivo image data with the flag indicating that the image is a priority image from the storage unit 55 under the control of the priority processing control unit 64, and performs position estimation processing. Execute. The position information generated thereby is stored in the storage unit 55 in association with the in-vivo image data.
- the display control unit 58 causes the display unit 59 to display a message notifying that the interpretation screen can be displayed.
- the display control unit 58 may display an interpretation screen (see FIGS. 6 and 7) on the display unit 59 when a predetermined time has elapsed after the display of the message.
- a screen that allows the user to select whether or not to display an interpretation screen may be displayed.
- the display control unit 58 causes the display unit 59 to display the interpretation screen when a signal for selecting the interpretation screen display is input by a user operation. Further, a reduced image of the priority image may be displayed in the thumbnail area 116 of the interpretation screen.
- step S25 the detection image processing unit 56 sequentially captures the in-vivo image data stored in the storage unit 55 and performs detection image processing.
- step S ⁇ b> 26 the position information acquisition unit 57 sequentially fetches related information of in-vivo image data other than the priority image from the storage unit 55 and executes position estimation processing.
- the display control unit 58 sequentially adds and displays a point 121 representing the imaging position of the in-vivo image subjected to the position estimation process in the position display area 118.
- the display unit 59 is controlled.
- the points 121 corresponding to the priority images and the points 121 corresponding to the in-vivo images other than the priority images may be displayed with different colors, marks, or luminances.
- the data processing by the image display device 5-2 is completed. Thereafter, the image display device 5-2 continues to display the interpretation screen when the interpretation screen display has already been started. If the interpretation screen display is not started after step S24, a screen for allowing the user to select whether or not to start the interpretation screen display may be displayed again.
- the position estimation process is preferentially performed on the in-vivo image (marked in-vivo image) that the user has paid attention during the examination. It is possible to grasp the determined imaging position of the in-vivo image at an early stage and efficiently perform interpretation.
- the position information acquisition unit 57 starts the position estimation process for the priority image after the display image processing for all the in-vivo image data is completed, but directly receives the related information from the temporary storage unit 52.
- the capturing and position estimation processing may be executed in parallel.
- the priority processing control unit 64 controls the position information acquisition unit 57 to interrupt the priority image and execute the position estimation process.
- FIG. 10 is a block diagram illustrating a configuration of the image display apparatus according to the third embodiment.
- the image display device 5-3 illustrated in FIG. 10 includes a first priority image determination unit 65, a first priority processing control unit 66, a second priority image determination unit 67, and a second priority processing control unit 68.
- Other configurations are the same as those shown in FIG.
- the user observes the in-vivo image in real time, and marks the desired in-vivo image.
- the first priority image determination unit 65 determines whether or not the marking information is included in the related information of the in-vivo image data subjected to the display image processing, and the priority image is determined for the in-vivo image data including the marking information. Add a flag to the effect.
- the first priority processing control unit 66 controls the detection image processing unit 56 to preferentially perform the detection image processing on the in-vivo image data to which the flag is attached in the priority image determination unit 65.
- the second priority image determination unit 67 determines whether or not a lesion or organ boundary has been extracted from the in-vivo image that has been subjected to the detection image processing, and uses the in-vivo image data from which the lesion or organ boundary has been extracted. On the other hand, a flag indicating that the image is a priority image is attached.
- the second priority processing control unit 68 controls the position information acquisition unit 57 to preferentially execute the position estimation process for the in-vivo image data to which the flag is attached in the second priority image determination unit 67.
- FIG. 11 is a flowchart showing the operation of the image display device 5-3.
- the display image processing unit 53 performs display image processing on the in-vivo image data stored in the temporary storage unit 52.
- the first priority image determination unit 65 adds a flag indicating that the in-vivo image data having marking information is a marked priority image.
- the in-vivo image data subjected to the display image processing is stored in the storage unit 55.
- step S33 under the control of the first priority processing control unit 66, the detection image processing unit 56 sequentially takes in-vivo image data with a flag indicating that it is a marked priority image from the storage unit 55, Image processing for detection is performed.
- step S ⁇ b> 34 the second priority image determination unit 67 sets a flag indicating that it is a priority image including a lesion or the like with respect to the in-vivo image data in which a lesion or an organ boundary is detected by the detection image processing unit 56. Attached.
- step S ⁇ b> 35 the position information acquisition unit 57 receives, from the storage unit 55, related information of the in-vivo image data with the flag indicating that the image is a priority image including a lesion under the control of the second priority processing control unit 68. Sequential capture and position estimation processing are executed. That is, in this step S35, the position estimation process is executed for the in-vivo image having the marking information and in which the lesion is detected.
- the display control unit 58 causes the display unit 59 to display a message notifying that the interpretation screen can be displayed in step S36.
- the display control unit 58 may display an interpretation screen (see FIGS. 6 and 7) on the display unit 59 when a predetermined time has elapsed after the display of the message.
- a screen that allows the user to select whether or not to display an interpretation screen may be displayed. In the latter case, the display control unit 58 causes the display unit 59 to display the interpretation screen when a signal for selecting the interpretation screen display is input by a user operation.
- step S37 the detection image processing unit 56 sequentially takes in-vivo image data other than the marked priority image from the storage unit 55, and performs detection image processing.
- step S ⁇ b> 38 the position information acquisition unit 57 sequentially retrieves related information of in-vivo image data other than the priority image including a lesion from the storage unit 55 and executes position estimation processing.
- the display control unit 58 sequentially adds and displays a point 121 representing the imaging position of the in-vivo image subjected to the position estimation process in the position display area 118.
- the display unit 59 is controlled.
- the points 121 may be displayed with different colors, marks, and brightness so that the points 121 corresponding to the priority images and the points corresponding to the in-vivo images other than the priority images can be distinguished.
- the data processing operation by the image display device 5-3 is finished. Thereafter, the image display device 5-3 continues to display the image interpretation screen when the image interpretation screen has already been displayed. If the interpretation screen display is not started after step S36, a screen for allowing the user to select whether or not to start the interpretation screen display may be displayed.
- the in-vivo image (marked in-vivo image) focused on by the user during the examination is subjected to detection image processing, and as a result, the in-vivo in which the lesion is detected.
- the position estimation process is preferentially executed for the image. Therefore, the user can grasp the imaging position of the in-vivo image determined to be important by the user himself and to be noticed by image processing at an early stage, and can efficiently interpret the image.
- the detection image processing unit 56 starts the detection image processing for the priority image after the display image processing for all the in-vivo image data is completed, but the display image processing is performed.
- the detection image processing may be executed in parallel when the in-vivo image data starts to be stored in the storage unit 55.
- the first priority processing control unit 66 controls the detection image processing unit 56 to interrupt the priority image and perform the detection image processing.
- the position information acquisition unit 57 starts the position estimation process for the priority image including the lesion after the detection image processing for the marked priority image is completed.
- the position information acquisition unit 57 may directly fetch the related information from the temporary storage unit 52 and execute the position estimation process in parallel with the display image process and the detection image process.
- the second priority processing control unit 68 controls the position information acquisition unit 57 to interrupt the priority image and execute the position estimation process.
- FIG. 12 is a block diagram showing a configuration of the image display apparatus according to the fourth embodiment.
- An image display device 5-4 illustrated in FIG. 12 includes a first priority image determination unit 65, a second priority image determination unit 67, and a priority processing control unit 69.
- the operations of the first priority image determination unit 65 and the second priority image determination unit 67 are the same as those described in the third embodiment.
- Other configurations are the same as those shown in FIG.
- the fourth embodiment as in the second embodiment, during the examination by the capsule endoscope 2, the in-vivo image is observed in real time by the user and the desired in-vivo image is marked.
- the priority processing control unit 69 includes the in-vivo image data flagged by the first priority image determination unit 65 (that is, the marked priority image) and the in-vivo image data flagged by the second priority image determination unit 67.
- the position information acquisition unit 57 is controlled so that the position estimation process is preferentially executed for (that is, a priority image including a lesion or the like).
- FIG. 13 is a flowchart showing the operation of the image display device 5-4.
- the display image processing unit 53 performs display image processing on the in-vivo image data stored in the temporary storage unit 52.
- the first priority image determination unit 65 attaches a flag indicating that the in-vivo image data having marking information is a marked priority image.
- the in-vivo image data subjected to the display image processing is stored in the storage unit 55.
- step S43 the detection image processing unit 56 sequentially captures the in-vivo image data stored in the storage unit 55 and performs detection image processing.
- the second priority image determination unit 67 sets a flag indicating that the in-vivo image data in which the lesion part or the organ boundary is detected by the detection image processing unit 56 is a priority image including a lesion or the like. Is attached.
- step S45 the position information acquisition unit 57, under the control of the priority processing control unit 69, gives priority information including related information of in-vivo image data with a flag indicating that it is a marked priority image, a lesion, and the like.
- the related information of the in-vivo image data attached with the flag indicating that is is sequentially fetched from the storage unit 55, and the position estimation process is executed. That is, in this step S45, a position estimation process is executed for an in-vivo image having marking information or having a lesion or the like detected.
- the display control unit 58 causes the display unit 59 to display a message notifying that the interpretation screen can be displayed in step S46.
- the display control unit 58 may display an interpretation screen (see FIGS. 6 and 7) on the display unit 59 when a predetermined time has elapsed after the display of the message. A screen that allows the user to select whether or not to display an interpretation screen may be displayed. In the latter case, the display control unit 58 causes the display unit 59 to display the interpretation screen when a signal for selecting the interpretation screen display is input by a user operation.
- step S47 the position information acquisition unit 57 sequentially fetches relevant information of in-vivo image data other than the priority image including the marked priority image and the lesion from the storage unit 55, and executes the position estimation process.
- the display control unit 58 sequentially adds and displays a point 121 representing the imaging position of the in-vivo image subjected to the position estimation process in the position display area 118.
- the display unit 59 may be controlled.
- the point 121 corresponding to the marked priority image, the point 121 corresponding to the priority image including a lesion, and the point 121 corresponding to the in-vivo image other than the priority image can be distinguished, for example, You may display those points 121 by a mutually different color, a mark, and brightness
- the data processing by the image display device 5-4 is completed. Thereafter, the image display device 5-4 continues to display the image interpretation screen when the image interpretation screen has already been displayed. In addition, if display of the interpretation screen is not started after step S46, a screen for allowing the user to select whether or not to start display of the interpretation screen may be displayed again.
- the position estimation process is preferentially executed for both the in-vivo image focused by the user during the examination and the in-vivo image in which a lesion or the like is detected, the user Therefore, it is possible to grasp the in-vivo image determined to be important by the user himself / herself and the imaging position of the in-vivo image determined to be noticed by the image processing widely and early, and to efficiently interpret the image.
- the detection image processing unit 56 starts the detection image processing after the display image processing for all the in-vivo image data is completed, but the in-vivo image data that has been subjected to the display image processing.
- the image processing for detection may be executed in parallel at the stage where the storage unit 55 starts to store.
- the position information acquisition unit 57 starts the position estimation process for the priority image after the detection image processing for all the in-vivo image data is completed.
- the position information acquisition unit 57 estimates the position in parallel with the detection image processing. Processing may be executed.
- the position information acquisition unit 57 may directly acquire related information from the temporary storage unit 52 and execute the position estimation process in parallel with the display image process and the detection image process. In these cases, when a priority image is generated by the determination by the first priority image determination unit 65 or the second priority image determination unit 67, the priority processing control unit 69 interrupts the priority image and executes position estimation processing.
- the position information acquisition unit 57 is controlled.
- FIG. 14 is a block diagram showing the configuration of the image display apparatus according to the fifth embodiment.
- An image display device 5-5 illustrated in FIG. 14 includes a priority image determination unit 71 and a priority processing control unit 72. Other configurations are the same as those shown in FIG.
- the image display device 5-5 is characterized in that the user selects a desired in-vivo image as a priority image using the operation input device 5b while observing an interpretation screen displayed on the display unit 59.
- the priority image determination unit 71 determines whether the position estimation process has been completed for the in-vivo image selected by the user based on the operation signal input from the outside via the interface unit 51, and the position estimation process is completed. If not, a flag indicating that the selected in-vivo image data is a priority image is attached.
- the priority processing control unit 72 controls the position information acquisition unit 57 to preferentially execute the position estimation process for the in-vivo image data to which the flag is attached in the priority image determination unit 71.
- FIG. 15 is a flowchart showing the operation of the image display device 5-5.
- the display image processing unit 53 performs display image processing on the in-vivo image data stored in the temporary storage unit 52.
- the in-vivo image data subjected to the display image processing is stored in the storage unit 55.
- the detection image processing unit 56 sequentially takes in-vivo image data on which the display image processing has been performed from the storage unit 55, and performs detection image processing.
- the in-vivo image can be displayed on the display unit 59 when the detection image processing is completed. Therefore, when the detection image processing in step S52 is completed, the display control unit 58 causes the display unit 59 to display a message notifying that the interpretation screen can be displayed in step S53. At this time, the display control unit 58 may display an interpretation screen on the display unit 59 when a predetermined time has elapsed after the display of the message, or whether or not to actually display the interpretation screen together with the message. You may display the screen which makes a user select. In the latter case, the display control unit 58 causes the display unit 59 to display the interpretation screen when a signal for selecting the interpretation screen display is input by a user operation.
- FIG. 16 is a schematic diagram showing a display example of an interpretation screen that can be displayed after step S53.
- the interpretation screen 140 includes a button 141 that is clicked when the user wants to display the imaging position of the in-vivo image being displayed in the main display area 113.
- step S54 the image display device 5-5 starts position estimation processing.
- step S55 the image display device 5-5 selects the in-vivo image being displayed by the user's operation using the operation input device 5b (for example, by clicking the “position confirmation” button 141). It is determined whether or not a signal is input.
- step S55 the priority image determination unit 71 determines whether or not the position estimation process has already been executed for the selected in-vivo image (step S56).
- step S56 the priority image determination unit 71 adds a flag indicating that the selected in-vivo image data is a priority image.
- step S57 the position information acquisition unit 57 takes in the relevant information of the in-vivo image data with the flag from the storage unit 55 under the control of the priority processing control unit 72.
- step S58 the position information acquisition unit 57 executes position estimation processing based on the acquired related information.
- step S59 the display control unit 58 causes the display unit 59 to display the imaging position of the in-vivo image on the interpretation screen based on the position information generated as a result of the position estimation process.
- step S60: Yes the operation returns to step S55.
- step S60: No the position estimation process ends (step S61).
- FIG. 17 is a schematic diagram showing an example of the imaging position display of the in-vivo image on the interpretation screen 140.
- the imaging position of the in-vivo image on which the position estimation processing has been executed so far is displayed by a point 142.
- the imaging position of the selected in-vivo image is displayed by a point 143.
- the point 143 may be displayed with a color, brightness, or mark different from that of the point 142.
- step S55 when the selection signal is not input (step S55: No), the position information acquisition unit 57 captures the related information of the in-vivo image data from the storage unit 55 in the normal order (for example, the transfer order of the in-vivo image data) (step S55). S62). Thereafter, the operation proceeds to step S58.
- step S55 Yes
- step S56 Yes
- step S63 the display control unit 58 selects the selected in-vivo image. Is extracted from the storage unit 55 (step S63). Thereafter, the operation proceeds to step S59.
- the position estimation process is preferentially performed on the in-vivo image selected by the user who is interpreting the image, the user can capture the in-vivo image determined to be necessary during the interpretation. It becomes possible to grasp the position at an early stage.
- the priority image determination unit 71 and the priority processing control unit 72 in the fifth embodiment may be provided in the image display devices according to the first to fourth embodiments.
- the position information acquisition unit 57 may execute position estimation processing of the in-vivo image selected by the user during interpretation, in preference to the priority image including the marked priority image or lesion.
- FIG. 18 is a block diagram illustrating a configuration of an image display apparatus according to the sixth embodiment.
- An image display device 5-6 shown in FIG. Other configurations are the same as those shown in FIG.
- the trajectory creation unit 73 creates a trajectory that passes between the time when the capsule endoscope 2 is introduced into the subject 10 and discharged after the position estimation processing for all in-vivo images is performed.
- the trajectory creation process is executed. Specifically, the trajectory creation unit 73 calculates two points that are temporally adjacent from a plurality of positions of the capsule endoscope 2 at the time of capturing the in-vivo image based on the position information generated by the position information acquisition unit 57. Extract and connect those two points.
- the trajectory creation unit 73 creates a total trajectory by sequentially connecting the positions estimated in this way. Information representing the created trajectory (trajectory information) is stored in the storage unit 55.
- various known methods can be applied as a specific method of the trajectory creation process.
- FIG. 19 is a schematic diagram showing a display example of the locus on the interpretation screen.
- a locus 151 drawn and displayed based on the locus information is displayed.
- a point 152 representing the imaging position of the in-vivo image currently displayed in the main display area 113 is displayed on the locus 151 in an overlapping manner.
- the trajectory creation unit 73 in the sixth embodiment may be provided in the image display devices according to the second to fifth embodiments.
Landscapes
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Surgery (AREA)
- Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Optics & Photonics (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Veterinary Medicine (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Physics & Mathematics (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Endoscopes (AREA)
Abstract
Description
図1は、本発明の実施の形態1に係るカプセル型内視鏡システムの概略構成を示す模式図である。このカプセル型内視鏡システム1は、被検体10の体内に導入されて撮像を行い、体内画像の画像データを受信装置3に無線送信するカプセル型内視鏡2と、カプセル型内視鏡2から無線送信された体内画像データを受信する受信装置3と、受信装置3が受信した体内画像データに基づいて体内画像を表示する画像表示装置5とを備える。
表示部59は、CRTディスプレイや液晶ディスプレイによって実現され、表示制御部58の制御の下で、被検体10の体内画像が順次表示される読影画面、種々の情報、ユーザへの通知メッセージ等を表示する。
まず、ステップS11において、表示用画像処理部53は、一時記憶部52に格納された体内画像データに対して表示用画像処理を施す。表示用画像処理を施された体内画像データは、記憶部55に格納される。
次に、本発明の実施の形態2に係る画像表示装置について説明する。図8は、実施の形態2に係る画像表示装置の構成を示すブロック図である。図8に示す画像表示装置5-2は、表示用画像処理部53の後段に設けられた優先画像判定部63と、位置情報取得部57を制御する優先処理制御部64とを備える。その他の構成については、図4に示すものと同様である。
ステップS21において、表示用画像処理部53は、一時記憶部52に格納された体内画像データに対して表示用画像処理を施す。続くステップS22において、優先画像判定部63は、マーキング情報を有する体内画像データに対し、優先画像である旨のフラグを附す。表示用画像処理が施された体内画像データは、記憶部55に格納される。
次に、本発明の実施の形態3に係る画像表示装置について説明する。図10は、実施の形態3に係る画像表示装置の構成を示すブロック図である。図10に示す画像表示装置5-3は、第1優先画像判定部65と、第1優先処理制御部66と、第2優先画像判定部67と、第2優先処理制御部68とを備える。その他の構成については、図4に示すものと同様である。また、実施の形態3においても、実施の形態2と同様に、カプセル型内視鏡2による検査中、ユーザに体内画像をリアルタイムで観察させ、所望の体内画像にマーキングさせる。
ステップS31において、表示用画像処理部53は、一時記憶部52に格納された体内画像データに対して表示用画像処理を施す。
次に、本発明の実施の形態4に係る画像表示装置について説明する。図12は、実施の形態4に係る画像表示装置の構成を示すブロック部である。図12に示す画像表示装置5-4は、第1優先画像判定部65と、第2優先画像判定部67と、優先処理制御部69とを備える。第1優先画像判定部65及び第2優先画像判定部67の動作は、実施の形態3において説明したものと同様である。また、その他の構成については、図4に示すものと同様である。また、実施の形態4においても、実施の形態2と同様に、カプセル型内視鏡2による検査中、ユーザに体内画像をリアルタイムで観察させ、所望の体内画像にマーキングさせる。
ステップS41において、表示用画像処理部53は、一時記憶部52に格納された体内画像データに対して表示用画像処理を施す。
次に、本発明の実施の形態5に係る画像表示装置について説明する。図14は、実施の形態5に係る画像表示装置の構成を示すブロック部である。図14に示す画像表示装置5-5は、優先画像判定部71と、優先処理制御部72とを備える。その他の構成については、図4に示すものと同様である。この画像表示装置5-5は、ユーザが表示部59に表示された読影画面を観察しながら、操作入力デバイス5bを用いて所望の体内画像を優先画像として選択することを特徴とする。
ステップS51において、表示用画像処理部53は、一時記憶部52に格納された体内画像データに対して表示用画像処理を施す。表示用画像処理が施された体内画像データは、記憶部55に格納される。続くステップS52において、検出用画像処理部56は、表示用画像処理が施された体内画像データを記憶部55から順次取り込み、検出用画像処理を施す。
その際にまず、ステップS55において、画像表示装置5-5は、操作入力デバイス5bを用いたユーザの操作(例えば、「位置確認」ボタン141をクリック)により、表示中の体内画像を選択する選択信号が入力されたか否かを判定する。
次に、本発明の実施の形態6に係る画像表示装置について説明する。図18は、実施の形態6に係る画像表示装置の構成を示すブロック図である。図18に示す画像表示装置5-6は、軌跡作成部73を備えている。その他の構成については、図4に示すものと同様である。
なお、実施の形態6における軌跡作成部73を、実施の形態2~5に係る画像表示装置に設けても良い。
2 カプセル型内視鏡
3 受信装置
4 アンテナユニット
5、5-2~5-6 画像表示装置
5a クレードル
5b 操作入力デバイス
10 被検体
20 カプセル型容器
20a 収容部
20b 光学ドーム
21 撮像部
21a 撮像素子
21b 光学系
22 照明部
23 回路基板
24、32 信号処理部
25、33 メモリ
26 送信部
27 アンテナ
28、37 バッテリ
31 受信部
34、51 インタフェース部
35 操作部
36 表示部
41a 受信アンテナ
52 一時記憶部
53 表示用画像処理部
54 検査情報作成部
55 記憶部
56 検出用画像処理部
57 位置情報取得部
58 表示制御部
59 表示部
61、63、65、67、71 優先画像判定部
62、64、66、68、69、72 優先処理制御部
73 軌跡作成部
110、130、140、150 読影画面
111 患者情報領域
112 診察情報領域
113 主表示領域
114 再生操作ボタン群
115 縮小画像
116 サムネイル領域
117 タイムバー
118 位置表示領域
120、132 人型画像
131 位置表示領域
133、134 分割領域
141 ボタン
121、142、143、152 ポイント
151 軌跡
Claims (9)
- 被検体の体内画像を撮像するカプセル型内視鏡から、該カプセル型内視鏡と無線通信を行う受信装置を介して取得された体内画像データに基づく体内画像群を表示する画像表示装置であって、
前記体内画像データと、該体内画像データに関連付けられ、前記被検体内における前記カプセル型内視鏡の位置に関連する情報とを格納する記憶部と、
前記位置に関連する情報に基づいて位置推定処理を実行することにより、各体内画像の撮像時における前記カプセル型内視鏡の位置情報を取得する位置情報取得部と、
前記各体内画像が所定の条件を満たすか否かを判定する判定部と、
前記判定部において前記所定の条件を満たすと判定された体内画像に対して優先的に前記位置推定処理を実行するように前記位置情報取得部を制御する優先処理制御部と、
を備えることを特徴とする画像表示装置。 - 前記記憶部に格納された体内画像データに対して所定の部位を検出する検出用画像処理を施す検出用画像処理部をさらに備え、
前記所定の条件は、前記検出用画像処理において前記所定の部位が検出されたことを特徴とする請求項1に記載の画像表示装置。 - 前記所定の条件は、前記カプセル型内視鏡が前記被検体内を通過している間に外部から前記受信装置に入力された信号に従って付加されたマーキング情報を有することを特徴とする請求項1に記載の画像表示装置。
- 前記記憶部に格納された体内画像データに対して所定の部位を検出する検出用画像処理を施す検出用画像処理部をさらに備え、
前記所定の条件は、前記カプセル型内視鏡が前記被検体内を通過している間に外部から前記受信装置に入力された信号に従って付加されたマーキング情報を有し、且つ、前記検出用画像処理部において前記所定の部位が検出されたことを特徴とする請求項1に記載の画像表示装置。 - 前記検出用画像処理部は、前記マーキング情報を有する体内画像に対して優先的に前記検出用画像処理を施すことを特徴とする請求項4に記載の画像表示装置。
- 前記記憶部に格納された体内画像データに対して所定の部位を検出する検出用画像処理を施す検出用画像処理部をさらに備え、
前記所定の条件は、前記カプセル型内視鏡が前記被検体内を通過している間に外部から前記受信装置に入力された信号に従って付加されたマーキング情報を有することと、前記検出用画像処理部において前記所定の部位が検出されたこととの内の少なくとも一方を満たすことを特徴とする請求項1に記載の画像表示装置。 - 前記記憶部に格納された体内画像データに対して表示用画像処理を施す表示用画像処理部と、
前記表示用画像処理が施された体内画像データに基づいて、前記体内画像群を構成する複数の体内画像が順次表示される読影画面を表示する表示部と、
前記優先処理制御部の制御の下で実行された前記位置推定処理の終了後、前記読影画面を前記表示部に表示させる表示制御部と、
をさらに備えることを特徴とする請求項1に記載の画像表示装置。 - 外部から当該画像表示装置に入力される信号を受け付ける入力信号受付部をさらに備え、
前記優先処理制御部は、前記入力信号受付部によって所定の体内画像を選択する選択信号が受け付けられた場合に、前記所定の体内画像について優先的に前記位置推定処理を実行させることを特徴とする請求項1に記載の画像表示装置。 - 被検体の体内に導入されて撮像を行い、該被検体の体内画像を表す体内画像データを生成するカプセル型内視鏡と、
前記カプセル型内視鏡により生成された体内画像データを無線通信により受信する受信装置と、
請求項1に記載の画像表示装置と、
を備えることを特徴とするカプセル型内視鏡システム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11840277.5A EP2556788B1 (en) | 2010-11-08 | 2011-10-21 | Image display apparatus and capsule endoscopy system |
CN201180016338.0A CN102843950B (zh) | 2010-11-08 | 2011-10-21 | 图像显示装置以及胶囊型内窥镜系统 |
JP2012521817A JP5044066B2 (ja) | 2010-11-08 | 2011-10-21 | 画像表示装置及びカプセル型内視鏡システム |
US13/466,167 US8711205B2 (en) | 2010-11-08 | 2012-05-08 | Image display device and capsule endoscope system |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-250074 | 2010-11-08 | ||
JP2010250074 | 2010-11-08 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/466,167 Continuation US8711205B2 (en) | 2010-11-08 | 2012-05-08 | Image display device and capsule endoscope system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012063623A1 true WO2012063623A1 (ja) | 2012-05-18 |
Family
ID=46050775
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/074345 WO2012063623A1 (ja) | 2010-11-08 | 2011-10-21 | 画像表示装置及びカプセル型内視鏡システム |
Country Status (5)
Country | Link |
---|---|
US (1) | US8711205B2 (ja) |
EP (1) | EP2556788B1 (ja) |
JP (1) | JP5044066B2 (ja) |
CN (1) | CN102843950B (ja) |
WO (1) | WO2012063623A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018008195A1 (ja) * | 2016-07-05 | 2018-01-11 | オリンパス株式会社 | 画像処理装置、画像処理システム、画像処理装置の作動方法、及び画像処理装置の作動プログラム |
JP6425868B1 (ja) * | 2017-09-29 | 2018-11-21 | オリンパス株式会社 | 内視鏡画像観察支援システム、内視鏡画像観察支援装置、内視鏡画像観察支援方法 |
WO2019064704A1 (ja) * | 2017-09-29 | 2019-04-04 | オリンパス株式会社 | 内視鏡画像観察支援システム、内視鏡画像観察支援装置、内視鏡画像観察支援方法 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5568198B1 (ja) * | 2012-10-24 | 2014-08-06 | オリンパスメディカルシステムズ株式会社 | 検査管理装置、検査管理システム、検査管理方法、及び検査管理プログラム |
JP5802861B2 (ja) * | 2013-08-28 | 2015-11-04 | オリンパス株式会社 | カプセル型内視鏡システム |
CN103593655B (zh) * | 2013-11-20 | 2017-08-22 | 深圳先进技术研究院 | 无线胶囊内窥镜图像识别方法及系统 |
CN107007242A (zh) * | 2017-03-30 | 2017-08-04 | 深圳市资福技术有限公司 | 一种胶囊式内窥镜控制方法及装置 |
DE102017130980B4 (de) * | 2017-12-21 | 2024-07-25 | Schölly Fiberoptic GmbH | Bildübertragungsanordnung und Verfahren zur Bildübertragung |
CN109259715B (zh) * | 2018-09-04 | 2021-04-27 | 北京理工大学 | 集成交互功能的胶囊内镜磁引导控制装置 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005168524A (ja) * | 2003-12-05 | 2005-06-30 | Olympus Corp | カプセル型内視鏡撮像画像ファイリング装置 |
JP2005218584A (ja) * | 2004-02-04 | 2005-08-18 | Olympus Corp | 画像情報の表示処理装置、その表示処理方法及び表示処理プログラム |
JP2006075301A (ja) | 2004-09-08 | 2006-03-23 | Olympus Corp | 画像処理装置、画像処理方法および画像処理プログラム |
JP2007111205A (ja) * | 2005-10-19 | 2007-05-10 | Olympus Corp | 受信装置およびこれを用いた被検体内情報取得システム |
JP2007283001A (ja) | 2006-04-19 | 2007-11-01 | Olympus Medical Systems Corp | カプセル型医療装置 |
JP2007319478A (ja) * | 2006-06-01 | 2007-12-13 | Fujifilm Corp | 医用画像表示装置及び方法、並びに内視鏡装置 |
JP2008301877A (ja) | 2007-06-05 | 2008-12-18 | Olympus Corp | 画像処理装置および画像処理プログラム |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2674349B1 (fr) * | 1991-03-20 | 1993-07-02 | Armines | Procede de traitement d'images par files d'attente hierarchisees. |
ES2365696T3 (es) * | 2001-03-14 | 2011-10-10 | Given Imaging Ltd. | Método y sistema para detectar anormalidades colorimétricas. |
JP4493386B2 (ja) * | 2003-04-25 | 2010-06-30 | オリンパス株式会社 | 画像表示装置、画像表示方法および画像表示プログラム |
JP3810381B2 (ja) * | 2003-04-25 | 2006-08-16 | オリンパス株式会社 | 画像表示装置、画像表示方法および画像表示プログラム |
EP1690491A4 (en) * | 2003-12-05 | 2011-04-13 | Olympus Corp | DISPLAY PROCESSING DEVICE |
EP2415389B1 (en) * | 2005-04-13 | 2016-05-04 | Olympus Corporation | Image processing apparatus and method for analysing mucosa, villi or feces |
JP4961475B2 (ja) * | 2007-06-20 | 2012-06-27 | オリンパスメディカルシステムズ株式会社 | 内視鏡システム |
-
2011
- 2011-10-21 WO PCT/JP2011/074345 patent/WO2012063623A1/ja active Application Filing
- 2011-10-21 JP JP2012521817A patent/JP5044066B2/ja active Active
- 2011-10-21 CN CN201180016338.0A patent/CN102843950B/zh active Active
- 2011-10-21 EP EP11840277.5A patent/EP2556788B1/en not_active Not-in-force
-
2012
- 2012-05-08 US US13/466,167 patent/US8711205B2/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005168524A (ja) * | 2003-12-05 | 2005-06-30 | Olympus Corp | カプセル型内視鏡撮像画像ファイリング装置 |
JP2005218584A (ja) * | 2004-02-04 | 2005-08-18 | Olympus Corp | 画像情報の表示処理装置、その表示処理方法及び表示処理プログラム |
JP2006075301A (ja) | 2004-09-08 | 2006-03-23 | Olympus Corp | 画像処理装置、画像処理方法および画像処理プログラム |
JP2007111205A (ja) * | 2005-10-19 | 2007-05-10 | Olympus Corp | 受信装置およびこれを用いた被検体内情報取得システム |
JP2007283001A (ja) | 2006-04-19 | 2007-11-01 | Olympus Medical Systems Corp | カプセル型医療装置 |
JP2007319478A (ja) * | 2006-06-01 | 2007-12-13 | Fujifilm Corp | 医用画像表示装置及び方法、並びに内視鏡装置 |
JP2008301877A (ja) | 2007-06-05 | 2008-12-18 | Olympus Corp | 画像処理装置および画像処理プログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP2556788A4 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018008195A1 (ja) * | 2016-07-05 | 2018-01-11 | オリンパス株式会社 | 画像処理装置、画像処理システム、画像処理装置の作動方法、及び画像処理装置の作動プログラム |
JP6326178B1 (ja) * | 2016-07-05 | 2018-05-16 | オリンパス株式会社 | 画像処理装置、画像処理システム、画像処理装置の作動方法、及び画像処理装置の作動プログラム |
US10726553B2 (en) | 2016-07-05 | 2020-07-28 | Olympus Corporation | Image processing apparatus, image processing system, operation method of image processing apparatus, and computer-readable recording medium |
JP6425868B1 (ja) * | 2017-09-29 | 2018-11-21 | オリンパス株式会社 | 内視鏡画像観察支援システム、内視鏡画像観察支援装置、内視鏡画像観察支援方法 |
WO2019064704A1 (ja) * | 2017-09-29 | 2019-04-04 | オリンパス株式会社 | 内視鏡画像観察支援システム、内視鏡画像観察支援装置、内視鏡画像観察支援方法 |
US11556731B2 (en) | 2017-09-29 | 2023-01-17 | Olympus Corporation | Endoscopic image observation system, endosopic image observation device, and endoscopic image observation method |
Also Published As
Publication number | Publication date |
---|---|
US8711205B2 (en) | 2014-04-29 |
CN102843950A (zh) | 2012-12-26 |
JPWO2012063623A1 (ja) | 2014-05-12 |
EP2556788A1 (en) | 2013-02-13 |
JP5044066B2 (ja) | 2012-10-10 |
CN102843950B (zh) | 2015-02-04 |
EP2556788A4 (en) | 2013-06-26 |
US20120274743A1 (en) | 2012-11-01 |
EP2556788B1 (en) | 2016-09-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5044066B2 (ja) | 画像表示装置及びカプセル型内視鏡システム | |
JP4724259B2 (ja) | 画像表示装置、読影支援システムおよび読影支援プログラム | |
JP4493386B2 (ja) | 画像表示装置、画像表示方法および画像表示プログラム | |
US20090051695A1 (en) | Image processing apparatus, computer program product, and image processing method | |
JP5015363B2 (ja) | 画像表示装置及びカプセル型内視鏡システム | |
JP4823659B2 (ja) | 生体内画像表示装置 | |
JP2009039449A (ja) | 画像処理装置 | |
JP5242852B2 (ja) | 画像表示装置及びカプセル型内視鏡システム | |
JP2004321603A (ja) | 画像表示装置、画像表示方法および画像表示プログラム | |
JP4956694B2 (ja) | 情報処理装置及びカプセル型内視鏡システム | |
JP2007236700A (ja) | カプセル内視鏡システム | |
CN114945314A (zh) | 医疗图像处理装置、内窥镜系统、诊断辅助方法及程序 | |
JP5593008B1 (ja) | 画像処理装置及び画像処理方法 | |
JP4554647B2 (ja) | 画像表示装置、画像表示方法および画像表示プログラム | |
JP2007307396A (ja) | 画像表示装置、画像表示方法および画像表示プログラム | |
JP5341257B2 (ja) | 画像処理装置、画像処理装置の作動方法、画像処理プログラムおよび内視鏡システム | |
JP2007307397A (ja) | 画像表示装置、画像表示方法および画像表示プログラム | |
JP2005131031A (ja) | 画像表示装置、画像表示方法、及び画像表示プログラム | |
JP2004350725A (ja) | 医療用画像記録装置 | |
JP2006167222A (ja) | 画像処理装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180016338.0 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012521817 Country of ref document: JP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11840277 Country of ref document: EP Kind code of ref document: A1 |
|
REEP | Request for entry into the european phase |
Ref document number: 2011840277 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011840277 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |