WO2018140062A1 - Method and apparatus for endoscope with distance measuring for object scaling - Google Patents

Method and apparatus for endoscope with distance measuring for object scaling Download PDF

Info

Publication number
WO2018140062A1
WO2018140062A1 PCT/US2017/015668 US2017015668W WO2018140062A1 WO 2018140062 A1 WO2018140062 A1 WO 2018140062A1 US 2017015668 W US2017015668 W US 2017015668W WO 2018140062 A1 WO2018140062 A1 WO 2018140062A1
Authority
WO
WIPO (PCT)
Prior art keywords
camera
image sequence
capsule
distance information
image
Prior art date
Application number
PCT/US2017/015668
Other languages
French (fr)
Inventor
Kang-Huai Wang
Chenyu Wu
Yi Xu
Original Assignee
CapsoVision, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CapsoVision, Inc. filed Critical CapsoVision, Inc.
Priority to PCT/US2017/015668 priority Critical patent/WO2018140062A1/en
Priority to CN201810091804.5A priority patent/CN108392165A/en
Priority to JP2018013654A priority patent/JP2018130537A/en
Priority to US15/883,337 priority patent/US10402992B2/en
Publication of WO2018140062A1 publication Critical patent/WO2018140062A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/041Capsule endoscopes for imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00002Operational features of endoscopes
    • A61B1/00004Operational features of endoscopes characterised by electronic signal processing
    • A61B1/00009Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00163Optical arrangements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/05Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by the image sensor, e.g. camera, being in the distal end portion
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/06Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
    • A61B1/0605Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements for spatially modulated illumination
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/06Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
    • A61B1/0646Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements with illumination filters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0077Devices for viewing the surface of the body, e.g. camera, magnifying lens
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0082Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes
    • A61B5/0084Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes for introduction into the body, e.g. by catheters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/07Endoradiosondes
    • A61B5/073Intestinal transmitters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/1032Determining colour for diagnostic purposes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/107Measuring physical dimensions, e.g. size of the entire body or parts thereof
    • A61B5/1072Measuring physical dimensions, e.g. size of the entire body or parts thereof measuring distances on the body, e.g. measuring length, height or thickness
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/107Measuring physical dimensions, e.g. size of the entire body or parts thereof
    • A61B5/1076Measuring physical dimensions, e.g. size of the entire body or parts thereof for measuring dimensions inside body cavities, e.g. using catheters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/42Detecting, measuring or recording for evaluating the gastrointestinal, the endocrine or the exocrine systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7225Details of analog processing, e.g. isolation amplifier, gain or sensitivity adjustment, filtering, baseline or drift compensation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/521Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30092Stomach; Gastric

Definitions

  • TITLE Method and Apparatus for Endoscope with Distance Measuring for Object Scaling
  • the present invention relates to the endoscope for capturing images of human gastrointestinal (GI) tract for diagnosis purpose.
  • the endoscope is enabled to measure distance of objects in the field of view of the camera.
  • the distance information can be used subsequently to process the image sequence captured, such as measuring a size of an object of interest or stitching the image sequence to reduce viewing time.
  • Endoscopes are flexible or rigid tubes that pass into the body through an orifice or surgical opening, typically into the esophagus via the mouth or into the colon via the rectum.
  • An image is formed at the distal end using a lens and transmitted to the proximal end, outside the body, either by a lens-relay system or by a coherent fiber-optic bundle.
  • a conceptually similar instrument might record an image electronically at the distal end, for example using a CCD or CMOS array, and transfer the image data as an electrical signal to the proximal end through a cable.
  • Endoscopes allow a physician control over the field of view and are well-accepted diagnostic tools.
  • Capsule endoscope is an alternative in vivo endoscope developed in recent years.
  • a camera is housed in a swallowable capsule, along with a radio transmitter for transmitting data, primarily comprising images recorded by the digital camera, to a base-station receiver or transceiver and data recorder outside the body.
  • the capsule may also include a radio receiver for receiving instructions or other data from a base-station transmitter.
  • radio-frequency transmission lower-frequency electromagnetic signals may be used.
  • Power may be supplied inductively from an external inductor to an internal inductor within the capsule or from a battery within the capsule.
  • the endoscope When the endoscope is used for imaging the human GI tract, one of the primary purposes is to identify any possible anomaly. If any anomaly is found, it is further of interest to determine characteristics of the anomaly, such as the size of the anomaly.
  • the captured images will be examined by medical profession for examination or diagnosis. The number of images captured is typically 25,000 or more. It will require a long reviewing time to look through the images even by skilled professionals. Accordingly, image stitching has been used to reduce the number of images to be viewed. For example, in PCT Patent Application Publication, Serial No. WO2014/193670 A2, published on Dec. 4, 2014, image stitching for images captured using a capsule camera is disclosed. It is desirable to develop methods or apparatus that are capable to further improve the efficiency of image stitching. BRIEF SUMMARY OF THE INVENTION
  • a method and apparatus for capturing images of a scene using a capsule camera are disclosed. After the capsule camera is swallowed by a patient, an image sequence is captured using the capsule camera when the capsule camera travels through a human gastrointestinal tract. Also, distance information associated with objects of the image sequence with respect to the capsule camera is captured when the capsule camera travels through the human gastrointestinal tract. Both the image sequence and the distance information are outputted.
  • the association information between the distance information and related images of the image sequence is outputted.
  • the association information may correspond to frame numbers or capture times of the related images of the image sequence.
  • the distance information can be determined based on Time-of-Flight or phase shift using a light or ultrasound source.
  • the distance information may also be derived from structured-light images by projecting structured light to one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera.
  • the present invention also discloses a method of determining the size of an object of interest in an image.
  • the image sequence captured by a capsule camera along with the distance information is received.
  • the size of an object of interest in a selected image in the image sequence can be determined based on the selected image and the distance information.
  • the size of the object of interest is determined according to the image size of the object of interest in a selected image scaled by a ratio of object distance to the capsule camera and focal length of the capsule camera.
  • the image size of the object of interest in the selected image can be measured in terms of a number of pixels of object of interest in the selected image.
  • the distance information can be in a form of structured-light images that are captured by projecting a structured light onto one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera, and the distance information is further derived from the structured-light images before the distance information is used for said determining the size information of the object of interest.
  • the present invention further discloses a method of stitching the image sequence utilizing information including the distance information to generate a stitched image sequence.
  • the distance information can be in a form of structured-light images.
  • the distance information is used to scale the objects of the image sequence for stitching the image sequence. In another embodiment, the distance information is used to adjust image intensities of the image sequence for stitching the image sequence.
  • Fig. 1 illustrates an example of relationship among the size of an object, the size of the corresponding object image, the object distance and the focal length of the camera.
  • Fig. 2 A and Fig. 2B illustrate an example of the different sizes of object images for a same object in two images captured at two different object distances.
  • FIG. 3 illustrates an exemplary flowchart for capturing an image sequence along with distance information according to an embodiment of the present invention.
  • Fig. 4 illustrates an exemplary flowchart for determining a size of an object of interest in the image sequence based on the image sequence and the distance information according to an embodiment of the present invention.
  • FIG. 5 illustrates an exemplary flowchart for stitching an image sequence utilizing information including the distance information to generate a stitched image sequence according to an embodiment of the present invention.
  • Endoscopes are normally inserted into the human body through a natural opening such as the mouth or anus. Therefore, endoscopes are preferred to be small sizes so as to be minimally invasive. As mentioned before, endoscopes can be used for diagnosis of human gastrointestinal (GI) tract. The captured image sequence can be viewed to identify any possible anomaly. If any anomaly is found, it is of interest to identify the characteristics of the anomaly, such as the size. Accordingly, an invention of the present invention discloses an endoscope including a distance measuring means to measure the object distances between the camera and various locations of an object in the field of view of the camera.
  • the distance measuring devices there are various known means for measuring the distance between the camera and various locations of an object in the field of view of the camera. For example, there is a class of distance measuring devices that determine the distance based on ToF (Time of Flight) or phase shift of a light source.
  • the light source may be a laser or LED (Light Emitting Diode).
  • a light sensor is used to detect the returned light.
  • the time difference or phase difference between the emitted light from the light source and the received light by the light detector is used to determine the distance.
  • Ultrasonic wave is also a signal source that can be used to measure the distance between an object and the camera for the intended GI imaging application.
  • the distance measuring means is well-known in the field and various literatures describing the distance measure based on ToF or phase shift using light or ultrasound sources are readily available. Therefore, details for distance measuring means based on ToF or phase shift using light or ultrasound sources are omitted in this disclosure.
  • the light for measuring the distance may interfere with the flash light illuminating the GI tract during image capture. In this case, the light for distance measuring and the flash light for image capture will not be applied at the same time, or at least one of the light sources needs to be substantially dimmed.
  • the distance information can be stored separately or stored with an associated image. The distance information can be captured before or after an associated image. If the distance information is stored separately, the related information (named as association information in this disclosure) for the associated image will also be stored so that the distance information can be properly used. The related information can be capture time, frame time or frame number of the associated image. If the ultrasound is used to measure the distance, the distance measuring using ultrasound and image capturing by applying the flash light to illuminating the GI tract may occur at the same time.
  • one technique for capturing depth information is to use a color filter placed on top of selected sensor pixels with the passband reasonably narrow and capture the color information and depth information simultaneously.
  • the environment light sources with spectrum in the filter passband will cause negligible amount of energy projected onto the sensor.
  • a fourth type of pixels may be added to capture light with the spectrum in the passband of the filter placed on top of these pixels. Then, the structured light that has the spectrum substantially in the passband can be projected onto the scene.
  • this approach will reduce the spatial resolution of the images or video captured using such image sensor and require the use of an unconventional color filter.
  • Another technique is to obtain the depth information as well as 3D topology by projecting structured-light patterns that are visible in the RGB sensors.
  • the real time image and/or video will be confounded by the structured light superimposed on it.
  • the depth or shape information for objects in the scene is derived.
  • the depth or shape information is then assumed by the image or images captured shortly before or after the structured light image. Since the regular images are captured by a capsule endoscope at a relatively slow frame rate (e.g. 5 frames per second), the scene corresponds to the image captured using the structured light and the scene corresponds to a regular image may be noticeably different due to the endoscope movement or the peristaltic motion of the intestines.
  • the structured images with shortened frame period is disclosed in U.S. Patent Application, Serial No. 14/884,788, filed on October 16, 2015. Since the structured light image is closer in time with the corresponding regular image, the depth information derived should be more accurate than that derived based on a structured-light image with a longer frame period.
  • the depth (i.e., distance) information will be derived from the structured-light images.
  • the raw distance information is stored in the form of structured-light images.
  • the distance information i.e., the structured-light images
  • the distance information can be stored separately or stored with an associated image taken under regular light.
  • the distance information can be captured before or after an associated image. If the distance information is stored separately, the related information (i.e., association information) for the associated image will also be stored so that the distance information can be properly used.
  • Techniques to derive the depth information from the structured-light images are known in the field. Details of the depth derivation from the structured-light images are omitted in this disclosure.
  • the focal length is known by design. If the distance (also named as object distance in this disclosure) between an object and the camera can be determined, the dimensions of an object can be determined simply using geometry.
  • Fig. 1 illustrates a simplified example of object dimension determination based on object-camera distance.
  • the image sensor is placed at the focal plane 120 behind the lens 110. The camera can capture a scene within the field of view extending an angle a.
  • the focal length is the distance between the lens and the image sensor.
  • the focal length often is fixed for endoscopic applications and is known by design.
  • the object distance D varies depending on the location of the capsule endoscope and its relative angles with respect to the GI wall being imaged. If the distance D is known, the dimension of an object can be determined from the captured image by measuring the size of the object image in the image. For example, if an object 130 with height H is at distance D from the camera, the object image height H can be derived from the object image height h in the image according to:
  • h is measured from the image, the focal length /is known by design, and the distance / ) is determined by a selected distance measuring means as mentioned above. Accordingly, if the distance can be determined, the object dimensions can be derived.
  • the object size in the image can be measured in physical dimension. However, the image is captured digitally and the size measurement may be more convenient in terms of the number of pixels. Since the physical dimension of image sensor surface and the optical footprint are known. Also, the number of pixels is known (e.g. 320x240). Therefore, the object image size in the image can be measured in a number of pixels and converted physical object image size in the image.
  • the object image size in the image depends on the actual object size and its distance from the camera. A smaller object at a closer distance may appear to have the same size in the image as a larger object at a farther distance.
  • the object 140 which is smaller but closer than object 130, appears to have the same height as object 130 in the image. Therefore, the distance is crucial information for determining the object size.
  • the distance measuring means disclosed above enables object size determination based on the images captured using an endoscope.
  • Distance information is also useful for image stitching.
  • object size variations in the captured images may be implicitly taken care by the registration process.
  • a corresponding object in different images can be identified and registered.
  • the different object sizes in different images due to distance variations are presumably taken into consideration by the registration process.
  • the object having a difference size in a target frame will be matched with the corresponding object in the reference frame.
  • a global motion model will be applied to the target image to scale the object so that the image can be scaled and stitched properly.
  • the images associated with objects in the GI tract environment are usually far from the ideal solid object models.
  • the iteration process may not always converge or converge to local minima when variables such as distance are involved in the optimization process.
  • the iteration process is usually used as part of the whole registration process.
  • the distance information is used for scaling.
  • the distance information is used to assist the image registration to improve the registration accuracy.
  • FIG. 2 A and Fig. 2B illustrate an example of the different sizes of object images for a same object in two images captured at two different object distances.
  • illustration 210 corresponds to the case that the capsule 211 is at a farther distance from an object of interest 213 of the GI tract 212.
  • Image 220 corresponds to the image captured for illustration 210.
  • illustration 230 corresponds to the case that the capsule 211 is at a closer distance from an object of interest 213 of the GI tract 212.
  • Image 240 corresponds to the image captured for illustration 230. Since image 240 is capture with the camera closer to the GI wall. Therefore, the object image in image 240 appears larger than the object image size in image 220. Therefore, the distance information can be used to scale the object in these two images.
  • the image intensities for two images to be registered will be adjusted according to the distance.
  • the pixel intensities are roughly proportional to the distance square inversely or another functional form.
  • the relation between the pixel intensities and the distance can also be tabulated instead of being represented in a functional form.
  • the intensities of an image at a closer distance can be adjusted down to match with those of another image at a farther distance.
  • the intensities of an image at a farther distance can be adjusted up to match with those of another image at a closer distance.
  • Fig. 3 illustrates an exemplary flowchart for capturing an image sequence along with distance information according to an embodiment of the present invention.
  • the capsule device is administered to a patient in step 310.
  • An image sequence is captured using the camera in the capsule device when the capsule device travels through a human gastrointestinal tract in step 320.
  • the structured-light images are captured using the camera by projecting structured light to one or more objects in a field of view of the camera when the capsule device travels through the human gastrointestinal tract in step 330, where the structured-light images are interleaved with regular images in the image sequence.
  • the distance information is derived from the structured-light images in step 340, where the distance information is associated with objects of the image sequence with respect to the camera.
  • the image sequence is outputted in step 350.
  • the distance information is outputted in step 360.
  • the distance information extracted from structure light images implies that the respective distance information is determined at more than one point in the image or field of view. It's the intention of the present invention to include one or more points
  • Fig. 4 illustrates an exemplary flowchart for determining a size of an object of interest in the image sequence based on the image sequence and the distance information according to an embodiment of the present invention.
  • An image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract is received in step 410.
  • the distance information with respect to the capsule camera associated with objects of a selected image in the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract is received in step 420.
  • the size of an object of interest in the selected image is determined based on contents of the selected image and the distance information in step 430.
  • the size information regarding the size of the object of interest is outputted in step 440.
  • Fig. 5 illustrates an exemplary flowchart for stitching an image sequence utilizing information including the distance information to generate a stitched image sequence according to an embodiment of the present invention.
  • An image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract is received in step 510.
  • the distance information with respect to the capsule camera associated with objects of the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract is also received in step 520.
  • the image sequence is stitched utilizing information including the distance information to generate a stitched image sequence in step 530.
  • the stitched image sequence is outputted in step 540.

Abstract

A method and apparatus for capturing images of a scene using a capsule device including a camera are disclosed. An image sequence is captured using the camera when the capsule device travels through a human gastrointestinal tract. Also, structured-light images are captured using the camera by projecting structured light to one or more objects in a field of view of the camera when the capsule device travels through the human gastrointestinal tract. The structured-light images are interleaved with regular images in the image sequence. The distance information with respect to the capsule camera associated with objects of the selected image is derived. Both the image sequence and the distance information are outputted. A method of determining the size of an object of interest utilizing the distance information is also disclosed. In another method, the distance information is used to scale object or adjust intensities.

Description

TITLE: Method and Apparatus for Endoscope with Distance Measuring for Object Scaling
Inventors: Kang-Huai Wang, Chenyu Wu and Yi Xu CROSS REFERENCE TO RELATED APPLICATIONS
[0001] The present invention is a continuation-in-part and claims priority to PCT Patent Application, Serial No. PCT/US 16/52975, filed on September 22, 2016, which claims priority to U.S. Patent Application, Serial No. 14/884,788, filed on October 16, 2015. The PCT Patent Application and the U.S. Patent Application are hereby incorporated by reference in their entireties.
FIELD OF THE INVENTION
[0002] The present invention relates to the endoscope for capturing images of human gastrointestinal (GI) tract for diagnosis purpose. In particular, the endoscope is enabled to measure distance of objects in the field of view of the camera. The distance information can be used subsequently to process the image sequence captured, such as measuring a size of an object of interest or stitching the image sequence to reduce viewing time.
BACKGROUND AND RELATED ART
[0003] Devices for imaging body cavities or passages in vivo are known in the art and include endoscopes and autonomous encapsulated cameras. Endoscopes are flexible or rigid tubes that pass into the body through an orifice or surgical opening, typically into the esophagus via the mouth or into the colon via the rectum. An image is formed at the distal end using a lens and transmitted to the proximal end, outside the body, either by a lens-relay system or by a coherent fiber-optic bundle. A conceptually similar instrument might record an image electronically at the distal end, for example using a CCD or CMOS array, and transfer the image data as an electrical signal to the proximal end through a cable. Endoscopes allow a physician control over the field of view and are well-accepted diagnostic tools.
[0004] Capsule endoscope is an alternative in vivo endoscope developed in recent years. For capsule endoscope, a camera is housed in a swallowable capsule, along with a radio transmitter for transmitting data, primarily comprising images recorded by the digital camera, to a base-station receiver or transceiver and data recorder outside the body. The capsule may also include a radio receiver for receiving instructions or other data from a base-station transmitter. Instead of radio-frequency transmission, lower-frequency electromagnetic signals may be used. Power may be supplied inductively from an external inductor to an internal inductor within the capsule or from a battery within the capsule.
[0005] An autonomous capsule camera system with on-board data storage was disclosed in the US Patent No. 7,983,458, entitled "In Vivo Autonomous Camera with On-Board Data Storage or Digital Wireless Transmission in Regulatory Approved Band," granted on July 19, 2011. The capsule camera with on-board storage archives the captured images in on-board non-volatile memory. The capsule camera is retrieved upon its exiting from the human body. The images stored in the non-volatile memory of the retrieved capsule camera are then accessed through an output port on in the capsule camera.
[0006] When the endoscope is used for imaging the human GI tract, one of the primary purposes is to identify any possible anomaly. If any anomaly is found, it is further of interest to determine characteristics of the anomaly, such as the size of the anomaly. The captured images will be examined by medical profession for examination or diagnosis. The number of images captured is typically 25,000 or more. It will require a long reviewing time to look through the images even by skilled professionals. Accordingly, image stitching has been used to reduce the number of images to be viewed. For example, in PCT Patent Application Publication, Serial No. WO2014/193670 A2, published on Dec. 4, 2014, image stitching for images captured using a capsule camera is disclosed. It is desirable to develop methods or apparatus that are capable to further improve the efficiency of image stitching. BRIEF SUMMARY OF THE INVENTION
[0007] A method and apparatus for capturing images of a scene using a capsule camera are disclosed. After the capsule camera is swallowed by a patient, an image sequence is captured using the capsule camera when the capsule camera travels through a human gastrointestinal tract. Also, distance information associated with objects of the image sequence with respect to the capsule camera is captured when the capsule camera travels through the human gastrointestinal tract. Both the image sequence and the distance information are outputted.
[0008] The association information between the distance information and related images of the image sequence is outputted. The association information may correspond to frame numbers or capture times of the related images of the image sequence. The distance information can be determined based on Time-of-Flight or phase shift using a light or ultrasound source. The distance information may also be derived from structured-light images by projecting structured light to one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera.
[0009] The present invention also discloses a method of determining the size of an object of interest in an image. The image sequence captured by a capsule camera along with the distance information is received. The size of an object of interest in a selected image in the image sequence can be determined based on the selected image and the distance information. The size of the object of interest is determined according to the image size of the object of interest in a selected image scaled by a ratio of object distance to the capsule camera and focal length of the capsule camera. The image size of the object of interest in the selected image can be measured in terms of a number of pixels of object of interest in the selected image. The distance information can be in a form of structured-light images that are captured by projecting a structured light onto one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera, and the distance information is further derived from the structured-light images before the distance information is used for said determining the size information of the object of interest. [0010] The present invention further discloses a method of stitching the image sequence utilizing information including the distance information to generate a stitched image sequence. Again, the distance information can be in a form of structured-light images. In one
embodiment, the distance information is used to scale the objects of the image sequence for stitching the image sequence. In another embodiment, the distance information is used to adjust image intensities of the image sequence for stitching the image sequence.
[0011] BRIEF DESCRIPTION OF THE DRAWINGS
[0012] Fig. 1 illustrates an example of relationship among the size of an object, the size of the corresponding object image, the object distance and the focal length of the camera.
[0013] Fig. 2 A and Fig. 2B illustrate an example of the different sizes of object images for a same object in two images captured at two different object distances.
[0014] Fig. 3 illustrates an exemplary flowchart for capturing an image sequence along with distance information according to an embodiment of the present invention.
[0015] Fig. 4 illustrates an exemplary flowchart for determining a size of an object of interest in the image sequence based on the image sequence and the distance information according to an embodiment of the present invention.
[0016] Fig. 5 illustrates an exemplary flowchart for stitching an image sequence utilizing information including the distance information to generate a stitched image sequence according to an embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION
[0017] It will be readily understood that the components of the present invention, as generally described and illustrated in the figures herein, may be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of the embodiments of the systems and methods of the present invention, as represented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. References throughout this specification to "one embodiment," "an embodiment," or similar language mean that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the present invention. Thus, appearances of the phrases "in one embodiment" or "in an embodiment" in various places throughout this specification are not necessarily all referring to the same embodiment.
[0018] Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. One skilled in the relevant art will recognize, however, that the invention can be practiced without one or more of the specific details, or with other methods, components, etc. In other instances, well-known structures, or operations are not shown or described in detail to avoid obscuring aspects of the invention. The illustrated embodiments of the invention will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout. The following description is intended only by way of example, and simply illustrates certain selected embodiments of apparatus and methods that are consistent with the invention as claimed herein.
[0019] Endoscopes are normally inserted into the human body through a natural opening such as the mouth or anus. Therefore, endoscopes are preferred to be small sizes so as to be minimally invasive. As mentioned before, endoscopes can be used for diagnosis of human gastrointestinal (GI) tract. The captured image sequence can be viewed to identify any possible anomaly. If any anomaly is found, it is of interest to identify the characteristics of the anomaly, such as the size. Accordingly, an invention of the present invention discloses an endoscope including a distance measuring means to measure the object distances between the camera and various locations of an object in the field of view of the camera.
[0020] There are various known means for measuring the distance between the camera and various locations of an object in the field of view of the camera. For example, there is a class of distance measuring devices that determine the distance based on ToF (Time of Flight) or phase shift of a light source. The light source may be a laser or LED (Light Emitting Diode). A light sensor is used to detect the returned light. The time difference or phase difference between the emitted light from the light source and the received light by the light detector is used to determine the distance. Ultrasonic wave is also a signal source that can be used to measure the distance between an object and the camera for the intended GI imaging application. The distance measuring means is well-known in the field and various literatures describing the distance measure based on ToF or phase shift using light or ultrasound sources are readily available. Therefore, details for distance measuring means based on ToF or phase shift using light or ultrasound sources are omitted in this disclosure.
[0021] If a light source is used to measure the distance, the light for measuring the distance may interfere with the flash light illuminating the GI tract during image capture. In this case, the light for distance measuring and the flash light for image capture will not be applied at the same time, or at least one of the light sources needs to be substantially dimmed. The distance information can be stored separately or stored with an associated image. The distance information can be captured before or after an associated image. If the distance information is stored separately, the related information (named as association information in this disclosure) for the associated image will also be stored so that the distance information can be properly used. The related information can be capture time, frame time or frame number of the associated image. If the ultrasound is used to measure the distance, the distance measuring using ultrasound and image capturing by applying the flash light to illuminating the GI tract may occur at the same time.
[0022] While the distance measuring means based on ToF or phase shift using light or ultrasound is well known, to fit such distance measuring means into an endoscope is a challenging task and costly with larger size not suitable for endoscope application.
Accordingly, other distance measuring means are based on image processing of images captured using an image sensor. [0023] For example, one technique for capturing depth information is to use a color filter placed on top of selected sensor pixels with the passband reasonably narrow and capture the color information and depth information simultaneously. The environment light sources with spectrum in the filter passband will cause negligible amount of energy projected onto the sensor. For the case of RGB pixels, a fourth type of pixels may be added to capture light with the spectrum in the passband of the filter placed on top of these pixels. Then, the structured light that has the spectrum substantially in the passband can be projected onto the scene. However this approach will reduce the spatial resolution of the images or video captured using such image sensor and require the use of an unconventional color filter.
[0024] Another technique is to obtain the depth information as well as 3D topology by projecting structured-light patterns that are visible in the RGB sensors. However the real time image and/or video will be confounded by the structured light superimposed on it. When a structured image is captured, the depth or shape information for objects in the scene is derived. The depth or shape information is then assumed by the image or images captured shortly before or after the structured light image. Since the regular images are captured by a capsule endoscope at a relatively slow frame rate (e.g. 5 frames per second), the scene corresponds to the image captured using the structured light and the scene corresponds to a regular image may be noticeably different due to the endoscope movement or the peristaltic motion of the intestines. In order to improve the accuracy of the derived depth information from the structured light images, the structured images with shortened frame period is disclosed in U.S. Patent Application, Serial No. 14/884,788, filed on October 16, 2015. Since the structured light image is closer in time with the corresponding regular image, the depth information derived should be more accurate than that derived based on a structured-light image with a longer frame period.
[0025] When the distance measuring means via structured light is used, the depth (i.e., distance) information will be derived from the structured-light images. In other words, the raw distance information is stored in the form of structured-light images. In this case, the distance information (i.e., the structured-light images) can be stored separately or stored with an associated image taken under regular light. The distance information can be captured before or after an associated image. If the distance information is stored separately, the related information (i.e., association information) for the associated image will also be stored so that the distance information can be properly used. Techniques to derive the depth information from the structured-light images are known in the field. Details of the depth derivation from the structured-light images are omitted in this disclosure.
[0026] In an endoscope, the focal length is known by design. If the distance (also named as object distance in this disclosure) between an object and the camera can be determined, the dimensions of an object can be determined simply using geometry. Fig. 1 illustrates a simplified example of object dimension determination based on object-camera distance. In a camera system, the image sensor is placed at the focal plane 120 behind the lens 110. The camera can capture a scene within the field of view extending an angle a. The focal length /is the distance between the lens and the image sensor. The focal length often is fixed for endoscopic applications and is known by design. However, when a capsule endoscope travels through the GI tract, the object distance D varies depending on the location of the capsule endoscope and its relative angles with respect to the GI wall being imaged. If the distance D is known, the dimension of an object can be determined from the captured image by measuring the size of the object image in the image. For example, if an object 130 with height H is at distance D from the camera, the object image height H can be derived from the object image height h in the image according to:
Figure imgf000009_0001
[0027] In the above equation, h is measured from the image, the focal length /is known by design, and the distance /) is determined by a selected distance measuring means as mentioned above. Accordingly, if the distance can be determined, the object dimensions can be derived. The object size in the image can be measured in physical dimension. However, the image is captured digitally and the size measurement may be more convenient in terms of the number of pixels. Since the physical dimension of image sensor surface and the optical footprint are known. Also, the number of pixels is known (e.g. 320x240). Therefore, the object image size in the image can be measured in a number of pixels and converted physical object image size in the image.
[0028] As shown above, the object image size in the image depends on the actual object size and its distance from the camera. A smaller object at a closer distance may appear to have the same size in the image as a larger object at a farther distance. For example, the object 140, which is smaller but closer than object 130, appears to have the same height as object 130 in the image. Therefore, the distance is crucial information for determining the object size.
Accordingly, the distance measuring means disclosed above enables object size determination based on the images captured using an endoscope.
[0029] Distance information is also useful for image stitching. In an ideal situation with a solid object model, object size variations in the captured images may be implicitly taken care by the registration process. A corresponding object in different images can be identified and registered. The different object sizes in different images due to distance variations are presumably taken into consideration by the registration process. The object having a difference size in a target frame will be matched with the corresponding object in the reference frame. Likely, a global motion model will be applied to the target image to scale the object so that the image can be scaled and stitched properly. Nevertheless, the images associated with objects in the GI tract environment are usually far from the ideal solid object models. Furthermore, the iteration process may not always converge or converge to local minima when variables such as distance are involved in the optimization process. As is known in the field, the iteration process is usually used as part of the whole registration process. In one embodiment, the distance information is used for scaling. In particular, the distance information is used to assist the image registration to improve the registration accuracy.
[0030] Fig. 2 A and Fig. 2B illustrate an example of the different sizes of object images for a same object in two images captured at two different object distances. In Fig. 2A, illustration 210 corresponds to the case that the capsule 211 is at a farther distance from an object of interest 213 of the GI tract 212. Image 220 corresponds to the image captured for illustration 210. In Fig. 2B, illustration 230 corresponds to the case that the capsule 211 is at a closer distance from an object of interest 213 of the GI tract 212. Image 240 corresponds to the image captured for illustration 230. Since image 240 is capture with the camera closer to the GI wall. Therefore, the object image in image 240 appears larger than the object image size in image 220. Therefore, the distance information can be used to scale the object in these two images.
[0031] In the GI tract environment, images are always captured using a light source to illuminate the field of view. When the camera is closer to an intestine wall, the object being imaged will be brighter and the image intensities will be higher. On the other hand, when the camera is farther from an intestine wall, the object being imaged will be dimmer and the image intensities will be lower. Therefore, the overall intensities of an image will depend on the distance between the object in the field of view and the camera. Since the object-camera distance is typically very short for the GI tract environment, the variation in the overall intensities of an image will be rather large. Such large intensity variation may degrade the registration performance and consequently lower the stitching performance.
[0032] Accordingly, in another embodiment of the present invention, the image intensities for two images to be registered will be adjusted according to the distance. The pixel intensities are roughly proportional to the distance square inversely or another functional form. The relation between the pixel intensities and the distance can also be tabulated instead of being represented in a functional form. The intensities of an image at a closer distance can be adjusted down to match with those of another image at a farther distance. Alternatively, the intensities of an image at a farther distance can be adjusted up to match with those of another image at a closer distance. After intensity adjustment to compensate the variation due to different distances, the registration and image stitching should perform better.
[0033] Fig. 3 illustrates an exemplary flowchart for capturing an image sequence along with distance information according to an embodiment of the present invention. The capsule device is administered to a patient in step 310. An image sequence is captured using the camera in the capsule device when the capsule device travels through a human gastrointestinal tract in step 320. The structured-light images are captured using the camera by projecting structured light to one or more objects in a field of view of the camera when the capsule device travels through the human gastrointestinal tract in step 330, where the structured-light images are interleaved with regular images in the image sequence. The distance information is derived from the structured-light images in step 340, where the distance information is associated with objects of the image sequence with respect to the camera. The image sequence is outputted in step 350. The distance information is outputted in step 360. The distance information extracted from structure light images implies that the respective distance information is determined at more than one point in the image or field of view. It's the intention of the present invention to include one or more points in image.
[0034] Fig. 4 illustrates an exemplary flowchart for determining a size of an object of interest in the image sequence based on the image sequence and the distance information according to an embodiment of the present invention. An image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract is received in step 410. The distance information with respect to the capsule camera associated with objects of a selected image in the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract is received in step 420. The size of an object of interest in the selected image is determined based on contents of the selected image and the distance information in step 430. The size information regarding the size of the object of interest is outputted in step 440.
[0035] Fig. 5 illustrates an exemplary flowchart for stitching an image sequence utilizing information including the distance information to generate a stitched image sequence according to an embodiment of the present invention. An image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract is received in step 510. The distance information with respect to the capsule camera associated with objects of the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract is also received in step 520. The image sequence is stitched utilizing information including the distance information to generate a stitched image sequence in step 530. The stitched image sequence is outputted in step 540.
[0036] The above description is presented to enable a person of ordinary skill in the art to practice the present invention as provided in the context of a particular application and its requirements. Various modifications to the described embodiments will be apparent to those with skill in the art, and the general principles defined herein may be applied to other embodiments. Therefore, the present invention is not intended to be limited to the particular embodiments shown and described, but is to be accorded the widest scope consistent with the principles and novel features herein disclosed. In the above detailed description, various specific details are illustrated in order to provide a thorough understanding of the present invention. Nevertheless, it will be understood by those skilled in the art that the present invention may be practiced.
[0037] The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims

1. A method of capturing images of a scene using a capsule device including a camera, the method comprising: administering the capsule device to a patient;
capturing an image sequence using the camera when the capsule device travels through a human gastrointestinal tract;
capturing structured-light images using the camera by projecting structured light to one or more objects in a field of view of the camera when the capsule device travels through the human gastrointestinal tract, wherein the structured-light images are interleaved with regular images in the image sequence;
deriving distance information from the structured-light images, wherein the distance information is associated with objects of the image sequence with respect to the camera;
outputting the image sequence; and
outputting the distance information.
2. The method of Claim 1 , wherein association information between the distance information and related images of the image sequence is outputted.
3. The method of Claim 2, wherein the association information corresponds to frame
numbers or capture times of the related images of the image sequence.
4. A capsule device for capturing images of a scene using a capsule device, the capsule device comprising: a camera;
a structured light;
one or more processors coupled to the camera and the structured light; one or more output interfaces coupled to one or more processors; and a housing adapted to be swallowed, wherein the housing encloses the camera and said one or more processors in a sealed environment;
wherein said one or more processors are configured to:
capture an image sequence using the camera when the capsule device travels through a human gastrointestinal tract;
capture structured-light images using the camera by projecting structured light to one or more objects in a field of view of the camera when the capsule device travels through the human gastrointestinal tract, wherein the structured-light images are interleaved with regular images of the image sequence;
deriving distance information from the structured-light images, wherein the distance information is associated with objects of the image sequence with respect to the camera;
output the image sequence through said one or more output interfaces; and output the distance information.
5. The capsule device of Claim 4, wherein said one or more output interfaces provide association information between the distance information and related images of the image sequence is outputted.
6. The capsule device of Claim 5, wherein the association information corresponds to frame numbers or capture times of the related images of the image sequence.
7. A method of processing images captured using a capsule camera, the method comprising: receiving an image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract;
receiving distance information with respect to the capsule camera associated with objects of a selected image in the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract;
determining a size of an object of interest in the selected image based on contents of the selected image and the distance information; and outputting size information regarding the size of the object of interest.
8. The method of Claim 7, wherein the size of the object of interest is determined according to an image size of the object of interest in the selected image scaled by a ratio of object distance to the capsule camera and focal length of the capsule camera.
9. The method of Claim 8, wherein the image size of the object of interest is measured in terms of a number of pixels of object of interest in the selected image.
10. The method of Claim 7, wherein the distance information is in a form of a
structured-light images that are captured by projecting structured light onto one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera, and the distance information is further derived from the structured-light images before the distance information is used for said determining the size information of the object of interest.
11. A method of processing images captured using a capsule camera, the method comprising: receiving an image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract;
receiving distance information with respect to the capsule camera associated with objects of the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract;
stitching the image sequence utilizing information including the distance information to generate a stitched image sequence; and
outputting the stitched image sequence.
12. The method of Claim 11, wherein the distance information is in a form of a structured-light images that are captured by projecting structured light onto one or more objects in a field of view of the capsule camera and capturing the structured-light images using the capsule camera, and the distance information is further derived from the structured-light images before the distance information is used for said stitching the image sequence.
13. The method of Claim 12, wherein the distance information is used to scale the objects of the image sequence for said stitching the image sequence.
14. The method of Claim 12, wherein the distance information is used to adjust image intensities of the image sequence for said stitching the image sequence.
15. The method of Claim 11, wherein association information between the distance information and related images of the image sequence is also received and used for said stitching the image sequence utilizing information.
16. An apparatus for processing images captured using a capsule camera, the apparatus comprising one or more electronic circuits or processors arranged to: receive an image sequence captured by the capsule camera when the capsule camera travelled through a human gastrointestinal tract;
receive distance information with respect to the capsule camera associated with objects of the image sequence captured by the capsule camera when the capsule camera travelled through the human gastrointestinal tract;
stitch the image sequence utilizing information including the distance information to generate a stitched image sequence; and
output the stitched image sequence.
PCT/US2017/015668 2015-10-16 2017-01-30 Method and apparatus for endoscope with distance measuring for object scaling WO2018140062A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
PCT/US2017/015668 WO2018140062A1 (en) 2017-01-30 2017-01-30 Method and apparatus for endoscope with distance measuring for object scaling
CN201810091804.5A CN108392165A (en) 2017-01-30 2018-01-30 Method and utensil for the introscope with the range measurement scaled for object
JP2018013654A JP2018130537A (en) 2017-01-30 2018-01-30 Method and apparatus used for endoscope with distance measuring function for object scaling
US15/883,337 US10402992B2 (en) 2015-10-16 2018-01-30 Method and apparatus for endoscope with distance measuring for object scaling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2017/015668 WO2018140062A1 (en) 2017-01-30 2017-01-30 Method and apparatus for endoscope with distance measuring for object scaling

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/884,788 Continuation-In-Part US9936151B2 (en) 2015-10-16 2015-10-16 Single image sensor for capturing mixed structured-light images and regular images

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/883,337 Continuation US10402992B2 (en) 2015-10-16 2018-01-30 Method and apparatus for endoscope with distance measuring for object scaling

Publications (1)

Publication Number Publication Date
WO2018140062A1 true WO2018140062A1 (en) 2018-08-02

Family

ID=62978402

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2017/015668 WO2018140062A1 (en) 2015-10-16 2017-01-30 Method and apparatus for endoscope with distance measuring for object scaling

Country Status (3)

Country Link
JP (1) JP2018130537A (en)
CN (1) CN108392165A (en)
WO (1) WO2018140062A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220318559A1 (en) * 2021-03-31 2022-10-06 Nvidia Corporation Generation of bounding boxes

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109730683B (en) * 2018-12-21 2021-11-05 重庆金山医疗技术研究院有限公司 Endoscope target size calculation method and analysis system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130002842A1 (en) * 2011-04-26 2013-01-03 Ikona Medical Corporation Systems and Methods for Motion and Distance Measurement in Gastrointestinal Endoscopy
US8617058B2 (en) * 2008-07-09 2013-12-31 Innurvation, Inc. Displaying image data from a scanner capsule
US20150119643A1 (en) * 2008-06-09 2015-04-30 Capso Vision, Inc. In Vivo CAMERA WITH MULTIPLE SOURCES TO ILLUMINATE TISSUE AT DIFFERENT DISTANCES

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4631057B2 (en) * 2004-02-18 2011-02-16 国立大学法人大阪大学 Endoscope system
JP4767685B2 (en) * 2005-12-28 2011-09-07 オリンパスメディカルシステムズ株式会社 In-subject observation system
CN102063714A (en) * 2010-12-23 2011-05-18 南方医科大学 Method for generating body cavity full-view image based on capsule endoscope images
JP2013013481A (en) * 2011-07-01 2013-01-24 Panasonic Corp Image acquisition device and integrated circuit
JP2013063179A (en) * 2011-09-16 2013-04-11 Olympus Medical Systems Corp Observation system
KR101390190B1 (en) * 2012-10-11 2014-04-29 삼성전자주식회사 X-ray photographing apparatus and method for using the same and x-ray image obtaining method
JP2014161355A (en) * 2013-02-21 2014-09-08 Olympus Corp Image processor, endoscope device, image processing method and program
CN103815858A (en) * 2014-02-26 2014-05-28 上海齐正微电子有限公司 Capsular endoscope with multiple built-in sensors
CN107072498B (en) * 2015-06-30 2019-08-20 奥林巴斯株式会社 Image processing apparatus, capsule-type endoscope system and endoscopic system
CN105996961B (en) * 2016-04-27 2018-05-11 安翰光电技术(武汉)有限公司 3D three-dimensional imagings capsule endoscope system and method based on structure light

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150119643A1 (en) * 2008-06-09 2015-04-30 Capso Vision, Inc. In Vivo CAMERA WITH MULTIPLE SOURCES TO ILLUMINATE TISSUE AT DIFFERENT DISTANCES
US8617058B2 (en) * 2008-07-09 2013-12-31 Innurvation, Inc. Displaying image data from a scanner capsule
US20130002842A1 (en) * 2011-04-26 2013-01-03 Ikona Medical Corporation Systems and Methods for Motion and Distance Measurement in Gastrointestinal Endoscopy

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220318559A1 (en) * 2021-03-31 2022-10-06 Nvidia Corporation Generation of bounding boxes

Also Published As

Publication number Publication date
JP2018130537A (en) 2018-08-23
CN108392165A (en) 2018-08-14

Similar Documents

Publication Publication Date Title
US10402992B2 (en) Method and apparatus for endoscope with distance measuring for object scaling
US11503991B2 (en) Full-field three-dimensional surface measurement
US11529197B2 (en) Device and method for tracking the position of an endoscope within a patient's body
JP6586211B2 (en) Projection mapping device
US10736559B2 (en) Method and apparatus for estimating area or volume of object of interest from gastrointestinal images
CN102247114B (en) Image processing apparatus and image processing method
US10346978B2 (en) Method and apparatus for area or volume of object of interest from gastrointestinal images
CN111035351B (en) Method and apparatus for travel distance measurement of capsule camera in gastrointestinal tract
Moglia et al. Recent patents on wireless capsule endoscopy
US20160073854A1 (en) Systems and methods using spatial sensor data in full-field three-dimensional surface measurement
US10580157B2 (en) Method and apparatus for estimating area or volume of object of interest from gastrointestinal images
WO2014136579A1 (en) Endoscope system and endoscope system operation method
JP6064106B1 (en) Image processing apparatus, capsule endoscope system, and endoscope system
US20080027329A1 (en) System, apparatus and method for measurement of motion parameters of an in-vivo device
JP5750669B2 (en) Endoscope system
JP2009521978A (en) System, apparatus, and method for estimating the size of a body lumen object
JP2009273893A (en) Device, system and method for magnetically maneuvering in-vivo device
US20020107444A1 (en) Image based size analysis
JP2011234871A (en) Endoscope system
WO2018140062A1 (en) Method and apparatus for endoscope with distance measuring for object scaling
EP3173010B1 (en) Passive capsule type endoscope for the intestine
US20230410336A1 (en) Method and Apparatus for Identifying Capsule Camera Location inside Gastrointestinal Tract
JP2003210390A (en) Endoscopic apparatus
KR20170048987A (en) Wireless capsule endoscopy with variable frame rate and method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17894078

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17894078

Country of ref document: EP

Kind code of ref document: A1