WO2004112593A1 - カプセル型内視鏡及びカプセル型内視鏡システム - Google Patents
カプセル型内視鏡及びカプセル型内視鏡システム Download PDFInfo
- Publication number
- WO2004112593A1 WO2004112593A1 PCT/JP2004/009267 JP2004009267W WO2004112593A1 WO 2004112593 A1 WO2004112593 A1 WO 2004112593A1 JP 2004009267 W JP2004009267 W JP 2004009267W WO 2004112593 A1 WO2004112593 A1 WO 2004112593A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal processing
- capsule endoscope
- data
- white balance
- processing data
- Prior art date
Links
- 238000012545 processing Methods 0.000 claims abstract description 163
- 230000005540 biological transmission Effects 0.000 claims abstract description 98
- 238000003384 imaging method Methods 0.000 claims abstract description 60
- 230000007547 defect Effects 0.000 claims abstract description 42
- 238000006243 chemical reaction Methods 0.000 claims abstract description 23
- 239000002775 capsule Substances 0.000 claims description 167
- 238000012937 correction Methods 0.000 claims description 45
- 238000000034 method Methods 0.000 claims description 25
- 230000008569 process Effects 0.000 claims description 9
- 230000015654 memory Effects 0.000 description 57
- 238000001514 detection method Methods 0.000 description 34
- 238000010586 diagram Methods 0.000 description 25
- 230000000875 corresponding effect Effects 0.000 description 13
- 238000005286 illumination Methods 0.000 description 12
- 238000000926 separation method Methods 0.000 description 10
- 238000004519 manufacturing process Methods 0.000 description 8
- 238000012986 modification Methods 0.000 description 8
- 230000004048 modification Effects 0.000 description 8
- 238000004891 communication Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 7
- 230000006835 compression Effects 0.000 description 6
- 238000007906 compression Methods 0.000 description 6
- 230000004044 response Effects 0.000 description 6
- 238000005070 sampling Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 238000007689 inspection Methods 0.000 description 4
- 230000002159 abnormal effect Effects 0.000 description 3
- 230000002596 correlated effect Effects 0.000 description 3
- 230000001360 synchronised effect Effects 0.000 description 3
- 238000012790 confirmation Methods 0.000 description 2
- 210000000056 organ Anatomy 0.000 description 2
- 230000009747 swallowing Effects 0.000 description 2
- 235000011293 Brassica napus Nutrition 0.000 description 1
- 240000008100 Brassica rapa Species 0.000 description 1
- 235000000540 Brassica rapa subsp rapa Nutrition 0.000 description 1
- 241000167880 Hirundinidae Species 0.000 description 1
- BQCADISMDOOEFD-UHFFFAOYSA-N Silver Chemical compound [Ag] BQCADISMDOOEFD-UHFFFAOYSA-N 0.000 description 1
- 230000005856 abnormality Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 230000035622 drinking Effects 0.000 description 1
- 210000003238 esophagus Anatomy 0.000 description 1
- 210000001035 gastrointestinal tract Anatomy 0.000 description 1
- 210000004798 organs belonging to the digestive system Anatomy 0.000 description 1
- 244000052769 pathogen Species 0.000 description 1
- 230000001717 pathogenic effect Effects 0.000 description 1
- 230000008855 peristalsis Effects 0.000 description 1
- 238000010248 power generation Methods 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 238000007789 sealing Methods 0.000 description 1
- 229910052709 silver Inorganic materials 0.000 description 1
- 239000004332 silver Substances 0.000 description 1
- 210000002784 stomach Anatomy 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/041—Capsule endoscopes for imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00011—Operational features of endoscopes characterised by signal transmission
- A61B1/00016—Operational features of endoscopes characterised by signal transmission using wireless means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0031—Implanted circuitry
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/51—Housings
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/54—Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/66—Remote control of cameras or camera parts, e.g. by remote control devices
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7232—Signal processing specially adapted for physiological signals or for diagnostic purposes involving compression of the physiological signal, e.g. to extend the signal recording period
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L1/00—Arrangements for detecting or preventing errors in the information received
- H04L1/004—Arrangements for detecting or preventing errors in the information received by using forward error control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/555—Constructional details for picking-up images in sites, inaccessible due to their dimensions or hazardous conditions, e.g. endoscopes or borescopes
Definitions
- the present invention relates to a capsule endoscope and a capsule endoscope system.
- Capsule endoscopes that can be inserted into the body cavity ⁇ ⁇ from the mouth and can capture information such as the inside of the body cavity by photographing the digestive organs such as the stomach (medical swallow-type tablet endoscopes) It has been known.
- an illumination means such as an LED, a solid-state image sensor such as a CCD or CMOS, and a power supply unit such as a battery for driving the illumination means and the solid-state image sensor are used.
- a power supply unit such as a battery for driving the illumination means and the solid-state image sensor
- Japanese Patent Application Laid-Open Publication No. 2000-244584 discloses a technique of a forceps endoscope equipped with a white balance function.
- the publication describes that the capsule endoscope has an image sensor, its scanning circuit, and a signal processing circuit integrated on the same chip, and the signal processing circuit has an auto white balance function.
- the processing such as white balance is applied to the signal processing circuit inside the capsule endoscope. In such a case, the circuit scale of the internal circuit becomes large, and there is a problem that the current consumption is increased.
- the power supply method of the capsule endoscope As for the power supply method of the capsule endoscope, a method using a battery and a method to supply power from outside the body by radio are proposed, but in both cases, the capsule endoscope is used. The same problem occurred when processing such as white balance was performed by the signal processing circuit inside the mirror.
- An object of the present invention is to reduce the power consumption without increasing the circuit scale of the internal circuit.
- the present invention provides a storage unit that stores signal processing data necessary for signal processing unique to an imaging element of a capsule endoscope, and a storage unit that stores the data for signal processing. And a transmission unit for transmitting the signal processing data.
- the present invention is characterized in that, in the capsule endoscope described above, the signal processing data is a value obtained in advance before the capsule endoscope is shipped.
- the data for signal processing is data of a white balance coefficient used when performing white balance processing of the image sensor.
- the data for signal processing is data of an image of the chart for color signal processing captured by the image sensor.
- the data for signal processing is data indicating an address of a pixel defect of the imaging element.
- the signal processing data is data indicating an offset value of a photoelectric conversion characteristic of the imaging device.
- the transmitting unit transmits the signal processing data together with image data captured by the image sensor.
- the transmitting unit includes at least a part of the signal processing data in each of frames that are transmission units when transmitting the image data. Transmission.
- the signal processing data is added to a rear end side of the frame.
- the signal processing data is added to a last end of the frame.
- the present invention is characterized in that, in the capsule endoscope described above, the transmitting unit transmits the signal processing data together with an error correction code of the signal processing data.
- the error correction code is obtained in advance before the capsule endoscope is shipped, and the data of the error correction code is stored in the storage unit. It is stored.
- the present invention provides a storage unit for storing signal processing data necessary for signal processing unique to an imaging device of a capsule endoscope, and a transmission for transmitting the signal processing data stored in the storage unit.
- a capsule endoscope system comprising: a capsule endoscope having a unit; and a receiver for receiving the signal processing data transmitted from the transmission unit. Is characterized in that the receiver performs signal processing unique to the image sensor based on the received signal processing data without performing signal processing unique to the image sensor.
- the present invention is characterized in that, in the capsule endoscope system described above, the data for signal processing is a value obtained in advance before the capsule endoscope is shipped.
- the self-signal processing data is data of a white balance coefficient used when performing white balance processing of the imaging device.
- the present invention provides the forceps / ray endoscope system described above, wherein the signal processing data is data of an image obtained by capturing a chart for color signal processing by the imaging device. And
- the data for signal processing is data indicating an address of a pixel defect of the image sensor.
- the signal processing data is data indicating an offset value of a photoelectric conversion characteristic of the imaging device.
- the transmission unit transmits the signal processing data together with image data captured by the image sensor.
- the transmission unit transmits at least a part of the signal processing data in each of frames that are transmission units when transmitting the imaging data. It is characterized by doing.
- the signal processing data is added to a rear end side of the frame.
- the signal processing data is added to a last end of the frame.
- the transmitting unit transmits the signal processing data together with an error correction code of the signal processing data.
- the error correction code is obtained in advance before the capsule endoscope is shipped, and the data of the error correction code is stored in the storage unit. Is stored.
- FIG. 1 is a side sectional view showing a capsule endoscope according to one embodiment of the present invention.
- FIG. 2 is a block diagram showing a capsule endoscope system according to one embodiment of the present invention.
- You. FIG. 3 is a block diagram showing a configuration of a capsule endoscope according to one embodiment of the present invention.
- FIG. 4 is a block diagram showing a configuration of a receiver according to one embodiment of the present invention.
- FIG. 5 is a block diagram showing a configuration of an image processing unit of the capsule endoscope according to one embodiment of the present invention.
- FIG. 6 is a block diagram showing a configuration of an image processing unit of the receiver according to one embodiment of the present invention.
- FIG. 1 is a side sectional view showing a capsule endoscope according to one embodiment of the present invention.
- FIG. 2 is a block diagram showing a capsule endoscope system according to one embodiment of the present invention.
- You. is a block diagram showing a configuration of a capsule endoscope according to
- FIG. 7 is a flowchart showing a procedure for obtaining a white balance coefficient in the capsule endoscope according to one embodiment of the present invention.
- FIG. 8 is a diagram showing a configuration of a transmission unit of transmission data transmitted from the force capsule endoscope according to one embodiment of the present invention.
- FIG. 9 is a flowchart showing the operation of the capsule endoscope system according to one embodiment of the present invention.
- FIG. 10 is a diagram showing another configuration of a transmission unit of transmission data transmitted from the force capsule endoscope according to one embodiment of the present invention.
- FIG. 11 is a flowchart illustrating a procedure of white balance processing performed by the receiver according to the embodiment of the present invention.
- FIG. 12 is a diagram showing still another configuration of a transmission unit of transmission data transmitted from the capsule endoscope according to one embodiment of the present invention.
- FIG. 13 is a flowchart showing a procedure for calculating an address of a pixel defect in a capsule endoscope according to another embodiment of the present invention.
- FIG. 14 is a diagram showing a transmission unit of transmission data including pixel defect address data in a capsenole type endoscope according to another embodiment of the present invention.
- FIG. 15 is a diagram for explaining a method of obtaining an offset value of a photoelectric conversion characteristic of a CMOS image sensor in a turnip endoscope according to another embodiment of the present invention.
- FIG. 16 is a diagram showing an example in which a white balance coefficient is added to the rear end of the image data.
- FIG. 17 is a block diagram showing a configuration of an image processing unit of a force-pressed endoscope according to yet another embodiment of the present invention.
- FIG. 18 is a block diagram showing a configuration of an image processing unit of a receiver according to still another embodiment of the present invention.
- FIG. 19 is a waveform diagram showing an output signal from the multiplexing unit shown in FIG.
- FIG. 20 is a waveform chart showing another example of the output signal from the multiplexing unit in FIG.
- FIG. 21 (a) is a waveform diagram showing still another example of the output signal from the multiplexing unit of FIG. 17, and
- FIG. 21 (b) is a multiplexing unit of FIG. FIG.
- FIG. 9 is a waveform chart showing still another example of the output signal from the controller.
- FIG. 22 shows the update of the output signal from the multiplex section shown in FIG.
- FIG. 7 is a waveform diagram showing another example.
- FIG. 23 is a block diagram showing a configuration of an image processing unit of a capsule endoscope according to still another embodiment of the present invention.
- FIG. 24 is a diagram showing an output signal from the multiplexing unit in FIG.
- FIG. 25 is a diagram showing an example in which a white balance coefficient is added to the rear end of a series of video signals.
- FIG. 26 is a block diagram showing a configuration of a capsule endoscope according to still another embodiment of the present invention.
- FIG. 27 is a block diagram showing a configuration of a receiver according to still another embodiment of the present invention.
- FIG. 1 is a schematic diagram showing the internal structure of a capsule endoscope according to the present embodiment.
- the capsenolle endoscope 10 includes an imaging unit 111 that can capture an image of the inside of a body cavity, and illumination units 1 12 a and 1 1 2 that irradiate light inside the body cavity.
- b a power supply unit 113 for supplying power to them
- a capsule housing in which at least an imaging unit 111, illumination units 112 a, 112, and b and a power supply unit 113 are disposed. It consists of body 14.
- the force capsule housing 14 is composed of a tip cover part 120 covering the imaging part 111 and the lighting parts 112a, 112b, a tip force par part 120, and a sealing member. It is provided in a watertight state via 1 2 1, and acts as a capsule body 1 2 2 with an imaging unit 1 1 1 etc. disposed inside, and a rear end cover 1 2 3 as necessary. It may be provided separately from the capsule body 122.
- the rear end force par section 123 is provided integrally with the capsule body 122 and has a flat shape.
- the shape is not limited.
- the shape may be a dome shape.
- the tip force par section 120 has an illumination window section 120a for transmitting the illumination light L from the illumination sections 112a and 112b, and an imaging window for imaging the illumination range. Part 1 2 0 b and May be configured to be clearly divided.
- the distal end cover 120 is entirely transparent, and the areas of the illumination window 120a and the imaging window 120b partially overlap.
- the imaging unit 111 is provided on an imaging board 124, and captures an image of a range illuminated by the illumination light L from the illumination units 112a and 112b.
- a sharp image is formed by the focus adjustment unit 128 using b.
- the imaging unit 111 is not limited to the above-described CCD, and an imaging unit such as CMOs can be used, for example. .
- the illuminating units 112a and 112b are provided on an illuminating substrate 130, and include, for example, light-emitting diodes (LEDs).
- the illuminating units 112a and 112b include an imaging lens 126 that constitutes the imaging unit 111. It is located at the center, and there are multiple units around it. In the present embodiment, as an example, a total of four illuminating units are arranged one above, below, left, and right with the imaging lens 126 as the center.
- the lighting units 112a and 112b are not limited to the LEDs described above, and other lighting means can be used.
- the power supply unit 113 is provided on a power supply board 132 provided with an internal switch 131.
- a button-type battery is used as the power supply 133.
- the power source 133 uses, for example, an Oishidani silver battery as the battery, but is not limited to this.
- a rechargeable battery, a power generation battery, or the like may be used.
- the internal switch 131 is provided to prevent an unnecessary current from flowing from the power source 133 before using the capsule endoscope.
- a wireless unit 142 for performing wireless communication with the outside is provided on the wireless board 141, and communication with the outside is performed via the wireless unit 142 as needed.
- the fountain part 142 is modulated by the modulator 211 as shown in FIGS.
- a transmission section 142a for amplifying the modulated modulation signal and an antenna 142b are provided.
- a control unit 143 is provided on the imaging board 124, and executes various processes in the capsule endoscope 10.
- the signal processing / control unit 143 includes an image processing unit 143a, a control unit 143b, a driving unit 143c, and a modulation unit 211.
- the image processing unit 143a controls an image signal processing function such as image data generation including, for example, correlated double sampling (usually including CDS), and power supply according to ON / OFF of the internal switch 131. It has a power supply control function. Further, the image processing unit 143a includes a parameter memory 208 for storing parameters such as a line 'frame and a parameter such as a white balance coefficient, and a multiplexing unit 209 for multiplexing a white balance coefficient and a video signal. Have.
- the control unit 143b has a timing generator and a sync generator 201 for generating various timing signals or synchronization signals.
- the control unit 143b controls the image processing unit 143a, the driving unit 143c, and the lighting units 112a and 112b based on the timing signal or the synchronization signal generated by the timing generator and the sync generator 201. I do.
- the lighting units 112a and 112b emit light at a predetermined timing in response to a timing signal or a synchronization signal from the control unit 143b.
- the driving unit 143c drives the CCD 125 based on a timing signal or a synchronization signal from the control unit 143b.
- the control unit 143b controls the timing at which the CCD 125 is driven and the timing at which the illumination units 112a and 112b emit light in synchronization, and controls the number of images captured by the CCD 125.
- the modulation section 211 has a modulation function of converting into, for example, PSK, MSK, GMSK, QMSK, ASK, AM, and FM systems, and outputs a modulated signal to the transmission section 142a.
- FIG. 2 is a schematic diagram of a capsule endoscope system according to the present embodiment. Inspection using the capsule endoscope 10 is performed using a forcepsell type endoscope system as shown in FIG.
- a capsule endoscope system 1 includes a forcepsell type endoscope 10 and a package 50 thereof, a jacket 3 worn by a patient, that is, a subject 2, It consists of a receiver 4 and a computer 5 that are detachable from the jacket 3.
- the jacket 3 is provided with antennas 31, 32, 33, and 34 for capturing radio waves transmitted from the antennas 14 2 b of the capsule endoscope 10. Communication between the forcepsell type endoscope 10 and the receiver 4 is possible via 33 and 34.
- the number of the antennas is not particularly limited to four, and it is sufficient if there are a plurality of the antennas, whereby the radio waves corresponding to the position accompanying the movement of the capsule endoscope 10 can be satisfactorily received.
- the position of the capsule endoscope 10 in the body cavity can be detected based on the reception intensity of each of the antennas 31, 32, 33, and 34.
- the receiver 4 includes a reception unit 41, a demodulation unit 301, an image processing unit (5300, an image compression unit 303, a card interface 300a). Is provided.
- the receiving section 41 amplifies the radio signal captured by the antennas 3!: 3 4 and outputs the signal to the demodulating section 301.
- the demodulation unit 301 demodulates the output from the reception unit 41.
- the image processing unit 300 performs signal separation on the signal demodulated by the demodulation unit 301, and detects a parameter such as a white balance coefficient based on the result of the signal separation.
- the image processing unit 300 performs a white balance process on the image data using the detected white balance coefficient.
- the image compression unit 303 compresses the image data that has been subjected to the white balance processing in the image processing unit 300.
- the card interface 303 a has a function as an interface for inputting and outputting image data between the CF memory card 44 as a large-capacity memory and the image compression unit 310.
- the receiver 4 detachably mounts the CF memory card 44.
- the CF memory card 44 stores the image data compressed by the image compression unit 310.
- the receiver 4 is provided with a display unit (not shown) for displaying information necessary for observation (inspection) and an input unit (not shown) for inputting information necessary for observation (inspection). ing.
- the computer 5 performs read-write of the CF memory card 44 and the like.
- the computer 5 has a processing function for a doctor or a nurse (examiner) to make a diagnosis based on an image of an organ or the like in a patient's body captured by the capsule endoscope 10.
- the capsule endoscope 10 is taken out of the package 50 before starting the inspection.
- the internal switch 13 1 of the capsule endoscope 10 becomes ON.
- the subject 2 swallows the capsule endoscope 10 in which the internal switch 13 1 is turned on from the mouth.
- the capsule endoscope 10 passes through the esophagus, advances in the body cavity by peristalsis of the digestive tract lumen, and sequentially captures an image of the body cavity.
- the radio waves of the captured image are output via the radio section 142 with respect to the imaging results, and the radio waves are transmitted by the antennas 3 1, 3 2, 3 3, 3 4 of the jacket 3. Is captured.
- the signal of the captured radio wave can be relayed from the antenna 31, 32, 33, or 34 to the receiver 4. At this time, depending on the position of the capsule endoscope 10, a difference in the received radio wave intensity occurs between the antennas 31, 33, 33, and 34.
- white balance processing is performed on the sequentially received captured image data, and the image data that has been subjected to white balance processing is stored in the CF memory card 44.
- the reception by the receiver 4 is performed when the imaging of the capsule endoscope 10 is started. It is not synchronized with the start, and the start and end of reception are controlled by operating the input unit of the receiver 4.
- the CF memory card 44 storing the captured image data is removed from the receiver 4 and inserted into the memory card slot of the computer 5. Put in. In the computer 5, the captured image data stored in the CF memory card 44 is read, and the captured image data is stored for each patient.
- the image processing unit 143a of the capsule endoscope 10 converts the analog surface image data output from the CCD 125 into a digital signal (digital transmission) and outputs the digital signal to the modulation unit 211.
- the image processing unit 143 a includes a CDS (Correlated Double Sampling) unit 203, an AMP unit 204, an AZD unit 205, a parameter memory 208, and a multiplexing unit 209. I have.
- the timing generator and the sync generator 201 supply the CCD 125 with a pulse signal 202 for driving the CCD 125 at a predetermined timing.
- the pulse (TG) signal 2.02 is a signal 202 serving as a reference for timing of an imaging system such as the CCD 125.
- the charge is converted from the CCD 125 and read out.
- the signal read from the CCD 125 is subjected to noise removal processing by correlated double sampling in the CDS unit 203, thereby generating image data.
- the image data is amplified by the AMP unit 204, AD-converted by the AZD unit 205, and sent to the multiplexing unit 209.
- the parameter memory 208 stores a white balance coefficient for correcting white balance.
- Each capsule endoscope 10 is tested in the manufacturing process, and a unique white balance coefficient is obtained for the capsenolle endoscope 10 (the method for obtaining the white balance coefficient will be described later).
- the white balance coefficient of each capsule type The white balance coefficients are written to the parameter memory 208 of the endoscope 10 and are shipped with the white balance coefficients stored in the parameter memory 208 of the capsule endoscope 10 at the time of shipment.
- the timing (SG) signal 210 is a signal serving as a reference for the timing of a display system constituting an image.
- the read white balance coefficient is superimposed (multiplexed) by the multiplexing unit 209 on the image signal output from the A / D unit 205, and then modulated by the modulation unit 211. Is done. As shown in FIG. 3, the modulation signal output from the modulation section 211 is transmitted to the outside of the forceps endoscope 10 via the radio section 142.
- FIG. 6 shows the configuration of the image processing unit 300 of the receiver 4 for digital transmission.
- the image processing unit 300 includes a signal separation unit 302, an image memory 303, a parameter detection unit 304, and an image signal processing unit 300. "
- the radio waves transmitted from the radio section 14 of the force capsule endoscope 10 are captured by the antennas 31 to 34, and the signals are amplified by the receiving section 41. Is demodulated.
- the signal demodulated by the demodulation unit 301 is separated by the signal separation unit 302, the image data is stored in the image memory 303, and the white balance coefficient is set by the parameter detection unit 304. Is detected.
- the image signal processing unit 305 corrects the image data stored in the image memory 303 based on the parameter (white balance coefficient) detected by the parameter detection unit 304. That is, the image signal processing unit 304 obtains white balance of image data based on the white balance coefficient detected by the parameter detection unit 304.
- the parameters detected by the parameter detection unit 304 are stored in the parameter memory 208 and multiplexed by the multiplexing unit 209 to the image data.
- the image signal processing unit 305 performs processing such as contour enhancement and LPF / gamma correction in addition to the above-described white balance image processing. Unlike the white balance processing, these processings such as contour enhancement, LPF, and gamma correction are commonly performed by all capsule endoscopes 10. Therefore, it is not necessary to separately store parameters for performing the common processing in the parameter memory 208 of each forcepsell-type endoscope 10 and are common to all capsule-type endoscopes 10 It is sufficient to store this data in the image signal processing unit 305. .
- the image data corrected by the image signal processing unit 305 is compressed by the image compression unit 306 and then stored in the CF memory card 44.
- FIG. 7 shows a procedure for obtaining a white balance coefficient for each capsule endoscope 10 in the manufacturing process.
- each capsenolle endoscope 10 captures a white chart as a reference.
- a correction coefficient (white balance coefficient) is calculated on the basis of G (green) so that the outputs of R (red) and B (pull one) become specified values.
- the calculated R and B correction coefficients are recorded in the parameter memory 208.
- step S A4 the correction coefficients recorded in the parameter memories 2 and 8 are checked.
- the confirmation is to read the correction coefficient from the parameter memory 208 and to check whether the read correction coefficient and the correction coefficient calculated in step S A2 are the same or not.
- step S A5 it is determined whether or not there is a problem (NG) case a plurality of times.
- NG a problem
- step SA5 when a predetermined number of times are reached (SA5—Y), it is indicated that the capsule endoscope 10 (particularly the parameter memory 208) has an abnormality. Is indicated (step SA 6). The capsule endoscope 10 determined to be abnormal is not shipped as it is.
- FIG. 8 shows a data format of transmission data (frame) serving as a transmission unit when transmitted from the capsule endoscope 10 in the case of performing digital transmission.
- the transmission unit 405 is composed of data corresponding to one line of CCD 125.
- the multiplexing unit 209 starts to construct a new transmission unit 405, and sets the horizontal identification data 406 and the R white balance coefficient 402 in that order.
- the new transmission unit 405 is used as a component, and after that, the image data 407 input from the AZD unit 205 before the horizontal identification data 406 is input is transmitted to the transmission unit 405. Add as a component.
- the image data 407 input from the A / D unit 205 before the horizontal identification data 406 is input refers to the horizontal shift register of the CCD 125 during the horizontal retrace period. (Not shown), which corresponds to the image data for one line of the CCD 125 to which the charges are transferred.
- the white balance coefficient 402 is added to a portion of one line of the CCD 125 corresponding to a time other than the effective imaging time.
- the multiplexing unit 209 starts to construct a new transmission unit 405, and transmits the horizontal identification data 406 and the white balance coefficient 403 of B in this order.
- the horizontal synchronizing signal 210 indicating the beginning of the transmission unit 405 and the TG signal 202 for determining the timing of reading the charge from the CCD 125 are generated by the timing generator 205 and the sync generator 201, respectively.
- the image data 407 for one line of D 125 is generated so as to be synchronized with each other so as to be sent to the multiplexing unit 209 at the timing when the parameter 402 or 403 is read from the parameter memory 208.
- the multiplexing unit 209 can detect that the timing at which the horizontal identification data 406 is input from the parameter memory 208 is a delimiter of the transmission unit 405, and the image input from the A / D unit 205 by the time of the detection.
- the data is used as a component of the transmission unit 405 as image data 407 for one line of the CCD 125.
- FIG. 9 is a flowchart showing an example of the operation of the capsule endoscope 10 and the receiver 4.
- step SB4 when the power is turned on (step SB1—Y) and imaging is started (step SB2), every time image data for one line of the CCD 125 is read (step SB1—Y), SB 3-Y), the image data for one line and one of the R and B white balance coefficients 402 and 403 stored in the parameter memory 208 are multiplexed (step SB4).
- the multiplexed data is as shown in FIG.
- the multiplexed data shown in FIG. 8 is transmitted after being modulated (steps SB5 and SB6).
- the operation performed for each line as described above is performed in the same manner for all lines included in one frame of the CCD 125. After the operation is performed for one frame, the operation is performed for the next one frame. The same operation is performed (steps SB7 and SB8). These operations are repeated until the imaging is stopped (step SB8).
- the receiver 4 transmits the data from the capsule endoscope 10 in step SB6.
- the received data is received (step SB11-Y)
- the image data and the white balance coefficient are separated and detected for each image data of one line of the CCD 125 (step SB12, SB1). 3).
- each transmission unit 405 includes an R or B correction coefficient 402 or 403.
- each transmission unit 405 is configured so as to include one bit of each of the R or B correction coefficient 402 or 403 composed of a plurality of bits (for example, 8 bits). Is also good.
- the configuration is such that the R or B correction coefficient 400 or 400 consisting of a plurality of (8 in this example) transmission units 4 05 and a plurality of bits (8 bits in this example). Good.
- one transmission unit of data transmitted from the forcepsell type endoscope 10 corresponds to image data of 0.15 lines of one transmission unit and one CCD 125.
- data (frame) 400 which is one transmission unit when transmitted from the capsule endoscope 10 is replaced by a CCD. It can be configured to correspond to one frame of image data of 125.
- the multiplexing unit 209 When the multiplexing unit 209 receives the vertical identification data 401, the multiplexing unit 209 starts to construct a new transmission unit 400, and the vertical identification data is read in order from the parameter memory 209. 4 0 1, R white balance coefficient 4 0 2, B white balance coefficient 4 0 3 as the components of the new transmission unit 4 0 0, followed by the vertical identification data 4 0 1 Image data output from the A / D section 205 until 4 04 is added as a component of the transmission unit 400.
- the image data 404 output from the AZD section 205 before the vertical identification data 401 is input is transmitted to the vertical shift register (not shown) of the CCD 125 during the vertical blanking period.
- the white balance coefficients 402 and 403 are added to the portion corresponding to the time before the valid start line of the CCD 125.
- the vertical synchronizing signal 210 indicating the start of the transmission unit 400, and the TG signal 202 that determines the timing for reading the electric charge from the CCD 125 are transmitted by the timing generator and the sync generator 201, respectively.
- the image data 404 that constitutes one frame of D125 is sent to the multiplexing unit 209 at the timing when the parameters 402 and 403 are read from the parameter memory 208. And are generated to synchronize with each other.
- the multiplexing unit 209 can detect that the timing at which the vertical identification data 410 was input from the parameter memory 208 is a delimiter of the transmission unit 400, and the AZD
- the image data input from the unit 205 is regarded as a component of the transmission unit 400 as image data 404 for one frame.
- the white balance coefficients 400 and 400 are added to the receiver 4 so that Sent. .
- the data regarding the white balance coefficient included in each transmission unit 400 is the correction coefficients 402 and 403 of R and B.
- data on the white balance coefficient included in each transmission unit 405 is based on the force that is the R or B acquisition coefficient 402 or 403, or the correction coefficient of R or B 402 Or, it is 1-bit data that composes 4003.
- the data on the white balance coefficient included in each transmission unit 400 is smaller than the data on the white balance coefficient included in each transmission unit 400, because the vertical synchronization This is because the frequency of occurrence of the horizontal synchronization signal 210 is higher than that of the signal 210.
- each transmission unit 405 is generated at a relatively high frequency. Based on 405, all information about the white balance coefficient of the capsule endoscope 10 can be obtained in a short time. As shown in FIGS. 8 and 10, each transmission unit 400, 405 is transmitted to the receiver 4 in a form in which the correction coefficients 402, 403 are added. You.
- the white balance coefficient of each capsule endoscope 10 is a value uniquely determined as a value stored in the parameter memory 208 during the manufacturing process, and is not a variable value.
- the reason why the white balance coefficient is transmitted to the receiver 4 for each transmission unit 400, 405 is that the white balance coefficient is transmitted only when the capsule endoscope 10 is activated. For example, when the power of the receiver 4 is not turned on when the capsule endoscope 10 is activated, the receiver 4 cannot receive the white balance coefficient, and then performs the white balance processing. This is to prevent a situation in which an image is displayed without being performed.
- FIG. 11 is a flowchart showing a procedure of a white balance process performed by the receiver 4.
- communication is performed from the capsule endoscope 10 to the receiver 4 using the transmission unit 405 shown in FIG. 8 and the operation according to the flowchart shown in FIG. 9 is performed. I do.
- the number of detections i is set to 0 in the parameter detection section 304 (step SC 1).
- the signal separation unit 302 of the receiver 4 detects the horizontal identification data 406 from the input data and outputs the horizontal identification data 406.
- Detect white balance coefficient 402 or 4003 immediately after 06.
- the signal separation unit 302 converts the horizontal identification data 400 and the white balance coefficient 402 or 4003 and the image data 407.
- the image data 407 is separated and output to the image memory 303, and the horizontal identification data 406 and the white balance coefficient 402 or 403 are output to the parameter detection section 304.
- the parameter detecting section 304 acquires the white balance coefficient 402 or 4003 immediately after the horizontal identification data 400, and obtains the obtained white balance coefficient 402 or 4003 as the parameter detecting section. It is stored in the parameter memory area k (i) of 304 (step SC2). Thereafter, the parameter detection unit 304 adds 1 to the number of detections i and increments it (step S C3).
- n corresponds to the number of CCD 125 lines.
- the transmission unit 400 shown in FIG. 10 is used for communication from the capsule endoscope 10 to the receiver 4, n is the number of image frames. Corresponding.
- Step SC2 and SC3 are repeated until the number of detections reaches 11, and the white balance coefficients 402 and 400 are stored in the n parameter memory areas k ( ⁇ ) of the parameter detector 304, respectively. Then, the process proceeds to Step SC5 (Step SC4 - ⁇ ).
- the parameter detection unit 304 selects the data with a high frequency of occurrence among the white balance coefficients 402, 403 detected ⁇ times by using the white balance coefficients RWB, BWB To be adopted.
- the white balance coefficients RWB, BWB To be adopted.
- the image signal processing unit 305 determines the white balance of the image data 407 based on the white balance coefficients RWB and BWB adopted by the parameter detection unit 304 in step SC5. Perform processing. For R pixels, a value R out obtained by multiplying the input data R in by the white balance coefficient RWB is the result of the white balance processing. For the pixel B, the value Bout obtained by multiplying the input data Bin by the white balance coefficient BWB is the white parameter. This is the result of the sensing process.
- the circuit scale of the internal circuit does not increase and power consumption may increase. Absent.
- the circuit size of the internal circuit does not increase.
- a white balance chart is imaged, and a surface image obtained by imaging the chart is received.
- a method of calculating the white balance coefficient of the capsenolle type endoscope 10 based on the image of the received chart in the receiver 4 after being transmitted to the receiver 4 will be considered. According to this method, when the receiver 4 cannot receive the captured image data relating to the white balance coefficient at the time when the chart is captured (for example, when the power of the receiver 4 is not turned on at that time), the user may be alerted. Assuming that the capsule endoscope is swallowed, the image captured by the capsule endoscope is not subjected to white balance processing.
- the capsule endoscope 10 even if the data transmitted from the capsule endoscope 10 cannot be received by the receiver 4 before drinking the capsule endoscope 10, the capsule endoscope 10 is not used. After that, since the mirror 1 ⁇ always sends the white balance coefficients RWB and BWB data to the receiver 4 together with the captured image data, even if the receiver 4 is turned on after swallowing, Then, based on the received white balance coefficients RWB and BWB, white balance processing of the captured image can be performed.
- the white balance coefficients RWB and BWB are stored in the parameter memory 208.
- the parameter memory 208 stores an R image (R data) obtained by photographing a white chart during the manufacturing process. Store the B image (B data) as it is.
- the R image (Rd ata) and the B image (B data) are included at the location of the white balance coefficient 402 in FIG. 8 or at the locations of the white balance coefficients 402 and 403 in FIG. 10, respectively.
- transmission units 405 and 400 are configured. The rest of the configuration and operation of the capsule endoscope 10 are the same as those of the first embodiment.
- the receiver 4 has a constant Gr serving as a reference for R and a constant Gb serving as a reference for B used in white balance processing.
- the receiver 4 inputs an R image (Rd ata) or a B image (Bd ata) and image data 407 from the received transmission unit 405.
- the receiver 4 inputs the R image (Rd a t a), the B image (B d a t a), and the image data 404 from the received transmission unit 400.
- the constant Gr serving as a reference for R and the constant Gb serving as a reference for B can be changed for each place (hospital) where the capsule endoscope 10 is used.
- the white balance processing result can be made different depending on the place where the capsule endoscope 10 is used.
- the constant Gr and the constant Gb can be changed depending on the region of ⁇ ′ imaged by the capsule endoscope 10.
- the constant G r and the constant G can be changed by reflecting the original color of each organ and the color of the pathogen to be found in each ⁇ .
- FIG. 12 is a modification of the transmission unit 400 of FIG. In the transmission unit 400 in FIG. 12, immediately after the R white balance coefficient 402, the R white balance An error correction code 408 for the white balance coefficient 403 was added, and an error correction code 409 for the white balance coefficient 403 of B was added immediately after the white balance coefficient 403 of B. I have.
- the error correction codes 408 and 409 are stored together when the white balance coefficients RWB and BWB are stored in the parameter memory 208 during the manufacturing process of the capsule endoscope 10. Instead, only the white balance coefficients RWB and BWB are stored in the parameter memory 208, and the error correction codes 408 and 409 are stored inside the force-pull endoscope 10. Alternatively, a configuration in which the error correction codes 408 and 409 are calculated based on the white balance coefficients RWB and BWB read from the parameter memory 208 may be employed.
- the receiver 4 can correct the white balance coefficient 402 of R based on the error correction code 408, and can correct the white balance coefficient 409 of B based on the error correction code 409.
- an error correction code for the R white balance coefficient 402 is added between the R white balance coefficient 402 and the image data 407 of the transmission unit 405 in FIG.
- an error correction code for the white balance coefficient of B can be added between the white balance coefficient of B and the image data 407.
- the error correction codes 408, 409 are the time before the effective start line of the CCD 125, together with the white balance coefficients 400, 403. Is added to the part corresponding to.
- the error correction code is added together with the white balance coefficient to a portion corresponding to a time other than the effective imaging time in one line of 0 ⁇ 125.
- the parameter memory 208 stores pixel defect address data indicating the address of a pixel defect in addition to the white balance coefficient.
- the pixel defect correction is to correct a pixel defect existing in an address of a pixel based on pixel data corresponding to an address around the address of the pixel defect.
- the rest of the configuration of the capsule endoscope 10 is the same as that of the first embodiment.
- the operation of the forcepsell type endoscope 10 and the configuration and operation of the receiver 4 are basically the same as those in the first embodiment.
- the multiplexing unit 209 multiplexes the image data, the white balance coefficient, and the pixel defect address data, and the multiplexed data is transmitted to the modulation unit 211 and the radio unit 142. It is transmitted from the capsule endoscope 10 through the interface.
- the parameter detection unit 304 detects each parameter of the white balance coefficient and the pixel defect address data
- the image signal processing unit 304 detects the parameters for the image data.
- White balance processing is performed based on the white balance coefficient
- pixel defect correction is performed based on the detected pixel defect address data.
- the image on which the white balance processing and the pixel defect correction have been performed is compressed by the image compressing unit 303, and the compressed image data is stored in the large-capacity memory 44.
- the address of the pixel defect is also tested in the manufacturing process for each capsule endoscope 10 to determine the address of the pixel defect of the capsule endoscope 10.
- Can be The pixel defect address data is written to the parameter endoscope 208 of each capsule endoscope 10, and at the time of shipment, each pixel defect address data is stored in the parameter memory 208 of each capsule endoscope 10. It is shipped in the state stored in.
- FIG. 13 is a flowchart showing a procedure for calculating the address of a pixel defect during a manufacturing process.
- the CCD 125 is set to a place set at 50 ° C (step SD 1). This is because white defects in CCD 125 are likely to occur under high temperature conditions. .
- the CCD 125 captures an image of the shading ( ⁇ room) to guide white defects (step SD 2).
- the address of a pixel above a specified value level from the base (black) is recorded in the parameter memory 208 as pixel defect address data (step SD3).
- a white chart is imaged by the CCD 125, and a black defect is derived (step SD4).
- the address of the pixel below the specified value level from the base (white) is recorded in the parameter memory 208 as pixel defect address data (step SD5).
- step SD6 the pixel defect address data recorded in the parameter memory 208 is confirmed.
- the confirmation is performed by reading the pixel defect address data from the parameter memory 208 and determining whether or not the read pixel defect address data and the data of the address of the pixel defect detected in steps SD3 and SD5 are the same. This is the content to check.
- step SD7 it is determined whether or not there is a problem (NG) case a plurality of times (step SD7).
- NG problem
- step SD7-N since the predetermined number of times has not been reached yet (SD7-N), the process returns to step SD1.
- step SD7 when a predetermined number of times are reached (SD7-Y) ', it is displayed that the capsule endoscope (particularly the parameter memory 208) is abnormal (step SD8).
- the capsule endoscope 10 determined to be abnormal is not shipped as it is.
- FIG. 14 is a diagram showing transmission data 400, which is a transmission unit when received from the capsule endoscope 10 in the second embodiment, and corresponds to FIG. 10 of the first embodiment.
- FIG. The same components as those in the first embodiment are denoted by the same reference numerals, and description thereof is omitted.
- pixel defect address data 410 is included in the transmission unit 400.
- pixel defect address data can be transmitted between the white balance coefficient 402 of R and the image data 407 of the transmission unit 405 in FIG. 8 of the first embodiment, Similarly, pixel defect address data can be added between the white balance coefficient of B and the image data 407.
- the pixel defect address data is added together with the white balance coefficients 402 and 403 to a location corresponding to the time before the effective start line of the CCD 125.
- the pixel defect address data is added together with the white balance coefficient to a portion corresponding to a time other than the effective imaging time in one line of the CCD 125.
- pixel defects of the CCD 125 can be corrected.Note that the first or Z and second modifications of the first embodiment may be applied to the second embodiment. It is possible.
- the parameter memory 208 can store data for correcting a defect caused by the variation of the CCD 125.
- White balance coefficients and pixel defect address data are examples.
- the CCD 125 is used for the capsule endoscope 10 .
- a CMOS image sensor is used instead of the CCD 125.
- an offset value of a photoelectric conversion characteristic which is a value unique to each CMOS image sensor, is individually stored. Otherwise, the configuration and operation of the capsule endoscope 10 and the configuration and operation of the receiver 4 are the same as those in the first embodiment.
- the multiplexing unit 209 multiplexes the image data and the offset value of the photoelectric conversion characteristic.
- the multiplexed data is transmitted from the capsule endoscope 10 via the modulation section 211 and the radio section 142.
- the parameter detection unit 304 detects the parameter of the offset value of the photoelectric conversion characteristic
- the image signal processing unit 304 detects the offset value of the detected photoelectric conversion characteristic with respect to the image data.
- the correction of the photoelectric conversion characteristics is performed based on.
- the image for which the photoelectric conversion characteristics have been corrected is compressed by the image compression unit 306, and the compressed image data is stored in the large-capacity memory 44.
- the offset value of the photoelectric conversion characteristic is also tested in the manufacturing process for each capsule endoscope 10 in the same manner as the white balance coefficient in the first embodiment, and the capsule endoscope 10 The offset value of the photoelectric conversion characteristic is obtained.
- the offset value of the photoelectric conversion characteristic is written to the parameter memory 208 of each capsule endoscope 10, and at the time of shipment, the offset value of each photoelectric conversion characteristic is stored in each capsule endoscope 10.
- the product is shipped stored in the parameter memory 208.
- FIG. 15 is a graph for explaining how to obtain the offset value of the photoelectric conversion characteristic of each image sensor (for example, a CMOS image sensor).
- image sensor for example, a CMOS image sensor.
- signal outputs are obtained when different amounts of light are incident on each image sensor, and plotted as points ⁇ and ⁇ . Points ⁇ and ⁇ are connected by a straight line, and the intersection with the ⁇ axis is determined as the offset value of the photoelectric conversion characteristics of the imaging element.
- the third embodiment it is possible to correct the photoelectric conversion characteristics when the imaging device is used as the solid-state imaging device of the capsule endoscope 10.
- the additional information such as the offset value of the image data is added before the image data 404 and transmitted
- FIG. 16 shows a configuration in which white balance coefficients 402 and 403 are added to the rear end of image data 404.
- the additional information is stored in the image data 40 If added to the end of 4, the receiver will be able to receive the signal in a state where the synchronization by the vertical synchronization signal is more reliably achieved.
- the additional information is at a location where stable simultaneous synchronization is achieved.
- the additional information is at most 2 bytes in the case of FIG. 16, but since this additional information has a great effect on the restoration of the image data, the additional information is added to the rear end of the image data 404. It is preferable that the receiver side can acquire additional information stably and reliably. (Fourth embodiment).
- the fourth embodiment In contrast to the first embodiment in which digital transmission is performed, the fourth embodiment 'performs analog transmission.
- the same components as those in the first embodiment are denoted by the same reference numerals, and description thereof is omitted.
- the capsule endoscope 10 The image processing unit 144 a sends the analog image data output from the CCD 125 to the modulation unit 211 as an analog signal. I do. Because of analog transmission, there is no A / D converter 205 as shown in FIG.
- the parameter memory 208 stores white balance coefficients RWB and BWB, as in the parameter memory 208 of the first embodiment.
- the multiplexing unit 209 of the image processing unit 144 a includes a mixer 2 12 and an adder 2 13.
- the white balance coefficients RWB and BWB are read out from the parameter memory 208 to the mixer 212, and the white balance coefficients RWB, BW B and the synchronization signal SG 1 are mixed.
- the adder 2 13 the mixing result by the mixer 2 12 and the image data are superimposed.
- the output from the adder 2 13 is frequency-modulated by the modulator 2 11.
- the synchronization signal SG1 output from the timing generator and the sync generator 201 is superimposed on the image data in the multiplex section 209 as it is, so that the image data is In multiple images contained in Make the boundaries of each image clear.
- FIG. 19 shows an output signal S1 from the multiplexing unit 209 of FIG. As shown in Fig. 19, during analog transmission, the signal is transmitted in the same signal waveform format as the NTSC composite video signal.
- a portion 601 above the reference level 600 is a video signal (corresponding to image data), and a portion below the reference level is a synchronization signal SG1.
- Reference numeral 602 is a horizontal synchronization signal.
- the white balance coefficients RWB and BWB are mixed with the synchronization signal SG1 lower than the reference level 600.
- Reference numeral 603 is a vertical synchronization signal. '' As shown in FIGS.
- the vertical synchronizing signal 603 and the horizontal synchronizing signal 602 are mixed with the white balance coefficients RWB and BWB.
- the signal is mixed with the video signal 601 by the adder 213.
- the white balance coefficients RWB and BWB are superimposed after the vertical synchronizing signal 603 and added at a position corresponding to the time before the effective start line of the CCD 125 (left side of the video signal 601). Is done.
- the vertical synchronizing signal 603 that is kept at the low level for a long time is detected by passing the signal through an LPF (low-pass filter) in the receiver 4.
- the horizontal synchronization signal 602 is detected by passing through a BPF (Pand Pass Filter) in the receiver 4.
- BPF Packet Filter
- FIG. 20 shows another example of the output signal S1 from the multiplexing unit 209 'in FIG.
- the white balance coefficients RWB and BWB are mixed with the synchronization signal SG1 (below the reference level 600) and superimposed on the vertical synchronization signal 603. It differs from FIG. 19 in that the portion to be mixed is after the video signal 601 (in FIG. 1.9, it is-before the video signal 601).
- coefficient identification signals 605a and 605b indicating the existence of white balance coefficients RWB and BWB are added immediately before each white balance coefficient RWB and BWB.
- the coefficient identification signals 605a and 605b by detecting the coefficient identification signals 605a and 605b, it can be recognized that the white balance coefficients RWB and BWB exist immediately after that.
- both the R and B white balance coefficients RWB and BWB are arranged consecutively, only the coefficient identification signal 605 a is sufficient, and the coefficient • identification signal 605 b is unnecessary. .
- the coefficient identification signals 605a and 605b can be added immediately before the white balance coefficients RWB and BWB in the case of FIG.
- FIG. 19 and FIG. 20 show the case where both the R and B white balance coefficients RWB and BWB are superimposed on each vertical synchronization signal 603.
- Fig. 21 and Fig. 22 show that only one bit of data of white balance coefficient RWB or BWB (assumed to be composed of 8 bits D7 to D0) is included in each horizontal synchronization signal 602. This shows the case where they are superimposed.
- One bit of data of the white balance coefficient RW or BWB is added to a portion corresponding to a time other than the effective imaging time in one line of the CCD 125.
- the white balance coefficient data superimposed on the horizontal synchronizing signal 60 2 is smaller than the white balance coefficient data superimposed on the vertical synchronizing signal 603 as described above. This is because the frequency of occurrence of the horizontal synchronization signal 602 is higher than the frequency of occurrence of 3.
- Fig. 21 (a) the white balance of one bit superimposed on each of the eight horizontal synchronization signals 602 and the coefficients (D7 to D0) are aligned, so that the white balance of R is obtained.
- the coefficient RWB is detected, and the white balance coefficient RWB for B is detected by the white balance coefficient (D7 to DO) for one bit superimposed on each of the next eight horizontal synchronization signals 602.
- Fig. 21 (b) shows an example in which the timing for superimposing the horizontal synchronization signal 6 • 2 is shifted.
- Fig. 21 (b) differs from Fig. 21 (a) in that it is just before the fall of the horizontal synchronizing signal. You are inserting data. This makes it easier to detect the white balance coefficient when detecting the horizontal synchronization signal at the rising edge. If the white balance coefficient is high level (H), the width of the horizontal synchronization signal becomes narrower, so that the inserted coefficient can be detected as H or L depending on the level width of the horizontal synchronization signal.
- H high level
- FIG. 22 is different from FIG. 21 in that the same 1-bit data having the same white balance coefficient RWB or BWB is superimposed on three consecutive horizontal synchronizing signals 602 in common.
- the receiver 4 detects the above-mentioned 1-bit data of the superimposed white balance coefficient RWB or BWB for each of the three ice flat synchronization signals 602.
- the receiver 4 cannot read the white balance coefficient superimposed on one horizontal synchronizing signal 602, an accurate white balance coefficient RWB or BWB cannot be obtained.
- the one bit superimposed on the second horizontal synchronization signal 602 was erroneously recognized as the one bit superimposed on the first horizontal synchronization signal 602.
- the above-mentioned 1 bit superimposed on the third horizontal synchronization signal 602 can be correctly recognized and recognized as D6.
- D7 when determining D7, it is assumed that data having a high frequency of occurrence is determined as a white balance coefficient by referring to a three-line coefficient from the first synchronization signal.
- the image processing section 300 'of the receiver 4 has an A / D conversion section 300. Is added.
- the signal separation unit 300 2 ′ of the image processing unit 300 comprises a clamp circuit 701, a synchronization signal separation unit 702, a vertical synchronization detection unit 703, and a horizontal synchronization detection unit 704. And a line number detector 705.
- the clamp circuit 701 clamps the output signal from the demodulation unit 301 and separates the synchronization signal (horizontal synchronization signal 602 and vertical synchronization signal 603) SG1 from the video signal 601
- the reference level 600 is detected.
- the synchronization signal separation unit 702 separates the synchronization signal SG1 and outputs the video signal 601 to the A / D conversion unit 307.
- the sync signal SG 1 is supplied to the vertical sync detector 703 and the horizontal
- the vertical synchronization signal is sent to the synchronization detector 704, the vertical synchronization detector 703 detects the vertical synchronization signal 603, and the horizontal synchronization detector 704 detects the horizontal synchronization signal 602.
- the detection results of the vertical synchronization detection section 703 and the horizontal synchronization detection section 704 are sent to the line number detection section 705. '
- the line number detection unit 705 determines the R white balance coefficient after a predetermined clock from the horizontal synchronization signal 602 on the second line from the vertical synchronization signal 603. It is known that RWB is included, and that a white balance coefficient BWB of B is included after a predetermined clock from the horizontal synchronization signal 602 on the third line.
- the line number detection unit 705 sends the parameter detection unit 304 a predetermined clock after the ice synchronization signal 602 on the second line from the vertical synchronization signal 603, and It outputs a sampling phase output to indicate after a predetermined clock from the horizontal synchronization signal 602 of the line.
- the parameter detection unit 304 can obtain the white balance coefficients RWB and BWB from the synchronization signal SG1 based on the sampling phase output.
- FIG. 23 is a diagram showing a modification of FIG.
- the multiplexing units 209 are provided with a mixer 212, an adder 213 ', and a D / A converter 214.
- the white balance coefficients RWB and BWB read out from the parameter memory 208 are converted into analog signals by the DZA converter 214 and then mixed with the image data by the mixer 212.
- the adder 2 13 the result of mixing by the mixer 2 1 2 ′ is superimposed on the synchronization signal SG 1.
- the output from the adder 2 1 3 ′ of ⁇ is frequency-modulated by the modulation section 2 1 1.
- FIG. 24 shows an output signal S2 from the multiplexing unit 209 ', shown in FIG.
- the white balance coefficients RWB and BWB are mixed in the mixer 2 1 2 ′ to the image data 600 1 above the reference level 600.
- the white balance coefficient RWB is superimposed on the image data 60 1 of the second line after the first horizontal synchronization signal 60 2 after the vertical synchronization signal 60 3 rises, and the white balance coefficient BWB is 2 It is superimposed on the third line of image data 601, after the horizontal synchronization signal 602.
- the actual video signal 601 is from the fourth line after the third horizontal synchronization signal 602.
- the white balance coefficients RW B and BWB are transmitted before or in a distributed manner before and after a series of video signals 601.
- FIG. 25 shows a configuration in which white balance coefficients RWB and BWB are added to the rear end of a series of 11 video signals 601.
- the additional information such as the white balance coefficients RWB and BWB is, for example, at most 2 bytes in the case of Fig. 16, but since this additional information has a great effect on the restoration of the image data, a series of video signals 6 It is preferable to add additional information to the rear end side of 01, and the receiver side can acquire stable and reliable additional information.
- additional information such as error correction codes 408, 409, pixel defect address data 410, offset values of photoelectric conversion characteristics, etc.
- it is added to the rear end side of 600.
- FIGS. 26 and 27 a fifth embodiment will be described with reference to FIGS. 26 and 27.
- the same components as those in the first embodiment are denoted by the same reference numerals, and description thereof will be omitted.
- the capsule endoscope 10 performs end-to-end analog transmission.
- the parameter memory 208 is stored in the parameter memory 208.
- the white balance coefficient is modulated and transmitted alone without transmitting the stored white balance coefficient to the image signal, and the image signal is modulated and transmitted alone.
- the receiver 4 demodulates each of the two modulated signals to obtain a white balance coefficient and an image signal.
- the image processing unit 144 a of the capsule endoscope 10 does not include the multiplexing unit 209 unlike FIG. This is because, in the fifth embodiment, the white balance coefficient and the image signal are not multiplexed.
- the signal processing / control section 144 'shown in FIG. 26 includes two modulation sections 2111a and 211b.
- the modulation unit 211a modulates the white balance coefficient stored in the parameter memory 208 with the carrier frequency f1.
- the modulation unit 21b modulates the image signal at the carrier frequency f2.
- the transmission section 142 a amplifies the white balance coefficient modulated signal output from the modulation section 211 a and the image signal modulation signal output from the modulation section 211 b.
- the shared antenna 144b transmits the modulated signals having different carrier frequencies f1 and f2 amplified by the transmission section 144a.
- the receiver 4 differs from FIG. 4 in that it has two demodulators 301a and 301b, and the parameter detector 304 is an image processor. Provided outside of 300. Radio wave signals (modulation signals of white balance coefficients and modulation signals of image signals) captured by the common antennas 31 to 34 are amplified by the reception unit 41, respectively.
- the demodulation section 301a demodulates the modulated signal of the carrier frequency f1, and sends the demodulated signal to the parameter detection section 304.
- the parameter detector 304 detects a white balance coefficient based on the input signal.
- the demodulation section 301b demodulates the modulated signal of the carrier frequency f2, and sends the demodulated signal to the image processing section 300.
- the signal separation unit 302 of the image processing unit 300 separates the plane image signal and the synchronization signal.
- the image processing unit 304 accesses the parameter detection unit 304 using the synchronization signal, and obtains a white balance coefficient from the parameter detection unit 304.
- the image processing unit 300 uses the white balance coefficient to Perform white balance processing on the image signal.
- the fifth embodiment can also be realized by digital transmission.
- the operation of the capsule endoscope 10 and the operation of the receiver 4 up to the demodulators 301a and 301b are the same in the case of digital transmission. Since the image processing section 300 of the receiver 4 at the time of digital transmission does not need to separate the image signal and the synchronization signal, the signal separating section 302 becomes unnecessary, and the detection is performed by the parameter detecting section 304.
- the white balance processing may be performed on the image signal using the efficient white balance coefficient.
- the same effects as in the first embodiment can be obtained.
- ADVANTAGE OF THE INVENTION According to the capsule endoscope of this invention, it is low and low power consumption regarding the signal processing peculiar to an image sensor. Industrial applicability
- the present invention relates to a medical endoscope, and is particularly suitable for a swallowable forcepsell endoscope for capturing an image of a body cavity and a forcepschel endoscope system using the same.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Veterinary Medicine (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Physics & Mathematics (AREA)
- Biophysics (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Computer Networks & Wireless Communication (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Optics & Photonics (AREA)
- Radiology & Medical Imaging (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physiology (AREA)
- Psychiatry (AREA)
- Endoscopes (AREA)
- Closed-Circuit Television Systems (AREA)
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA002530718A CA2530718C (en) | 2003-06-24 | 2004-06-24 | Capsule endoscope and capsule endoscope system |
KR1020057024658A KR100757620B1 (ko) | 2003-06-24 | 2004-06-24 | 캡슐형 내시경 및 캡슐형 내시경 시스템 |
JP2005507337A JPWO2004112593A1 (ja) | 2003-06-24 | 2004-06-24 | カプセル型内視鏡及びカプセル型内視鏡システム |
EP04746736A EP1637064A4 (en) | 2003-06-24 | 2004-06-24 | CAPSULE ENDOSCOPE AND CAPSULE ENDOSCOPE SYSTEM |
CN2004800175725A CN1809309B (zh) | 2003-06-24 | 2004-06-24 | 胶囊型内窥镜以及胶囊型内窥镜系统 |
AU2004249063A AU2004249063B2 (en) | 2003-06-24 | 2004-06-24 | Encapsulated endoscope and encapsulated endoscope system |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2003180138 | 2003-06-24 | ||
JP2003-180138 | 2003-06-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2004112593A1 true WO2004112593A1 (ja) | 2004-12-29 |
Family
ID=33535120
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2004/009267 WO2004112593A1 (ja) | 2003-06-24 | 2004-06-24 | カプセル型内視鏡及びカプセル型内視鏡システム |
Country Status (8)
Country | Link |
---|---|
US (1) | US20050049461A1 (ja) |
EP (1) | EP1637064A4 (ja) |
JP (1) | JPWO2004112593A1 (ja) |
KR (1) | KR100757620B1 (ja) |
CN (1) | CN1809309B (ja) |
AU (1) | AU2004249063B2 (ja) |
CA (1) | CA2530718C (ja) |
WO (1) | WO2004112593A1 (ja) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006304885A (ja) * | 2005-04-26 | 2006-11-09 | Olympus Medical Systems Corp | 被検体内導入装置及び被検体内情報取得システム |
JP2007089837A (ja) * | 2005-09-29 | 2007-04-12 | Fujinon Corp | 電子内視鏡システム |
JP2007089839A (ja) * | 2005-09-29 | 2007-04-12 | Fujinon Corp | 電子内視鏡システム |
WO2007078003A1 (ja) | 2006-01-06 | 2007-07-12 | Olympus Medical Systems Corp. | 経自然開口的または経皮的な医療システム |
JP2007319442A (ja) * | 2006-06-01 | 2007-12-13 | Fujifilm Corp | カプセル内視鏡システム、および画像処理装置 |
JP2007330794A (ja) * | 2006-06-13 | 2007-12-27 | Given Imaging Ltd | 生体内撮像のための方法および生体内検知システム |
JP2008275786A (ja) * | 2007-04-26 | 2008-11-13 | Olympus Medical Systems Corp | 撮像ユニット |
WO2009069401A1 (ja) * | 2007-11-28 | 2009-06-04 | Olympus Medical Systems Corp. | 被検体内情報取得システムおよび被検体内導入装置 |
WO2009133875A1 (ja) | 2008-04-30 | 2009-11-05 | 学校法人自治医科大学 | 自然開口部越管腔内視鏡手術(notes)用外科手術システム及び外科手術方法 |
JP2010516088A (ja) * | 2007-01-09 | 2010-05-13 | カプソ・ビジョン・インコーポレイテッド | カプセルカメラにおける製造ばらつき及び設計的不完全性を補正する方法 |
JP2010110639A (ja) * | 2010-01-15 | 2010-05-20 | Olympus Corp | 受信装置および被検体内情報取得システム |
EP2407083A1 (en) * | 2004-09-16 | 2012-01-18 | Olympus Corporation | Capsule-type endoscope |
JP2013078591A (ja) * | 2012-11-21 | 2013-05-02 | Toshiba Corp | 撮像装置、撮像装置の作動方法及び内視鏡装置 |
Families Citing this family (47)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE60228266D1 (de) * | 2001-06-18 | 2008-09-25 | Given Imaging Ltd | Schluckbare in vivo-erfassungskapsel mit einer starre und flexible abschnitte aufweisenden leiterplatte |
US7833151B2 (en) | 2002-12-26 | 2010-11-16 | Given Imaging Ltd. | In vivo imaging device with two imagers |
JP4550048B2 (ja) * | 2003-05-01 | 2010-09-22 | ギブン イメージング リミテッド | パノラマ視野の撮像装置 |
IL162740A (en) * | 2003-06-26 | 2010-06-16 | Given Imaging Ltd | Device, method and system for reduced transmission imaging |
JP4150663B2 (ja) * | 2003-12-25 | 2008-09-17 | オリンパス株式会社 | 被検体内位置検出システム |
EP2572626B1 (en) * | 2004-01-19 | 2016-03-23 | Olympus Corporation | Capsule type endoscope |
WO2005067782A1 (ja) * | 2004-01-19 | 2005-07-28 | Olympus Corporation | カプセル型医療装置 |
JP4262632B2 (ja) * | 2004-04-19 | 2009-05-13 | オリンパス株式会社 | 受信装置 |
US8500630B2 (en) * | 2004-06-30 | 2013-08-06 | Given Imaging Ltd. | In vivo device with flexible circuit board and method for assembly thereof |
US7336833B2 (en) * | 2004-06-30 | 2008-02-26 | Given Imaging, Ltd. | Device, system, and method for reducing image data captured in-vivo |
JP2006026234A (ja) * | 2004-07-20 | 2006-02-02 | Olympus Corp | 生体内撮像装置および生体内撮像システム |
US20060104057A1 (en) * | 2004-10-28 | 2006-05-18 | Jerome Avron | Device and method for in-vivo illumination |
US20090281389A1 (en) * | 2004-12-30 | 2009-11-12 | Iddan Gavriel J | Device, system, and method for adaptive imaging |
US7584534B2 (en) * | 2005-01-10 | 2009-09-08 | Perceptron, Inc. | Remote inspection device |
EP1857036B1 (en) * | 2005-03-09 | 2012-07-25 | Olympus Corporation | Device to be introduced into subject and system to be introduced into subject |
US20060217593A1 (en) * | 2005-03-24 | 2006-09-28 | Zvika Gilad | Device, system and method of panoramic multiple field of view imaging |
US20060224040A1 (en) * | 2005-03-31 | 2006-10-05 | Given Imaging Ltd. | In vivo imaging device and method of manufacture thereof |
DE102005046086A1 (de) * | 2005-09-26 | 2007-04-05 | Leutron Vision Gmbh | Elektronisches Bildaufnahmesystem |
EP2033569B1 (en) * | 2006-06-29 | 2012-06-13 | Olympus Medical Systems Corp. | Capsule medical device and capsule medical device system |
US20080004532A1 (en) | 2006-06-30 | 2008-01-03 | Kevin Rubey | System and method for transmitting identification data in an in-vivo sensing device |
US8472795B2 (en) * | 2006-09-19 | 2013-06-25 | Capso Vision, Inc | System and method for capsule camera with on-board storage |
JP2008119145A (ja) | 2006-11-09 | 2008-05-29 | Olympus Medical Systems Corp | 画像表示方法および画像表示装置 |
KR100837588B1 (ko) * | 2006-11-16 | 2008-06-13 | 아이쓰리시스템 주식회사 | 아날로그 전기신호를 이용한 체내외 간 고속통신방법 및 그시스템 |
KR100876647B1 (ko) * | 2006-11-22 | 2009-01-08 | 주식회사 코렌 | 캡슐형 촬영 장치 및 이를 이용한 체내 촬영 방법 |
US20080161639A1 (en) * | 2006-12-28 | 2008-07-03 | Olympus Medical Systems Corporation | Capsule medical apparatus and body-cavity observation method |
US8702591B2 (en) * | 2007-01-12 | 2014-04-22 | Olympus Medical Systems Corp. | Capsule medical apparatus |
US10499029B2 (en) * | 2007-01-09 | 2019-12-03 | Capso Vision Inc | Methods to compensate manufacturing variations and design imperfections in a display device |
US9007478B2 (en) * | 2007-01-09 | 2015-04-14 | Capso Vision, Inc. | Methods to compensate manufacturing variations and design imperfections in a capsule camera |
JP5340557B2 (ja) * | 2007-05-08 | 2013-11-13 | オリンパスメディカルシステムズ株式会社 | カプセル型医療装置 |
JP4918438B2 (ja) * | 2007-08-31 | 2012-04-18 | オリンパスメディカルシステムズ株式会社 | 被検体内情報取得システム |
US20090105532A1 (en) * | 2007-10-22 | 2009-04-23 | Zvika Gilad | In vivo imaging device and method of manufacturing thereof |
JP5377888B2 (ja) * | 2008-06-03 | 2013-12-25 | オリンパスメディカルシステムズ株式会社 | 撮像装置および被検体内画像取得装置 |
US8516691B2 (en) * | 2009-06-24 | 2013-08-27 | Given Imaging Ltd. | Method of assembly of an in vivo imaging device with a flexible circuit board |
WO2011047339A2 (en) * | 2009-10-15 | 2011-04-21 | Inventio Llc | Disposable and reusable complex shaped see-through endoscope |
US9610133B2 (en) * | 2010-03-16 | 2017-04-04 | Covidien Lp | Wireless laparoscopic camera |
CN103781394B (zh) | 2011-04-05 | 2016-09-28 | 可视化气囊有限责任公司 | 用于内窥镜的气囊状进入设备 |
US9833126B2 (en) | 2011-04-05 | 2017-12-05 | Visualization Balloons, Llc | Balloon access device with features for engaging an endoscope |
JP5927039B2 (ja) * | 2012-05-28 | 2016-05-25 | 富士フイルム株式会社 | 電子内視鏡装置及びその撮像モジュール |
US9538909B2 (en) * | 2013-07-08 | 2017-01-10 | Omnivision Technologies, Inc. | Self-illuminating CMOS imaging package |
KR101858877B1 (ko) * | 2013-08-28 | 2018-06-28 | 도시바 라이프스타일 가부시키가이샤 | 저장고용 카메라 장치 및 이를 구비한 저장고 |
US10463235B2 (en) | 2014-02-24 | 2019-11-05 | Visualization Balloons, Llc | Gastrointestinal endoscopy with attachable intestine pleating structures |
JP6633067B2 (ja) * | 2015-05-27 | 2020-01-22 | オリンパス株式会社 | 撮像装置および内視鏡 |
EP3135189A1 (en) * | 2015-08-25 | 2017-03-01 | Capso Vision, Inc. | Methods to compensate manufacturing variations and design imperfections in a display device |
DE112018001670T5 (de) * | 2017-03-30 | 2019-12-19 | Hoya Corporation | Elektronische endoskopvorrichtung |
CN109567728B (zh) * | 2018-11-27 | 2021-12-17 | 重庆金山医疗技术研究院有限公司 | 用于电子内窥镜的光耦合热量处理装置 |
KR102248552B1 (ko) * | 2019-03-22 | 2021-05-04 | 재단법인대구경북과학기술원 | 생물학적 물질의 운반 디바이스 |
US11534544B2 (en) | 2019-03-22 | 2022-12-27 | Daegu Gyeongbuk Institute Of Science And Technology | Device for conveying biological material |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0070169A1 (en) | 1981-07-09 | 1983-01-19 | Olympus Optical Co., Ltd. | Endoscope with monitor |
JPS6349128A (ja) * | 1986-08-19 | 1988-03-01 | 株式会社東芝 | 内視鏡装置 |
JPH06335449A (ja) * | 1993-05-31 | 1994-12-06 | Olympus Optical Co Ltd | 電子内視鏡装置 |
JPH11298907A (ja) | 1998-04-10 | 1999-10-29 | Asahi Optical Co Ltd | 電子内視鏡システム |
WO2001035813A1 (en) | 1999-11-15 | 2001-05-25 | Given Imaging Ltd. | Method for activating an image collecting process |
JP2001245844A (ja) | 2000-03-03 | 2001-09-11 | Asahi Optical Co Ltd | カプセル内視鏡 |
WO2003010967A1 (en) | 2001-07-26 | 2003-02-06 | Given Imaging Ltd. | Diagnostic device using data compression |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7453490B2 (en) * | 1997-01-31 | 2008-11-18 | Gyrus Acmi, Inc. | Correction of image signals characteristic of non-uniform images in an endoscopic imaging system |
US6100920A (en) * | 1997-01-31 | 2000-08-08 | Circon Corporation | Video signal compensator for compensating differential picture brightness of an optical image due to uneven illumination and method |
US5984875A (en) * | 1997-08-22 | 1999-11-16 | Innotek Pet Products, Inc. | Ingestible animal temperature sensor |
JPH11164831A (ja) * | 1997-12-03 | 1999-06-22 | Aloka Co Ltd | 超音波診断装置 |
JP2000232653A (ja) * | 1999-02-08 | 2000-08-22 | Fuji Photo Film Co Ltd | 撮影装置 |
US7053941B1 (en) * | 1999-08-19 | 2006-05-30 | Canon Kabushiki Kaisha | Image input apparatus |
WO2001065995A2 (en) * | 2000-03-08 | 2001-09-13 | Given Imaging Ltd. | A device and system for in vivo imaging |
JP2001318745A (ja) * | 2000-05-11 | 2001-11-16 | Sony Corp | データ処理装置およびデータ処理方法、並びに記録媒体 |
JP3962550B2 (ja) * | 2001-02-23 | 2007-08-22 | フジノン株式会社 | 電子内視鏡装置 |
US6939292B2 (en) * | 2001-06-20 | 2005-09-06 | Olympus Corporation | Capsule type endoscope |
US20050187433A1 (en) * | 2001-07-26 | 2005-08-25 | Given Imaging Ltd. | In-vivo imaging device providing constant bit rate transmission |
IL151049A0 (en) * | 2001-08-02 | 2003-04-10 | Given Imaging Ltd | In vivo imaging methods and devices |
US7123288B2 (en) * | 2001-09-28 | 2006-10-17 | Fujinon Corporation | Electronic endoscope eliminating influence of light distribution in optical zooming |
JP2004350963A (ja) * | 2003-05-29 | 2004-12-16 | Olympus Corp | カプセル型医療装置 |
US20050075537A1 (en) * | 2003-10-06 | 2005-04-07 | Eastman Kodak Company | Method and system for real-time automatic abnormality detection for in vivo images |
-
2004
- 2004-06-24 JP JP2005507337A patent/JPWO2004112593A1/ja active Pending
- 2004-06-24 US US10/876,812 patent/US20050049461A1/en not_active Abandoned
- 2004-06-24 CA CA002530718A patent/CA2530718C/en not_active Expired - Fee Related
- 2004-06-24 EP EP04746736A patent/EP1637064A4/en not_active Withdrawn
- 2004-06-24 WO PCT/JP2004/009267 patent/WO2004112593A1/ja active Application Filing
- 2004-06-24 CN CN2004800175725A patent/CN1809309B/zh not_active Expired - Fee Related
- 2004-06-24 KR KR1020057024658A patent/KR100757620B1/ko not_active IP Right Cessation
- 2004-06-24 AU AU2004249063A patent/AU2004249063B2/en not_active Ceased
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0070169A1 (en) | 1981-07-09 | 1983-01-19 | Olympus Optical Co., Ltd. | Endoscope with monitor |
JPS6349128A (ja) * | 1986-08-19 | 1988-03-01 | 株式会社東芝 | 内視鏡装置 |
JPH06335449A (ja) * | 1993-05-31 | 1994-12-06 | Olympus Optical Co Ltd | 電子内視鏡装置 |
JPH11298907A (ja) | 1998-04-10 | 1999-10-29 | Asahi Optical Co Ltd | 電子内視鏡システム |
WO2001035813A1 (en) | 1999-11-15 | 2001-05-25 | Given Imaging Ltd. | Method for activating an image collecting process |
JP2001245844A (ja) | 2000-03-03 | 2001-09-11 | Asahi Optical Co Ltd | カプセル内視鏡 |
WO2003010967A1 (en) | 2001-07-26 | 2003-02-06 | Given Imaging Ltd. | Diagnostic device using data compression |
Non-Patent Citations (1)
Title |
---|
See also references of EP1637064A4 |
Cited By (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8421853B2 (en) | 2004-09-16 | 2013-04-16 | Olympus Corporation | Capsule endoscope system |
EP2407083A1 (en) * | 2004-09-16 | 2012-01-18 | Olympus Corporation | Capsule-type endoscope |
JP4602828B2 (ja) * | 2005-04-26 | 2010-12-22 | オリンパスメディカルシステムズ株式会社 | 被検体内情報取得システム |
JP2006304885A (ja) * | 2005-04-26 | 2006-11-09 | Olympus Medical Systems Corp | 被検体内導入装置及び被検体内情報取得システム |
JP2007089837A (ja) * | 2005-09-29 | 2007-04-12 | Fujinon Corp | 電子内視鏡システム |
JP2007089839A (ja) * | 2005-09-29 | 2007-04-12 | Fujinon Corp | 電子内視鏡システム |
WO2007078003A1 (ja) | 2006-01-06 | 2007-07-12 | Olympus Medical Systems Corp. | 経自然開口的または経皮的な医療システム |
US8475361B2 (en) | 2006-01-06 | 2013-07-02 | Olympus Medical Systems Corp. | Percutaneous or natural-orifice medical procedure and system therefor |
JP2007319442A (ja) * | 2006-06-01 | 2007-12-13 | Fujifilm Corp | カプセル内視鏡システム、および画像処理装置 |
JP2007330794A (ja) * | 2006-06-13 | 2007-12-27 | Given Imaging Ltd | 生体内撮像のための方法および生体内検知システム |
JP2010516088A (ja) * | 2007-01-09 | 2010-05-13 | カプソ・ビジョン・インコーポレイテッド | カプセルカメラにおける製造ばらつき及び設計的不完全性を補正する方法 |
JP2013243679A (ja) * | 2007-01-09 | 2013-12-05 | Capso Vision Inc | カプセルカメラにおける製造ばらつき及び設計的不完全性を補正する方法 |
JP2015053683A (ja) * | 2007-01-09 | 2015-03-19 | カプソ・ビジョン・インコーポレイテッドCapso Vision, Inc. | カプセルカメラにおける製造ばらつき及び設計的不完全性を補正する方法 |
JP2008275786A (ja) * | 2007-04-26 | 2008-11-13 | Olympus Medical Systems Corp | 撮像ユニット |
US8587718B2 (en) | 2007-04-26 | 2013-11-19 | Olympus Medical Systems Corp. | Image pickup unit and manufacturing method of image pickup unit |
JP2009131320A (ja) * | 2007-11-28 | 2009-06-18 | Olympus Medical Systems Corp | 被検体内情報取得システムおよび被検体内導入装置 |
US8366608B2 (en) | 2007-11-28 | 2013-02-05 | Olympus Medical Systems Corp. | In-vivo information acquiring system and body-insertable apparatus |
WO2009069401A1 (ja) * | 2007-11-28 | 2009-06-04 | Olympus Medical Systems Corp. | 被検体内情報取得システムおよび被検体内導入装置 |
WO2009133875A1 (ja) | 2008-04-30 | 2009-11-05 | 学校法人自治医科大学 | 自然開口部越管腔内視鏡手術(notes)用外科手術システム及び外科手術方法 |
JP2010110639A (ja) * | 2010-01-15 | 2010-05-20 | Olympus Corp | 受信装置および被検体内情報取得システム |
JP2013078591A (ja) * | 2012-11-21 | 2013-05-02 | Toshiba Corp | 撮像装置、撮像装置の作動方法及び内視鏡装置 |
Also Published As
Publication number | Publication date |
---|---|
AU2004249063B2 (en) | 2008-12-11 |
KR100757620B1 (ko) | 2007-09-10 |
CA2530718C (en) | 2009-09-01 |
KR20060030051A (ko) | 2006-04-07 |
EP1637064A1 (en) | 2006-03-22 |
AU2004249063A1 (en) | 2004-12-29 |
CA2530718A1 (en) | 2004-12-29 |
US20050049461A1 (en) | 2005-03-03 |
CN1809309A (zh) | 2006-07-26 |
CN1809309B (zh) | 2010-11-10 |
EP1637064A4 (en) | 2010-07-28 |
JPWO2004112593A1 (ja) | 2006-07-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2004112593A1 (ja) | カプセル型内視鏡及びカプセル型内視鏡システム | |
AU2005283435B2 (en) | Capsule-type endoscope | |
US8279276B2 (en) | Receiving apparatus | |
AU2005244523B2 (en) | A Device and System for In Vivo Imaging | |
JP4663230B2 (ja) | 小さな横断面面積を有するインビボ・イメージング・デバイスおよびその構成方法 | |
JP4422679B2 (ja) | カプセル内視鏡およびカプセル内視鏡システム | |
JP5164473B2 (ja) | 内視鏡装置 | |
JP4918438B2 (ja) | 被検体内情報取得システム | |
JP4445799B2 (ja) | 被検体内導入装置および医療装置 | |
JP2005021516A (ja) | 電子内視鏡システム | |
JP5160621B2 (ja) | カプセル型内視鏡システム | |
JP4823621B2 (ja) | 受信装置、送信装置及び送受信システム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2005507337 Country of ref document: JP |
|
REEP | Request for entry into the european phase |
Ref document number: 2004746736 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2004746736 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2004249063 Country of ref document: AU |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020057024658 Country of ref document: KR Ref document number: 20048175725 Country of ref document: CN Ref document number: 2530718 Country of ref document: CA |
|
ENP | Entry into the national phase |
Ref document number: 2004249063 Country of ref document: AU Date of ref document: 20040624 Kind code of ref document: A |
|
WWP | Wipo information: published in national office |
Ref document number: 2004249063 Country of ref document: AU |
|
WWP | Wipo information: published in national office |
Ref document number: 2004746736 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1020057024658 Country of ref document: KR |