WO2019155628A1 - Image processing device, image processing method, and recording medium - Google Patents

Image processing device, image processing method, and recording medium Download PDF

Info

Publication number
WO2019155628A1
WO2019155628A1 PCT/JP2018/004691 JP2018004691W WO2019155628A1 WO 2019155628 A1 WO2019155628 A1 WO 2019155628A1 JP 2018004691 W JP2018004691 W JP 2018004691W WO 2019155628 A1 WO2019155628 A1 WO 2019155628A1
Authority
WO
WIPO (PCT)
Prior art keywords
object region
heat map
image processing
image
detection target
Prior art date
Application number
PCT/JP2018/004691
Other languages
French (fr)
Japanese (ja)
Inventor
雅弘 西光
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Priority to PCT/JP2018/004691 priority Critical patent/WO2019155628A1/en
Priority to JP2019570260A priority patent/JP6988926B2/en
Publication of WO2019155628A1 publication Critical patent/WO2019155628A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/045Control thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection

Definitions

  • the present invention relates to an image processing apparatus for facilitating confirmation of a detection target object in an image.
  • Non-Patent Document 1 and Non-Patent Document 2 disclose examples of this image processing method.
  • this image processing method operates as follows in an apparatus including an object region candidate generation unit and an object region candidate output unit. That is, when an image (in the case of a moving image, each frame image constituting the moving image) is input, the object region candidate generation unit is configured to output a plurality of candidates (in the region where the object to be detected included in the image is considered to exist) Object region candidate) is generated.
  • the object region candidates are rectangles or circles circumscribing the object to be detected, and these are variables indicating the shape of the object (for example, center coordinates (x, y), height (h) and width (w)), and This is expressed by the reliability indicating the content of the detection target object in the object region candidate.
  • the object area candidate output means outputs a rectangle or a circle circumscribing the object to be detected based on the degree of overlap or reliability of each object area candidate from the object area candidates generated by the object area candidate generation means. To do.
  • Patent Document 1 discloses a technique related to the present application.
  • the detection target object has a regular shape and is an object that can be predicted to some extent.
  • the size, shape, etc. of the detection target object such as a human face, is regularly predicted to some extent, and the predetermined height (h) and width (w) from the center coordinates (x, y) in the image It is an object that fits in.
  • the object to be detected cannot be predicted irregularly in its size, shape, etc., and may not fit within a predetermined height and width preset from the center coordinates in the image (hereinafter referred to as this). Are called irregular objects).
  • Non-Patent Document 1 and Non-Patent Document 2 it is a non-rigid body that does not fit in a certain shape, such as a lesion inside or outside an organ (such as a tumor), and can be irregularly deformed into various shapes as the disease progresses.
  • a certain shape such as a lesion inside or outside an organ (such as a tumor)
  • one object region candidate is set so as to include the detection target object as much as possible in the region and not to include anything other than the detection target object as much as possible.
  • a plurality of object region candidates partially overlap and are generated in succession.
  • the detection target object is a round shape
  • a single object region candidate centered on the center of gravity is often sufficient
  • the detection target object is a starfish type
  • a plurality of object region candidates are sufficient.
  • the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2 circumscribe the entire irregular object (or the detection target object). Rectangle or circle area candidates (including the whole) are difficult to output.
  • Non-Patent Document 1 and Non-Patent Document 2 a part of a plurality of object region candidates are generated by overlapping one irregular detection target object.
  • the operator for example, “doctor” in the case of a lesion
  • the frame representing the object region candidate does not appropriately represent the detection target object for some reason, there is a problem that the observation by the operator is misleaded to the frame and the detection target object to be discovered is overlooked.
  • Non-Patent Document 1 since region candidates are output based on the overlap and reliability of each region candidate, the object region candidate circumscribing the detection target object has higher reliability than other object region candidates. Become. However, the candidate object region circumscribing an irregular object contains many non-object regions, which generally results in lower reliability than a partially circumscribing rectangle, and thus accurate. There arises a problem that an object region candidate cannot be extracted.
  • Non-Patent Document 2 a problem is that high-speed image processing is difficult because an object region of an image is generated or regenerated using a neural network that requires a large amount of calculation. is there.
  • an object of the present invention is to provide an image processing apparatus or the like that displays a frame indicating the entire irregular object at high speed.
  • the image processing apparatus provides: An object region candidate generation unit that generates a plurality of object region candidates indicating positions where at least a part of the detection target object exists in the image; Using the generated object region candidate, a heat map generation unit that generates a heat map corresponding to the position in the image, And an object region output unit that outputs an object region that surrounds the entire detection target object, which is made up of one or more object region candidates, using the generated heat map.
  • An image processing method includes: Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image; Using the generated object region candidate, generate a heat map corresponding to the position in the image, Using the generated heat map to output an object region that includes one or more object region candidates and surrounds the entire detection target object.
  • An image processing program is: Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image; Using the generated object region candidate, generate a heat map corresponding to the position in the image, An image processing program for causing a computer to output an object region that surrounds the entire detection target object, which is composed of one or more object region candidates, using a generated heat map.
  • the image processing program may be stored in a non-transitory computer-readable storage medium.
  • an image processing apparatus or the like that displays a frame indicating the entire irregular object at high speed.
  • FIG. 1 is a block diagram illustrating a configuration example of an image processing apparatus according to a first embodiment of the present invention.
  • 3 is a flowchart showing an operation in the image processing apparatus according to the first embodiment of the present invention. It is a figure which shows the object area
  • an image processing apparatus or the like that mainly targets a detection target object having an irregular shape and an unpredictable object region will be described.
  • the image processing apparatus can also detect an object having a regular shape and a predictable object region.
  • an irregular detection target object is a non-rigid body, for example, a lesion (tumor) that occurs inside or outside an organ.
  • the user of the image processing apparatus is a medical person such as a doctor.
  • This image processing apparatus is incorporated in, for example, a real-time tumor detection system for endoscopic moving images.
  • a moving image or still image hereinafter simply referred to as an image
  • the doctor can check that all or part of the internal or external organ where the lesion occurs is normal. By observing whether or not there is a mutation, the mutation type, malignancy and progression of the tumor are determined.
  • the image processing apparatus in the embodiment of the present invention displays a frame circumscribing the entire tumor accurately and at high speed as information indicating the position of the tumor. Thereby, it is possible to assist the diagnosis and treatment by the doctor quickly and accurately.
  • the image processing apparatus 100 includes an object region candidate generation unit 11, a heat map generation unit 12, and an object region output unit 13.
  • the object area candidate generation unit 11 generates a plurality of object area candidates indicating positions where at least a part of the detection target object exists in the image. Specifically, the object region candidate generation unit 11 inputs an image captured by an external camera (not shown in FIG. 1), for example, an endoscope camera, and there is a detection target object in the input image. A plurality of region candidates (object region candidates) are generated. As a method for generating a plurality of object region candidates in a region where an object to be detected included in an image is supposed to exist, existing methods described in Non-Patent Document 1, Non-Patent Document 2, and the like may be used. When there are too many object area candidates, the object area candidate generation unit 11 generates some combinations of the object area candidates suitable as the object area candidates, and outputs one of them as an object area candidate (group). May be.
  • the heat map generation unit 12 generates a heat map corresponding to the position in the image using the generated object region candidate. Specifically, the heat map generation unit 12 inputs one or more images including object region candidates, and generates a heat map related to the image based on the reliability of each object region candidate.
  • the reliability is a value indicating the probability that the detection target object exists in the object region candidate. For example, the degree of overlap of each object region candidate or the content rate of the detection target object in a certain object region candidate, etc. It is calculated based on the degree indicating Further, the heat map refers to a diagram in which a data matrix or the like indicating each value of a section (for example, pixel) in an image is visibly expressed as a color in two dimensions. In the heat map, the reliability score is set to be high in the region where the object exists.
  • the object region output unit 13 inputs the generated heat map, and outputs an object region (rectangular frame) indicating the entire irregular detection target object in the image based on the heat map.
  • the object region output unit 13 includes, for each of a plurality of sections (for example, for each pixel) into which the heat map is divided, an object including a set of sections whose reliability is equal to or higher than a predetermined threshold.
  • a region candidate is output as the object region.
  • step S101 the image processing apparatus 100 inputs an image to the object region candidate generation unit 11, and generates an object region candidate.
  • the image processing apparatus 100 inputs an image to the object region candidate generation unit 11, and generates an object region candidate.
  • a specific operation example of this step will be described below.
  • the object region candidate generation unit 11 receives an image including a tumor, and for this image, An object region candidate indicating a region including a tumor is generated.
  • a technique for realizing high-speed object recognition a technique that uses a rectangle that circumscribes an object to be detected as an object region is adopted.
  • the drawing function (Rectangle) the central coordinates (x, y), width (w), height (h), and the reliability (conf ) Shall be used.
  • a certain nth (n is a positive integer) object region candidate (Rectangle) can be expressed by the function of Equation 1.
  • Rectangle n [x n, y n, w n, h n, conf n] ... ( Equation 1)
  • the object region is an example of a rectangle surrounded so as to be in contact with the entire outer edge (outer periphery) of the detection target object (hereinafter, this state is also referred to as circumscribing), but this is not limited to a rectangle.
  • Yen may be used.
  • the object region may be enclosed so as to be in contact with the entire inner periphery of the detection target object, instead of circumscribing the object region.
  • the center of gravity of the object may be used as the center coordinate. Further, a variable representing rotation may be added and used.
  • FIG. 3 shows an example in which a plurality of object region candidates obtained by inputting a plurality of variables (n) to the function of Expression (1) are represented on a heat map.
  • 3 is a rectangular area candidate surrounded by the thick line frame so as to be in contact with the entire detection target object (tumor) (hereinafter, this state is also referred to as circumscribing), and other rectangles indicate a part of the detection target object.
  • This is a rectangular area candidate shown.
  • the rectangular area candidates are set so as to include the detection target object as much as possible in the area and not to include anything other than the detection target object as much as possible.
  • step S102 the generated plurality of object region candidates are received, and the heat map generation unit 12 generates a heat map based on the plurality of object region candidates.
  • the heat map generation unit 12 generates a heat map based on the plurality of object region candidates.
  • (i, j) represents the coordinate position of the pixel in the heat map.
  • FIG. 4 shows an example of a heat map generated by the heat map generator 12.
  • the value of the object region candidate is set using the reliability conf n value of each of the plurality of pixels included in each object region candidate of the heat map.
  • the value of the reliability conf n of a certain object region candidate is set as the value of a pixel existing in the object region candidate.
  • the reliability value of the portion C is the object region candidate A And a value obtained by adding the reliability of B.
  • the object area output unit 13 that has acquired the generated heat map selects an object area candidate that circumscribes the entire detection target object from among a plurality of object area candidates, and outputs the object area candidate as an object area.
  • the input heat map image (FIG. 4) has a value indicating the probability (reliability) that an object exists in each pixel.
  • the threshold value may be set in advance by the designer, or may be set to be changeable by the operator.
  • the heat map determined that the detection target object exists is as shown in FIG. It is assumed that the detection target object (tumor) exists as shown in FIG. At this time, the object area output unit 13 outputs a frame (frame 1 in FIGS.
  • the output image is displayed on an external monitor or the like and viewed by an operator such as a doctor.
  • the output object region is output as a frame (rectangular frame) along the shape of the detection target object.
  • the image processing apparatus 100 can display a frame indicating the entire irregular object at high speed. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the image processing apparatus 100 outputs each object region using a heat map based on the reliability of each object region candidate, and thus correctly outputs the object region circumscribing the entire detection target object. Further, since the object region is output by comparing the pixel value in the heat map generated from the reliability of the object region candidate generated by the object region candidate generating unit 11 with a preset threshold value, the neural network calculation is performed. This is because the object region is output at high speed with a much smaller calculation amount than the amount.
  • the image processing apparatus 100 outputs the object region of the detection target object based on the heat map, but the output of the object region includes a heat map and a corresponding object region candidate. May be used simultaneously. Further, in the process in which the object region candidate generation unit 11 of the first embodiment generates object region candidates, if the number of the object region candidates is too large, it takes a lot of processing time to calculate the combination of the object region candidates. Was. Therefore, in the second embodiment of the present invention, a method of outputting the object region of the detection target object by using the heat map and the object region candidate will be described by simplifying the processing to be generated.
  • the image processing apparatus 200 includes an object region candidate generation unit 21, a heat map generation unit 22, and an object region output unit 23.
  • the object region output unit 23 uses the object region candidate generated by the object region candidate generation unit 21 and the heat map generated by the heat map generation unit 22 corresponding to the object region candidate (heat map on which the generated object region candidates are superimposed). Then, an object region candidate circumscribing the entire detection target object is selected from among a plurality of object region candidates and output as an object region.
  • the object region candidate generation unit 21 generates an object region candidate. However, the object region candidate generation unit 21 is different from the first embodiment (object region candidate generation unit 11) when the number of object region candidates is too large. The combination of is not generated.
  • the generated object area candidate outputs the object area candidate to the heat map generation unit 22 and the object area output unit 23.
  • the heat map generator 22 is the same as that of the first embodiment (heat map generator 12).
  • Steps S201 and S202 are the same as steps S101 and S102 in FIG. 2, respectively.
  • step S203 the object region output unit 23 acquires the object region candidate from the object region candidate generation unit 21 and the heat map from the heat map generation unit 22, and outputs the object region based on these.
  • a specific operation example of this step will be described below.
  • FIG. 9 shows a plurality of object region candidates (frames) drawn on the heat map (see FIG. 4).
  • Rectangle n [x n , y n , w n , h n , conf n ])), based on the value (reliability) of each pixel in the heat map corresponding to the pixel in the object area
  • the object region output unit 23 compares the score of the object region candidate (sum of reliability of all the pixels in the object region candidate) with a preset threshold value, and the object region candidate having a score equal to or higher than the threshold value. Is output as the object region.
  • the object area output unit 23 extracts object area candidates whose object area candidate score is equal to or greater than a threshold value from the object area candidates, and outputs the area surrounded by the outer edges as the object area.
  • FIG. 10 shows an object region 2c composed of two extracted object region candidates 2a and 2b. It should be noted that the object region candidates 2a and 2b may be displayed as they are as long as they are numbers that can be easily visually recognized by the operator (for example, two in FIG. 10), without necessarily being combined into one. Further, the threshold value may be set in advance by the designer, or may be set to be changeable by the operator.
  • the image processing apparatus 200 can display a frame indicating the entire irregular object at high speed. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the image processing apparatus 200 outputs each object region using the heat map score based on the reliability of each object region candidate, and thus correctly outputs the object region circumscribing the entire detection target object. is there. Further, since the object region is output by comparing the pixel value in the heat map generated from the reliability of the object region candidate generated by the object region candidate generating unit 21 with a preset threshold value, the neural network calculation is performed. This is because the object region is output at high speed with a much smaller calculation amount than the amount.
  • the processing of the object region candidate generation unit 21 is simplified in comparison with the first embodiment, so that the entire processing time can be shortened.
  • the heat map generation unit 22 needs to perform aggregation processing for calculating the score from the heat map, but the object region output unit 23 configures the object region from the object region candidates using the heat map.
  • the processing in the object region candidate generation unit 11 is simplified. For this reason, even when the number of object region candidates is large, it is possible to avoid the processing time of the object region candidate generating unit 11 from being prolonged, and as a result, the processing time as a whole can be shortened.
  • the detection target object is stationary during detection, but the detection target object may move.
  • the object to be detected is a tumor
  • the tumor itself may move in conjunction with the movement of the surrounding involuntary muscles without moving.
  • the position of a tumor in the image moves as the camera moves in accordance with an endoscope operation by a doctor. Therefore, in the third embodiment, an image processing apparatus capable of outputting an object region at high speed even when a detection target object moves during detection will be described.
  • the image processing apparatus 300 includes an object region candidate generation unit 31, a heat map generation unit 32, and an object region output unit 33.
  • the image processed by the image processing apparatus 300 is a moving image (a series of temporally continuous images).
  • the basic operation of the heat map generation unit 32 is the same as that of the second embodiment, except that the output of the heat map generation unit 32 is the same as that of the heat map generation unit 32 as represented by the dotted arrows in FIG. To be input.
  • generation part 32 is utilized for the heat map production
  • the heat map generation unit 32 uses the heat map generated at a past time and the object region candidate corresponding to the current time generated by the object region candidate generation unit 31 in the image showing a certain tumor, and the current time A heat map corresponding to is generated.
  • the object region candidate generation unit 31 and the object region output unit 33 are the same as the object region candidate generation unit 21 and the object region output unit 23 of the second embodiment.
  • Step S301 is substantially the same as step S201 in FIG. 8, but the object region candidate generation unit 31 inputs an image to the object region candidate generation unit 31 at predetermined intervals (for example, 1 second). Alternatively, an image may be input when there is a change in the moving image.
  • step S302 when the object region candidate generated by the object region candidate generating unit 31 and the heat map of the past image are received, the heat map generating unit 32 generates a heat map based on these.
  • the heat map generating unit 32 generates a heat map based on these.
  • each object region candidate generated by the object region candidate generation unit 31 for the image at time t + 1 is displayed.
  • Rectangle t, n [x t, n , y t, n , w t, n , h t, n , conf t, n ] (Formula 2)
  • the heat map at time t Score (Rectangle t, n ) ⁇ t, w, h heatmap score (t, i, j) (Equation 3)
  • Formula (4) Can be expressed as
  • the object region candidate generation unit 31 generates a heat map at the current time based on the object region candidate for the image at the current time and the heat map generated from the image at the past time.
  • the similarity of the heat map image may be set to w t .
  • a similarity indicating a degree of coincidence between a heat map generated in the past and a heat map currently generated may be calculated and used as a weight. In this case, the degree of coincidence increases as there is no movement in the past and the present.
  • Step S303 is the same as step S203 in FIG.
  • the image processing apparatus 300 can display a frame indicating the entire irregular object at high speed even when the detection target object moves and is captured as a moving image. it can. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the heat map generation unit 32 generates a heat map using the heat map generated by the heat map generation unit 32 at the past time in addition to the object region candidate at the current time, and the object region output unit This is because 33 uses this heat map to select an object area from object area candidates.
  • the image processing apparatuses 100, 200, 300, and 400 may be integrated with an apparatus such as a camera that captures an image, or a camera and an information processing apparatus (such as a personal computer). May be connected to control the overall operation of the image processing on the information processing apparatus side. That is, an application program for image processing on the information processing apparatus side may be used under the following configuration.
  • the information processing apparatus 500 includes the following configuration as an example.
  • CPU Central Processing Unit
  • ROM Read Only Memory
  • RAM Random Access Memory
  • a storage device 505 for storing the program 504 and other data
  • a drive device 507 for reading / writing the recording medium 506
  • Communication interface 508 connected to the communication network 509
  • An input / output interface 510 for inputting / outputting data -Bus 511 connecting each component
  • Each component of the image processing apparatus in each embodiment of the present application is realized by the CPU 501 acquiring and executing a program 504 that realizes these functions.
  • a program 504 that realizes the functions of each component of the image processing apparatus (for example, the object region candidate generation unit 11, the heat map generation unit 12, and the object region output unit 13) is stored in advance in the storage device 505 or the RAM 503, for example.
  • the CPU 501 reads the data as necessary.
  • the program 504 may be supplied to the CPU 501 via the communication network 509 or may be stored in the recording medium 506 in advance, and the drive device 507 may read the program and supply it to the CPU
  • the image processing apparatus may be realized by an arbitrary combination of a separate information processing apparatus and a program for each component.
  • a plurality of components included in the image processing apparatus may be realized by an arbitrary combination of one information processing apparatus 500 and a program.
  • the constituent elements of the image processing apparatus are realized by other general-purpose or dedicated circuits, processors, or combinations thereof. These may be configured by a single chip or may be configured by a plurality of chips connected via a bus.
  • Some or all of the components of the image processing apparatus may be realized by a combination of the above-described circuit and the like and a program.
  • the plurality of information processing apparatuses and circuits may be centrally arranged or distributedly arranged. May be.
  • the information processing apparatus, the circuit, and the like may be realized as a form in which each is connected via a communication network, such as a client and server system and a cloud computing system.
  • the present invention can be applied to applications such as an image processing apparatus that recognizes a non-rigid object having an irregular and unpredictable shape included in an image, and a program for realizing the apparatus on a computer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Surgery (AREA)
  • Pathology (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Optics & Photonics (AREA)
  • Molecular Biology (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Image Analysis (AREA)
  • Endoscopes (AREA)

Abstract

The present invention provides an image processing device and the like capable of quickly displaying a frame indicating the whole of an irregular object. An image processing device 100 is provided with: an object area candidate generation unit 11 for generating a plurality of object area candidates indicating a position in an image where at least a part of an object to be detected exists; a heatmap generation unit 12 for generating a heatmap corresponding to the position in the image by using the generated object area candidates; and an object area output unit 13 for outputting an object area surrounding the whole of the object to be detected by using the generated heatmap, the object area being constituted of one or more object area candidates.

Description

画像処理装置、画像処理方法および記録媒体Image processing apparatus, image processing method, and recording medium
 本発明は、画像中における検知対象物体の確認を容易にするための画像処理装置等に関する。 The present invention relates to an image processing apparatus for facilitating confirmation of a detection target object in an image.
 動画像中における検知対象物体の発見を補助するための画像処理方法がある。非特許文献1および非特許文献2は、この画像処理方法の例を開示する。例えば、この画像処理方法は、物体領域候補生成手段および物体領域候補出力手段を備える装置において、以下のように動作する。即ち、物体領域候補生成手段は、画像(動画の場合、動画を構成する各フレーム画像)が入力されると、その画像中に含まれる検知対象の物体が存在すると思われる領域に複数の候補(物体領域候補)を生成する。物体領域候補とは、検知対象の物体を外接する矩形もしくは円であり、これらは物体の形状を示す変数(たとえば中心座標(x, y)、高さ(h)および幅(w))、更に、物体領域候補における検知対象物体の含有率等を示す信頼度等によって表現される。物体領域候補出力手段は、物体領域候補生成手段で生成された物体領域候補の中から、各物体領域候補の重なり具合又は信頼度等に基づいて、検知対象の物体に外接する矩形もしくは円を出力する。 There is an image processing method for assisting in finding a detection target object in a moving image. Non-Patent Document 1 and Non-Patent Document 2 disclose examples of this image processing method. For example, this image processing method operates as follows in an apparatus including an object region candidate generation unit and an object region candidate output unit. That is, when an image (in the case of a moving image, each frame image constituting the moving image) is input, the object region candidate generation unit is configured to output a plurality of candidates (in the region where the object to be detected included in the image is considered to exist) Object region candidate) is generated. The object region candidates are rectangles or circles circumscribing the object to be detected, and these are variables indicating the shape of the object (for example, center coordinates (x, y), height (h) and width (w)), and This is expressed by the reliability indicating the content of the detection target object in the object region candidate. The object area candidate output means outputs a rectangle or a circle circumscribing the object to be detected based on the degree of overlap or reliability of each object area candidate from the object area candidates generated by the object area candidate generation means. To do.
 その他、特許文献1は、本願に関連する技術を開示する。 In addition, Patent Document 1 discloses a technique related to the present application.
特開2016-206995号公報JP 2016-206995 A
 非特許文献1および非特許文献2に開示される画像処理方法においては、検知対象物体は、規則的形状であり、ある程度予測可能な物体である。例えば、人の顔等のように、検知対象物体の大きさ、形状等がある程度規則的に予測され、画像内の中心座標(x, y)から所定の高さ(h)および幅(w)に収まる物体である。しかしながら、検知対象物体には、その大きさ、形状等が不規則で予測できず、画像内の中心座標から予め設定される所定の高さおよび幅に収まらない可能性のある物体(以下、これを不規則物体と称呼する)もある。例えば、臓器内外の病変(腫瘍など)のように、一定の形状に収まらず、病状進行のステージに伴い、様々な形状に不規則に変形可能な非剛体などである。この場合、非特許文献1および非特許文献2に開示される画像処理方法を用いて、物体全体に外接する物体領域を判定し、出力することは困難となる。これは、非特許文献1および非特許文献2に開示される画像処理方法においては、不規則物体の形状では、物体と少なくとも一部が外接する物体領域候補が多数生成されるからである。この理由は、一般的に、ひとつの物体領域候補は、その領域内においてできるだけ検知対象物体を含み、かつ、できるだけ検知対象物体以外のものを含まないように設定されるからであり、この結果、1つの不規則物体に対し、複数の物体領域候補はその一部が重なり、連なって生成されることとなる。例えば、検知対象物体が丸型であれば、その重心を中心とする1つの物体領域候補で足りることが多いが、検知対象物体がヒトデ型であれば、複数の物体領域候補(その重心を中心とする中心近辺をカバーする物体領域候補と、当該ヒトデ型に含まれる複数の足部をカバーする複数の物体領域候補)にて表わされることが多い。更に、後者の物体領域候補は、上記設定により検知対象物体を高い比率にて含んでいるため、信頼度が高くなる傾向がある。これらの理由により、検知対象が不規則物体の場合は、非特許文献1および非特許文献2に開示される画像処理方法においては、不規則物体の全体にて外接する(又は、検知対象物体の全体を含む)矩形もしくは円の領域候補は出力されにくい。 In the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2, the detection target object has a regular shape and is an object that can be predicted to some extent. For example, the size, shape, etc. of the detection target object, such as a human face, is regularly predicted to some extent, and the predetermined height (h) and width (w) from the center coordinates (x, y) in the image It is an object that fits in. However, the object to be detected cannot be predicted irregularly in its size, shape, etc., and may not fit within a predetermined height and width preset from the center coordinates in the image (hereinafter referred to as this). Are called irregular objects). For example, it is a non-rigid body that does not fit in a certain shape, such as a lesion inside or outside an organ (such as a tumor), and can be irregularly deformed into various shapes as the disease progresses. In this case, it is difficult to determine and output an object region circumscribing the entire object using the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2. This is because the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2 generate a large number of object region candidates that are at least partially circumscribed with the object in the shape of an irregular object. This is because, generally, one object region candidate is set so as to include the detection target object as much as possible in the region and not to include anything other than the detection target object as much as possible. As a result, For one irregular object, a plurality of object region candidates partially overlap and are generated in succession. For example, if the detection target object is a round shape, a single object region candidate centered on the center of gravity is often sufficient, but if the detection target object is a starfish type, a plurality of object region candidates (centered on the center of gravity) are sufficient. Are often represented by object region candidates covering the vicinity of the center and a plurality of object region candidates covering a plurality of feet included in the starfish type). Further, since the latter object region candidate includes the detection target objects at a high ratio by the above setting, the reliability tends to be high. For these reasons, when the detection target is an irregular object, the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2 circumscribe the entire irregular object (or the detection target object). Rectangle or circle area candidates (including the whole) are difficult to output.
 これに伴う問題として、非特許文献1および非特許文献2に開示される画像処理方法において、1つの不規則な検知対象物体に対し、複数の物体領域候補の一部が重なり連なって生成されると、この検知対象物体を観察するオペレータ(例えば、病変の場合は「医師」)にとって、物体領域候補を表わす複数重なった枠によって検知対象物体の視認が困難となり、当該検知対象物体の一部または全体の形状の判断がしづらいという問題がある。または、物体領域候補を表わす枠が何らかの理由により適切に検知対象物体を表わしていない場合に、オペレータによる観察が当該枠にミスリードされ、発見されるべき検知対象物体を見過ごす等の問題がある。 As a problem associated with this, in the image processing methods disclosed in Non-Patent Document 1 and Non-Patent Document 2, a part of a plurality of object region candidates are generated by overlapping one irregular detection target object. For the operator (for example, “doctor” in the case of a lesion) observing the detection target object, it is difficult to visually recognize the detection target object due to the multiple overlapping frames representing the object region candidates. There is a problem that it is difficult to judge the overall shape. Or, when the frame representing the object region candidate does not appropriately represent the detection target object for some reason, there is a problem that the observation by the operator is misleaded to the frame and the detection target object to be discovered is overlooked.
 非特許文献1に示された技術では、各領域候補の重なりおよび信頼度に基づき領域候補が出力されるため、検知対象物体と外接する物体領域候補が他の物体領域候補よりも高い信頼度になる。しかしながら、不規則な物体と外接する物体領域候補内では、物体ではない領域を多く含むこととなるから、一般的に部分的に外接する矩形と比べて信頼度が低くなってしまい、ひいては正確な物体領域候補が抽出できないという問題が生じる。 In the technique disclosed in Non-Patent Document 1, since region candidates are output based on the overlap and reliability of each region candidate, the object region candidate circumscribing the detection target object has higher reliability than other object region candidates. Become. However, the candidate object region circumscribing an irregular object contains many non-object regions, which generally results in lower reliability than a partially circumscribing rectangle, and thus accurate. There arises a problem that an object region candidate cannot be extracted.
 非特許文献2に開示される画像処理方法においては、多大な計算量を必要とするニューラルネットワークを用いて画像の物体領域を生成または再生成するため、高速な画像処理が困難であるという問題がある。 In the image processing method disclosed in Non-Patent Document 2, a problem is that high-speed image processing is difficult because an object region of an image is generated or regenerated using a neural network that requires a large amount of calculation. is there.
 そこで、本発明は、上述した課題に鑑み、不規則物体の全体を示す枠を高速に表示する画像処理装置等を提供することを目的とする。 Therefore, in view of the above-described problems, an object of the present invention is to provide an image processing apparatus or the like that displays a frame indicating the entire irregular object at high speed.
 上記問題点を鑑みて、本発明の第1の観点である画像処理装置は、
 画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成する物体領域候補生成部と、
生成された物体領域候補を用いて、画像中における位置に対応するヒートマップを生成するヒートマップ生成部と、
 生成されたヒートマップを用いて、1つ以上の物体領域候補から成る、検知対象物体の全体を囲う物体領域を出力する物体領域出力部
とを備える。
In view of the above problems, the image processing apparatus according to the first aspect of the present invention provides:
An object region candidate generation unit that generates a plurality of object region candidates indicating positions where at least a part of the detection target object exists in the image;
Using the generated object region candidate, a heat map generation unit that generates a heat map corresponding to the position in the image,
And an object region output unit that outputs an object region that surrounds the entire detection target object, which is made up of one or more object region candidates, using the generated heat map.
 本発明の第2の観点である画像処理方法は、
 画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成し、
 生成された物体領域候補を用いて、画像中における位置に対応するヒートマップを生成し、
 生成されたヒートマップを用いて、1つ以上の物体領域候補から成る、検知対象物体の全体を囲う物体領域を出力する
ことを備える。
An image processing method according to a second aspect of the present invention includes:
Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image;
Using the generated object region candidate, generate a heat map corresponding to the position in the image,
Using the generated heat map to output an object region that includes one or more object region candidates and surrounds the entire detection target object.
 本発明の第3の観点である画像処理プログラムは、
 画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成し、
 生成された物体領域候補を用いて、画像中における位置に対応するヒートマップを生成し、
 生成されたヒートマップを用いて、1つ以上の物体領域候補から成る、検知対象物体の全体を囲う物体領域を出力する
ことをコンピュータに実現させるための画像処理プログラムである。
An image processing program according to a third aspect of the present invention is:
Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image;
Using the generated object region candidate, generate a heat map corresponding to the position in the image,
An image processing program for causing a computer to output an object region that surrounds the entire detection target object, which is composed of one or more object region candidates, using a generated heat map.
 画像処理プログラムは、非一時的なコンピュータ可読の記憶媒体に格納されていてもよい。 The image processing program may be stored in a non-transitory computer-readable storage medium.
 本発明によれば、不規則物体の全体を示す枠を高速に表示する画像処理装置等を提供することができる。 According to the present invention, it is possible to provide an image processing apparatus or the like that displays a frame indicating the entire irregular object at high speed.
本発明の第1の実施形態にかかる画像処理装置の構成例を示すブロック図である。1 is a block diagram illustrating a configuration example of an image processing apparatus according to a first embodiment of the present invention. 本発明の第1の実施形態にかかる画像処理装置における動作を示すフローチャートである。3 is a flowchart showing an operation in the image processing apparatus according to the first embodiment of the present invention. 画像内における物体領域候補を示す図である。It is a figure which shows the object area | region candidate in an image. 画像に対応するヒートマップを示す図である。It is a figure which shows the heat map corresponding to an image. 画像内における物体領域を示す図である。It is a figure which shows the object area | region in an image. 画像内における物体領域を検知対象物体と共に示す図である。It is a figure which shows the object area | region in an image with a detection target object. 本発明の第2の実施形態にかかる画像処理装置の構成例を示すブロック図である。It is a block diagram which shows the structural example of the image processing apparatus concerning the 2nd Embodiment of this invention. 本発明の第2の実施形態にかかる画像処理装置における動作を示すフローチャートである。It is a flowchart which shows operation | movement in the image processing apparatus concerning the 2nd Embodiment of this invention. 画像に対応するヒートマップおよび物体領域候補を示す図である。It is a figure which shows the heat map and object area | region candidate corresponding to an image. 画像内における物体領域を検知対象物体と共に示す図である。It is a figure which shows the object area | region in an image with a detection target object. 本発明の第3の実施形態にかかる画像処理装置の構成例を示すブロック図である。It is a block diagram which shows the structural example of the image processing apparatus concerning the 3rd Embodiment of this invention. 本発明の第3の実施形態にかかる画像処理装置における動作を示すフローチャートである。It is a flowchart which shows operation | movement in the image processing apparatus concerning the 3rd Embodiment of this invention. 各実施形態において適用可能な情報処理装置の構成例を示すブロック図である。It is a block diagram which shows the structural example of the information processing apparatus applicable in each embodiment.
 以下の実施形態においては、主に、形状が不規則的で物体領域が予測不可能な検知対象物体を対象とする画像処理装置等について説明する。尚、この画像処理装置は、形状が規則的で物体領域が予測可能な物体も検知対象とできることは勿論である。以下の説明において、不規則な検知対象物体とは、非剛体、一例として、臓器内外に発生する病変(腫瘍)とする。この場合、画像処理装置の使用者は医師等の医療関係者となる。 In the following embodiment, an image processing apparatus or the like that mainly targets a detection target object having an irregular shape and an unpredictable object region will be described. Of course, the image processing apparatus can also detect an object having a regular shape and a predictable object region. In the following description, an irregular detection target object is a non-rigid body, for example, a lesion (tumor) that occurs inside or outside an organ. In this case, the user of the image processing apparatus is a medical person such as a doctor.
 この画像処理装置は、例えば、内視鏡動画像におけるリアルタイム腫瘍検知システム等に組み込まれる。当該内視鏡を介し撮影される動画または静止画(以下単に画像と記載)において検知された病変がある場合、医師は当該病変が発生している臓器内外の全部又は一部の形状が正常時と比して変異しているかを観察することで、腫瘍の変異タイプ、悪性度および進行度等を判断する。又、医師が、腫瘍の摘出(切除)を行うにあたり、ある臓器における腫瘍部分と正常な部分の境界を正確に認識する必要がある。このため、本発明の実施形態における画像処理装置は、腫瘍の位置を示す情報として、腫瘍全体に外接する枠を的確かつ高速に表示する。これにより医師による診断および処置が素早く正確になるよう補助することができる。 This image processing apparatus is incorporated in, for example, a real-time tumor detection system for endoscopic moving images. When there is a lesion detected in a moving image or still image (hereinafter simply referred to as an image) taken through the endoscope, the doctor can check that all or part of the internal or external organ where the lesion occurs is normal. By observing whether or not there is a mutation, the mutation type, malignancy and progression of the tumor are determined. In addition, when a doctor removes (removes) a tumor, it is necessary to accurately recognize the boundary between a tumor portion and a normal portion in a certain organ. For this reason, the image processing apparatus in the embodiment of the present invention displays a frame circumscribing the entire tumor accurately and at high speed as information indicating the position of the tumor. Thereby, it is possible to assist the diagnosis and treatment by the doctor quickly and accurately.
 以下、図面を参照して、本発明の各実施形態を説明する。以下の図面の記載において、同一又は類似の部分には同一又は類似の符号を付している。ただし、図面は本発明の実施形態における構成を概略的に表している。更に以下に記載される本発明の実施形態は一例であり、その本質を同一とする範囲において適宜変更可能である。 Hereinafter, embodiments of the present invention will be described with reference to the drawings. In the following description of the drawings, the same or similar parts are denoted by the same or similar reference numerals. However, the drawings schematically show the configuration in the embodiment of the present invention. Furthermore, the embodiment of the present invention described below is an example, and can be appropriately changed within a range in which the essence is the same.
 <第1の実施形態>
 (画像処理装置)
 本発明の第1の実施形態に係る画像処理装置100は、図1に示すように、物体領域候補生成部11、ヒートマップ生成部12、および、物体領域出力部13を備える。
<First Embodiment>
(Image processing device)
As illustrated in FIG. 1, the image processing apparatus 100 according to the first embodiment of the present invention includes an object region candidate generation unit 11, a heat map generation unit 12, and an object region output unit 13.
 物体領域候補生成部11は、画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成する。具体的に、物体領域候補生成部11は、外部カメラ(図1には不図示)、例えば内視鏡カメラ等にて撮影された画像を入力し、入力された画像内における検知対象物体が存在する領域の候補(物体領域候補)を複数生成する。画像中に含まれる検知対象の物体が存在すると思われる領域に複数の物体領域候補を生成する手法としては、非特許文献1および非特許文献2等に記載の既存の手法を用いてよい。物体領域候補の数が多すぎる場合、物体領域候補生成部11は、当該物体領域候補としてふさわしい幾つかの当該物体領域候補の組み合わせを生成し、その1つを物体領域候補(群)として出力してもよい。 The object area candidate generation unit 11 generates a plurality of object area candidates indicating positions where at least a part of the detection target object exists in the image. Specifically, the object region candidate generation unit 11 inputs an image captured by an external camera (not shown in FIG. 1), for example, an endoscope camera, and there is a detection target object in the input image. A plurality of region candidates (object region candidates) are generated. As a method for generating a plurality of object region candidates in a region where an object to be detected included in an image is supposed to exist, existing methods described in Non-Patent Document 1, Non-Patent Document 2, and the like may be used. When there are too many object area candidates, the object area candidate generation unit 11 generates some combinations of the object area candidates suitable as the object area candidates, and outputs one of them as an object area candidate (group). May be.
 ヒートマップ生成部12は、生成された物体領域候補を用いて、画像中における位置に対応するヒートマップを生成する。具体的に、ヒートマップ生成部12は、物体領域候補を含む画像を1つ以上入力し、各物体領域候補の信頼度に基づき、当該画像に関するヒートマップを生成する。ここで、信頼度とは、物体領域候補内に検知対象物体が存在する確率を示す値であり、例えば、各物体領域候補の重なり具合、又は、ある物体領域候補における検知対象物体の含有率等を示す度合いを基に算出される。また、ヒートマップとは、画像内の区画(例えばピクセル)の各々の値を示すデータ行列等を、色として可視的に二次元で表現した図のことを指す。ヒートマップでは、物体が存在する領域においては信頼度のスコアが高くなるように設定される。 The heat map generation unit 12 generates a heat map corresponding to the position in the image using the generated object region candidate. Specifically, the heat map generation unit 12 inputs one or more images including object region candidates, and generates a heat map related to the image based on the reliability of each object region candidate. Here, the reliability is a value indicating the probability that the detection target object exists in the object region candidate. For example, the degree of overlap of each object region candidate or the content rate of the detection target object in a certain object region candidate, etc. It is calculated based on the degree indicating Further, the heat map refers to a diagram in which a data matrix or the like indicating each value of a section (for example, pixel) in an image is visibly expressed as a color in two dimensions. In the heat map, the reliability score is set to be high in the region where the object exists.
 物体領域出力部13は、生成されたヒートマップを入力し、当該ヒートマップを基に、当該画像内における、不規則な検知対象物体の全体を示す物体領域(矩形枠)を出力する。具体的に、物体領域出力部13は、ヒートマップを区分した複数の区画毎(例えばピクセル毎)において、当該区画に紐付けられた前記信頼度が所定の閾値以上である区画の集合を含む物体領域候補を、前記物体領域として出力する。 The object region output unit 13 inputs the generated heat map, and outputs an object region (rectangular frame) indicating the entire irregular detection target object in the image based on the heat map. Specifically, the object region output unit 13 includes, for each of a plurality of sections (for example, for each pixel) into which the heat map is divided, an object including a set of sections whose reliability is equal to or higher than a predetermined threshold. A region candidate is output as the object region.
 (画像処理装置の動作)
 図2に示すフローチャートを参照して、画像処理装置100の具体的な動作について説明する。
(Operation of image processing device)
The specific operation of the image processing apparatus 100 will be described with reference to the flowchart shown in FIG.
 先ずステップS101において、画像処理装置100は、画像を物体領域候補生成部11に入力し、物体領域候補を生成する。このステップの具体的な一動作例を、以下に説明する。 First, in step S101, the image processing apparatus 100 inputs an image to the object region candidate generation unit 11, and generates an object region candidate. A specific operation example of this step will be described below.
 画像処理装置100が、例えば、内視鏡画像中の病変(腫瘍)を認識するための装置であるとすると、物体領域候補生成部11は、腫瘍を含む画像を受け付け、この画像に対して、腫瘍が含まれている領域を示す物体領域候補を生成する。高速な物体認識を実現する一手法として、物体領域として検知対象の物体を外接する矩形を用いる手法を採用するとする。この場合、描画関数(Rectangle)において、矩形を表現する変数として、中心座標(x、y)、幅(w)、高さ(h)、矩形内に物体を含む確からしさを示す信頼度(conf)を用いるものとする。この場合、あるn番目(nは正の整数)の物体領域候補(Rectangle)は式1の関数で表すことができる。
Rectangle = [x, y, w, h, conf] …(式1)
 本例では、物体領域は、検知対象物体の全体の外縁(外周)と接するように囲った(以下、この状態を外接とも称呼する)矩形を例とするが、これは矩形に限られず、例えば円でもよい。また物体領域の外接ではなく、検知対象物体の全体の内周と接するように囲ったものでもよい。また物体の重心を中心座標としてもよい。また、回転を表わす変数を加えて用いてもよい。
If the image processing apparatus 100 is, for example, an apparatus for recognizing a lesion (tumor) in an endoscopic image, the object region candidate generation unit 11 receives an image including a tumor, and for this image, An object region candidate indicating a region including a tumor is generated. As a technique for realizing high-speed object recognition, a technique that uses a rectangle that circumscribes an object to be detected as an object region is adopted. In this case, in the drawing function (Rectangle), the central coordinates (x, y), width (w), height (h), and the reliability (conf ) Shall be used. In this case, a certain nth (n is a positive integer) object region candidate (Rectangle) can be expressed by the function of Equation 1.
Rectangle n = [x n, y n, w n, h n, conf n] ... ( Equation 1)
In this example, the object region is an example of a rectangle surrounded so as to be in contact with the entire outer edge (outer periphery) of the detection target object (hereinafter, this state is also referred to as circumscribing), but this is not limited to a rectangle. Yen may be used. Further, the object region may be enclosed so as to be in contact with the entire inner periphery of the detection target object, instead of circumscribing the object region. The center of gravity of the object may be used as the center coordinate. Further, a variable representing rotation may be added and used.
 図3は、複数の変数(n個)を式(1)の関数に入力し得られた、複数の物体領域候補を、ヒートマップ上に表わした例を示す。図3中の太線枠が検知対象物体(腫瘍)の全体と接するように囲った(以下、この状態を外接とも称呼する)矩形領域候補であり、それ以外の矩形は検知対象物体の一部を示す矩形領域候補となっている。矩形領域候補は、その領域内においてできるだけ検知対象物体を含み、かつ、できるだけ検知対象物体以外のものを含まないように設定される。 FIG. 3 shows an example in which a plurality of object region candidates obtained by inputting a plurality of variables (n) to the function of Expression (1) are represented on a heat map. 3 is a rectangular area candidate surrounded by the thick line frame so as to be in contact with the entire detection target object (tumor) (hereinafter, this state is also referred to as circumscribing), and other rectangles indicate a part of the detection target object. This is a rectangular area candidate shown. The rectangular area candidates are set so as to include the detection target object as much as possible in the area and not to include anything other than the detection target object as much as possible.
 ステップS102においては、生成された複数の物体領域候補を受け取り、ヒートマップ生成部12は、これらの複数の物体領域候補を基にヒートマップを生成する。このステップの具体的な一動作例を、以下に説明する。 In step S102, the generated plurality of object region candidates are received, and the heat map generation unit 12 generates a heat map based on the plurality of object region candidates. A specific operation example of this step will be described below.
 入力されたn個の各物体領域候補、即ち、Rectangle = [x, y, w, h, conf]、Rectangle = [x, y, w, h, conf]、…、Rectangle = [x, y, w, h, conf]は、領域候補内に物体が存在する確からしさを示す信頼度conf、conf、…、confを備える。この信頼度を、画像中の物体領域候補内のヒートマップの各ピクセルの値と紐付ける。即ち、ヒートマップの各ピクセルの値を conf = heatmapscore(i,j)と設定する。ここで(i,j)はヒートマップにおけるピクセルの座標位置を表す。これにより、画像を構成する各ピクセルにおいて信頼度を持つヒートマップが生成される。 Each of n input object region candidates, that is, Rectangle 1 = [x 1 , y 1 , w 1 , h 1 , conf 1 ], Rectangle 2 = [x 2 , y 2 , w 2 , h 2 , conf 2], ..., Rectangle n = [x n, y n, w n, h n, conf n] is confidence conf 1, conf 2 indicating the likelihood that an object exists in the area candidate, ..., conf n Is provided. This reliability is associated with the value of each pixel of the heat map in the object region candidate in the image. That is, the value of each pixel of the heat map is set as conf n = heatmap score (i, j) . Here, (i, j) represents the coordinate position of the pixel in the heat map. Thereby, a heat map having reliability in each pixel constituting the image is generated.
 図4は、ヒートマップ生成部12で生成したヒートマップの例を示す。本例においては、ヒートマップの各物体領域候補に含まれる複数のピクセルの各々の信頼度confの値を用いて、その物体領域候補の値を設定する。例えば、ある物体領域候補の信頼度confの値を、当該物体領域候補内に存在するピクセルの値とする。尚、複数の物体領域候補、例えば、物体領域候補AとBとが一部重なり合った部分(例えばこの部分をCとする)がある場合、Cの部分の信頼度の値は、物体領域候補AおよびBの信頼度を加算した値であってもよい。 FIG. 4 shows an example of a heat map generated by the heat map generator 12. In this example, the value of the object region candidate is set using the reliability conf n value of each of the plurality of pixels included in each object region candidate of the heat map. For example, the value of the reliability conf n of a certain object region candidate is set as the value of a pixel existing in the object region candidate. When there are a plurality of object region candidates, for example, a portion where the object region candidates A and B partially overlap (for example, this portion is referred to as C), the reliability value of the portion C is the object region candidate A And a value obtained by adding the reliability of B.
 ヒートマップが、ピクセル値が高いほど色が白くなるように設定されている場合、図3にて多数の物体領域候補が確認される画像の左下側の物体領域候補のピクセルの値が大きくなり、その結果、図4に示すように画像が白っぽくなる。 When the heat map is set so that the higher the pixel value is, the whiter the color is, the value of the pixel of the object region candidate on the lower left side of the image in which many object region candidates are confirmed in FIG. As a result, the image becomes whitish as shown in FIG.
 最後にステップS103において、生成されたヒートマップを取得した物体領域出力部13は、複数の物体領域候補の中から検知対象物体の全体と外接する物体領域候補を選択し、物体領域として出力する。 Finally, in step S103, the object area output unit 13 that has acquired the generated heat map selects an object area candidate that circumscribes the entire detection target object from among a plurality of object area candidates, and outputs the object area candidate as an object area.
 物体領域出力部13の動作について詳細に説明する。入力されたヒートマップ画像(図4)は各ピクセルにおいて物体が存在する確からしさ(信頼度)を示す値を持つ。ここで、その値が所定の閾値以上のピクセルには物体が存在するとし、当該閾値より小さいピクセルには物体が存在しないと判断する。閾値は設計者によって予め設定されていても良いし、オペレータにより変更可能に設定されていてもよい。この場合、検知対象物体が存在すると判断されたヒートマップは図5に示すもののようになる。尚、検知対象物体(腫瘍)は図6に示すように存在しているものとする。このとき、物体領域出力部13は、この物体が存在する領域の全てを囲う枠(図5、図6の枠1)を物体領域の枠として出力する。出力された画像は、外部のモニタ等に映し出され、医師等のオペレータに閲覧される。尚、出力される物体領域は、検知対象物体の形状に沿った枠(矩形枠)として出力する。 The operation of the object area output unit 13 will be described in detail. The input heat map image (FIG. 4) has a value indicating the probability (reliability) that an object exists in each pixel. Here, it is determined that an object exists in a pixel whose value is equal to or greater than a predetermined threshold value, and no object exists in a pixel smaller than the threshold value. The threshold value may be set in advance by the designer, or may be set to be changeable by the operator. In this case, the heat map determined that the detection target object exists is as shown in FIG. It is assumed that the detection target object (tumor) exists as shown in FIG. At this time, the object area output unit 13 outputs a frame (frame 1 in FIGS. 5 and 6) surrounding all of the area where the object exists as a frame of the object area. The output image is displayed on an external monitor or the like and viewed by an operator such as a doctor. The output object region is output as a frame (rectangular frame) along the shape of the detection target object.
 (第1の実施形態の効果)
 本発明の第1の実施形態における画像処理装置100は、不規則物体の全体を示す枠を高速に表示することができる。ひいてはオペレータによる不規則物体の発見および判断を容易にすることができる。この理由は、画像処理装置100が、各物体領域候補の信頼度に基づくヒートマップを用いて各物体領域を出力するため、検知対象物体の全体に外接する物体領域を正しく出力するからである。また、物体領域候補生成部11で生成される物体領域候補の信頼度から生成されるヒートマップ内のピクセルの値と予め設定される閾値とを比較して物体領域を出力するため、ニューラルネットワーク計算量と比して、遥かに少ない計算量で、高速に物体領域を出力するからである。
(Effects of the first embodiment)
The image processing apparatus 100 according to the first embodiment of the present invention can display a frame indicating the entire irregular object at high speed. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the image processing apparatus 100 outputs each object region using a heat map based on the reliability of each object region candidate, and thus correctly outputs the object region circumscribing the entire detection target object. Further, since the object region is output by comparing the pixel value in the heat map generated from the reliability of the object region candidate generated by the object region candidate generating unit 11 with a preset threshold value, the neural network calculation is performed. This is because the object region is output at high speed with a much smaller calculation amount than the amount.
 <第2の実施形態>
 本発明の第1の実施形態においては、画像処理装置100はヒートマップを基に検知対象物体の物体領域を出力したが、当該物体領域の出力にはヒートマップとこれに対応する物体領域候補とを同時に用いても良い。また、第1の実施形態の物体領域候補生成部11が物体領域候補を生成する処理において、当該物体領域候補の数が多すぎる場合、当該物体領域候補の組み合わせの計算に多くの処理時間を要していた。そこで本発明の第2の実施形態においては、上記の生成する処理を簡略化して、ヒートマップおよびその物体領域候補を用いて検知対象物体の物体領域を出力する手法について説明する。
<Second Embodiment>
In the first embodiment of the present invention, the image processing apparatus 100 outputs the object region of the detection target object based on the heat map, but the output of the object region includes a heat map and a corresponding object region candidate. May be used simultaneously. Further, in the process in which the object region candidate generation unit 11 of the first embodiment generates object region candidates, if the number of the object region candidates is too large, it takes a lot of processing time to calculate the combination of the object region candidates. Was. Therefore, in the second embodiment of the present invention, a method of outputting the object region of the detection target object by using the heat map and the object region candidate will be described by simplifying the processing to be generated.
 (画像処理装置)
 本発明の第2の実施形態に係る画像処理装置200は、図7に示すように、物体領域候補生成部21、ヒートマップ生成部22、および、物体領域出力部23を備える。
(Image processing device)
As shown in FIG. 7, the image processing apparatus 200 according to the second embodiment of the present invention includes an object region candidate generation unit 21, a heat map generation unit 22, and an object region output unit 23.
 物体領域出力部23は、物体領域候補生成部21が生成した物体領域候補およびそれに対応するヒートマップ生成部22が生成したヒートマップ(生成された物体領域候補が重畳されたヒートマップ)を用いて、複数の物体領域候補の中から検知対象物体の全体と外接する物体領域候補を選択し、物体領域として出力する。 The object region output unit 23 uses the object region candidate generated by the object region candidate generation unit 21 and the heat map generated by the heat map generation unit 22 corresponding to the object region candidate (heat map on which the generated object region candidates are superimposed). Then, an object region candidate circumscribing the entire detection target object is selected from among a plurality of object region candidates and output as an object region.
 物体領域候補生成部21は、物体領域候補を生成するが、第1の実施形態(物体領域候補生成部11)と異なる点は、当該物体領域候補の数が多すぎる場合に、当該物体領域候補の組み合わせを生成しないことである。生成された物体領域候補は、当該物体領域候補をヒートマップ生成部22および物体領域出力部23に出力する。 The object region candidate generation unit 21 generates an object region candidate. However, the object region candidate generation unit 21 is different from the first embodiment (object region candidate generation unit 11) when the number of object region candidates is too large. The combination of is not generated. The generated object area candidate outputs the object area candidate to the heat map generation unit 22 and the object area output unit 23.
 ヒートマップ生成部22は第1の実施形態(ヒートマップ生成部12)と同様である。 The heat map generator 22 is the same as that of the first embodiment (heat map generator 12).
 (画像処理装置の動作)
 図8に示すフローチャートを参照して、画像処理装置200の具体的な動作について説明する。
(Operation of image processing device)
The specific operation of the image processing apparatus 200 will be described with reference to the flowchart shown in FIG.
 ステップS201およびS202については、夫々、図2のステップS101およびS102と同様である。 Steps S201 and S202 are the same as steps S101 and S102 in FIG. 2, respectively.
 ステップS203において、物体領域出力部23は、物体領域候補生成部21から物体領域候補を、ヒートマップ生成部22からヒートマップを取得し、これらを基に、物体領域を出力する。このステップの具体的な一動作例を、以下に説明する。 In step S203, the object region output unit 23 acquires the object region candidate from the object region candidate generation unit 21 and the heat map from the heat map generation unit 22, and outputs the object region based on these. A specific operation example of this step will be described below.
 まず、物体領域候補生成部21で生成された物体領域候補(Rectangle = [x, y, w, h, conf])と、ヒートマップ生成部22で生成されたヒートマップ(図4参照)が、物体領域出力部23に入力される。図9は、ヒートマップ(図4参照)上に描かれた複数の物体領域候補(枠)を示す。物体領域出力部23は、各物体領域候補(Rectangle = [x, y, w, h, conf]、Rectangle = [x, y, w, h, conf]、…、Rectangle = [x, y, w, h, conf])の物体領域内のピクセルに対応する、ヒートマップの各ピクセルの値(信頼度)を基に、各物体領域候補のスコアを、集計のための関数を用いて計算する。例えば、入力された物体領域候補(Rectangle)のスコアを、当該物体領域内に対応するヒートマップの各ピクセルの信頼度の和とする場合、Rectangleを変数とする関数を用いて、Score(Rectangle)=Σ(w,h)heatmapscore(i,j)とする。この後、物体領域出力部23は、物体領域候補のスコア(物体領域候補内の全ピクセルの信頼度の和)と予め設定された閾値とを比較し、当該閾値以上のスコアとなる物体領域候補を物体領域として出力する。物体領域出力部23は、物体領域候補の中から、物体領域候補のスコアが閾値以上となる物体領域候補を抽出し、それらの外縁で囲われた領域を物体領域として出力する。例えば、図10は、抽出された2つの物体領域候補2a、2bから構成される物体領域2cを表わしている。尚、物体領域候補2a、2bは、必ず1つにまとめずとも、オペレータが容易に視認可能な数であれば(例えば図10では2つ)、そのまま表示しても良い。又、閾値は設計者によって予め設定されていても良いし、オペレータによって変更可能に設定されていても良い。 First, object region candidates (Rectangle n = [x n , y n , w n , h n , conf n ]) generated by the object region candidate generation unit 21, and a heat map ( 4) is input to the object region output unit 23. FIG. 9 shows a plurality of object region candidates (frames) drawn on the heat map (see FIG. 4). The object region output unit 23 outputs each object region candidate (Rectangle 1 = [x 1 , y 1 , w 1 , h 1 , conf 1 ], Rectangle 2 = [x 2 , y 2 , w 2 , h 2 , conf 2]. ], ..., Rectangle n = [x n , y n , w n , h n , conf n ])), based on the value (reliability) of each pixel in the heat map corresponding to the pixel in the object area The score of the object region candidate is calculated using a function for tabulation. For example, when the score of the input object region candidate (Rectangle n ) is the sum of the reliability of each pixel of the heat map corresponding to the object region, a score (Rectangle) n ) = Σ (w, h) heatmap score (i, j) . Thereafter, the object region output unit 23 compares the score of the object region candidate (sum of reliability of all the pixels in the object region candidate) with a preset threshold value, and the object region candidate having a score equal to or higher than the threshold value. Is output as the object region. The object area output unit 23 extracts object area candidates whose object area candidate score is equal to or greater than a threshold value from the object area candidates, and outputs the area surrounded by the outer edges as the object area. For example, FIG. 10 shows an object region 2c composed of two extracted object region candidates 2a and 2b. It should be noted that the object region candidates 2a and 2b may be displayed as they are as long as they are numbers that can be easily visually recognized by the operator (for example, two in FIG. 10), without necessarily being combined into one. Further, the threshold value may be set in advance by the designer, or may be set to be changeable by the operator.
 (第2の実施形態の効果)
 本発明の第2の実施形態に係る画像処理装置200は、不規則物体の全体を示す枠を高速に表示することができる。ひいてはオペレータによる不規則物体の発見および判断を容易にすることができる。この理由は、画像処理装置200が、各物体領域候補の信頼度に基づくヒートマップのスコアを用いて各物体領域を出力するため、検知対象物体の全体に外接する物体領域を正しく出力するからである。また、物体領域候補生成部21で生成される物体領域候補の信頼度から生成されるヒートマップ内のピクセルの値と予め設定される閾値とを比較して物体領域を出力するため、ニューラルネットワーク計算量と比して、遥かに少ない計算量で、高速に物体領域を出力するからである。
 尚、第1の実施形態と比して第2の実施形態では、物体領域候補生成部21の処理が簡略化されたため、全体の処理時間を早めることができる。第2の実施形態ではヒートマップ生成部22がヒートマップからスコアを算出する集計処理が必要となるが、物体領域出力部23が、ヒートマップを用いて物体領域候補の中から物体領域を構成する物体領域候補の選択を行うことにより、物体領域候補生成部11における処理が簡略化される。このため、物体領域候補数が多い場合でも、物体領域候補生成部11の処理時間が長期化することを避けられ、ひいては、全体としての処理時間を短縮することができる。
(Effect of 2nd Embodiment)
The image processing apparatus 200 according to the second embodiment of the present invention can display a frame indicating the entire irregular object at high speed. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the image processing apparatus 200 outputs each object region using the heat map score based on the reliability of each object region candidate, and thus correctly outputs the object region circumscribing the entire detection target object. is there. Further, since the object region is output by comparing the pixel value in the heat map generated from the reliability of the object region candidate generated by the object region candidate generating unit 21 with a preset threshold value, the neural network calculation is performed. This is because the object region is output at high speed with a much smaller calculation amount than the amount.
In the second embodiment, the processing of the object region candidate generation unit 21 is simplified in comparison with the first embodiment, so that the entire processing time can be shortened. In the second embodiment, the heat map generation unit 22 needs to perform aggregation processing for calculating the score from the heat map, but the object region output unit 23 configures the object region from the object region candidates using the heat map. By selecting the object region candidate, the processing in the object region candidate generation unit 11 is simplified. For this reason, even when the number of object region candidates is large, it is possible to avoid the processing time of the object region candidate generating unit 11 from being prolonged, and as a result, the processing time as a whole can be shortened.
 <第3の実施形態>
 第1および第2の実施形態においては、主に、検知中に検知対象物体が静止している場合を想定しているが、検知対象物体は動くこともある。例えば検知対象物体が腫瘍の場合、腫瘍そのものは動かなくても周囲の不随意筋の動きに連動して動く場合もある。また、本発明の一利用例となる内視鏡動画像におけるリアルタイム腫瘍検知システムでは、医師による内視鏡操作に伴いカメラが動くことによって、画像中の腫瘍位置が動く。よって、第3の実施形態においては、検知中に検知対象物体が動く場合であっても高速に物体領域を出力可能な画像処理装置について説明する。
<Third Embodiment>
In the first and second embodiments, it is mainly assumed that the detection target object is stationary during detection, but the detection target object may move. For example, when the object to be detected is a tumor, the tumor itself may move in conjunction with the movement of the surrounding involuntary muscles without moving. In the real-time tumor detection system in an endoscope moving image that is an example of use of the present invention, the position of a tumor in the image moves as the camera moves in accordance with an endoscope operation by a doctor. Therefore, in the third embodiment, an image processing apparatus capable of outputting an object region at high speed even when a detection target object moves during detection will be described.
 (画像処理装置)
 本発明の第3の実施形態に係る画像処理装置300は、図11に示すように、物体領域候補生成部31、ヒートマップ生成部32、および、物体領域出力部33を備える。尚、本実施形態においては、画像処理装置300が処理する画像は動画像(時間的に連続した一連の画像)であることを前提とする。
(Image processing device)
As shown in FIG. 11, the image processing apparatus 300 according to the third embodiment of the present invention includes an object region candidate generation unit 31, a heat map generation unit 32, and an object region output unit 33. In the present embodiment, it is assumed that the image processed by the image processing apparatus 300 is a moving image (a series of temporally continuous images).
 ヒートマップ生成部32の基本動作は第2の実施形態と同様であるが、異なる点は、図11の点線矢印に表わされるように、ヒートマップ生成部32の出力が、ヒートマップ生成部32の入力となることである。本実施形態においては、動画像を処理するため、ヒートマップ生成部32で生成された過去の画像のヒートマップを、現在の画像のヒートマップ生成に利用する。ヒートマップ生成部32は、ある腫瘍を映す画像において、過去の時刻に生成されたヒートマップと、物体領域候補生成部31が生成する現在の時刻に対応する物体領域候補を用いて、現在の時刻に対応するヒートマップを生成する。 The basic operation of the heat map generation unit 32 is the same as that of the second embodiment, except that the output of the heat map generation unit 32 is the same as that of the heat map generation unit 32 as represented by the dotted arrows in FIG. To be input. In this embodiment, in order to process a moving image, the heat map of the past image produced | generated in the heat map production | generation part 32 is utilized for the heat map production | generation of the present image. The heat map generation unit 32 uses the heat map generated at a past time and the object region candidate corresponding to the current time generated by the object region candidate generation unit 31 in the image showing a certain tumor, and the current time A heat map corresponding to is generated.
 物体領域候補生成部31および物体領域出力部33は第2の実施形態の物体領域候補生成部21および物体領域出力部23と同様である。 The object region candidate generation unit 31 and the object region output unit 33 are the same as the object region candidate generation unit 21 and the object region output unit 23 of the second embodiment.
 (画像処理装置の動作)
 図12に示すフローチャートを参照して、上記画像処理装置300の具体的な動作について説明する。
(Operation of image processing device)
A specific operation of the image processing apparatus 300 will be described with reference to a flowchart shown in FIG.
 ステップS301については、図8のステップS201とほぼ同じ動作であるが、物体領域候補生成部31は、所定の間隔(例えば1秒)毎に、画像を物体領域候補生成部31に入力する。または動画像中に変化があった場合に画像を入力しても良い。 Step S301 is substantially the same as step S201 in FIG. 8, but the object region candidate generation unit 31 inputs an image to the object region candidate generation unit 31 at predetermined intervals (for example, 1 second). Alternatively, an image may be input when there is a change in the moving image.
 ステップS302において、物体領域候補生成部31で生成した物体領域候補と過去の画像のヒートマップと受け取ると、ヒートマップ生成部32はこれらを基に、ヒートマップを生成する。このステップの具体的な一動作例を、以下に説明する。 In step S302, when the object region candidate generated by the object region candidate generating unit 31 and the heat map of the past image are received, the heat map generating unit 32 generates a heat map based on these. A specific operation example of this step will be described below.
 本実施形態では、ヒートマップ生成部32は、時刻tの入力画像に対して物体領域候補生成部31およびヒートマップ生成部32によって得られた過去のヒートマップを、Score(Rectanglet,n) =Σt,i,j heatmapscore(t,i,j)とする。 In the present embodiment, the heat map generation unit 32 uses the past heat map obtained by the object region candidate generation unit 31 and the heat map generation unit 32 for the input image at time t as Score (Rectangle t, n ) = Let Σ t, i, j heatmap score (t, i, j) .
 ここで、例えば、時刻t+1の画像に対して物体領域候補生成部31が生成する各物体領域候補を
Rectanglet,n = [xt,n, yt,n, wt,n, ht,n, conft,n] …(式2)
とし、上記時刻tにおけるヒートマップを
Score(Rectanglet,n)=Σt,w,hheatmapscore(t,i,j) …(式3)
とした場合、式(2)および式(3)を用いて、時刻t+1におけるヒートマップを、
Score(Rectanglet+1,n)=Σt,w,hheatmapscore(t,i,j)t+1,i,jheatmapscore(t+1,i,j)…式(4)
と表わすことができる。
Here, for example, each object region candidate generated by the object region candidate generation unit 31 for the image at time t + 1 is displayed.
Rectangle t, n = [x t, n , y t, n , w t, n , h t, n , conf t, n ] (Formula 2)
And the heat map at time t
Score (Rectangle t, n ) = Σ t, w, h heatmap score (t, i, j) (Equation 3)
When using the equation (2) and the equation (3), the heat map at time t + 1 is
Score (Rectangle t + 1, n ) = Σt , w, h heatmap score (t, i, j) + Σt + 1, i, j heatmap score (t + 1, i, j) … Formula (4)
Can be expressed as
 即ち、物体領域候補生成部31は、現時刻の画像に対する物体領域候補と過去の時刻の画像で生成されたヒートマップとを基に、現時刻のヒートマップを生成する。 That is, the object region candidate generation unit 31 generates a heat map at the current time based on the object region candidate for the image at the current time and the heat map generated from the image at the past time.
 尚、本例では、時刻が1つ前の画像のヒートマップを単純に加算する場合を示したが、これに限らない。例えば過去の時刻n(例えば、秒)前までの時刻のヒートマップを利用してもよい。また加算する場合、過去の時刻の各ヒートマップを等価に加算する必要はなく、各時刻に重みを与えてもよい。すなわち、
Score(Rectanglet+1,n)=wtΣt,w,hheatmapscore(t,i,j)+wt+1Σt+1,i,jheatmapscore(t+1,i,j)…(式6)
としてもよい。この各時刻の重みwtは、たとえばwt=1/tと設定してもよいし、各時刻tのヒートマップに応じて決定してもよい。例えばヒートマップ画像の類似度をwtとしてもよい。重み付けの一例として、過去に生成されたヒートマップと現在生成されたヒートマップとの一致度合いを示す類似度を算出して、重みとして利用してもよい。この場合、過去と現在において動きがない程一致度合いは高くなる。
In this example, the case where the heat map of the image with the previous time is simply added is shown, but the present invention is not limited to this. For example, you may use the heat map of the time until the past time n (for example, second). Moreover, when adding, it is not necessary to add each heat map of the past time equivalently, You may give a weight to each time. That is,
Score (Rectangle t + 1, n ) = w t Σ t, w, h heatmap score (t, i, j) + w t + 1 Σ t + 1, i, j heatmap score (t + 1, i, j ) (Formula 6)
It is good. The weight w t at each time may be set as w t = 1 / t, for example, or may be determined according to the heat map at each time t. For example, the similarity of the heat map image may be set to w t . As an example of weighting, a similarity indicating a degree of coincidence between a heat map generated in the past and a heat map currently generated may be calculated and used as a weight. In this case, the degree of coincidence increases as there is no movement in the past and the present.
 ステップS303については、図8のステップS203と同様である。 Step S303 is the same as step S203 in FIG.
 (第3の実施形態の効果)
 本発明の第3の実施形態における画像処理装置300は、検知対象物体が動き、且つ、動画像で撮影された場合であっても、不規則物体の全体を示す枠を高速に表示することができる。ひいてはオペレータによる不規則物体の発見および判断を容易にすることができる。この理由は、ヒートマップ生成部32が、現時刻の物体領域候補に加え、過去の時刻にヒートマップ生成部32にて生成されたヒートマップも用いて、ヒートマップを生成し、物体領域出力部33が、このヒートマップを用いて、物体領域候補の中から物体領域を選択するからである。更に、各時刻に重みを与え、現時刻の類似度の基づき重みづけを行うことで、検知対象物体の動きに対して安定的に不規則物体の全体を示す枠を高速に表示することができる。ひいてはオペレータによる不規則物体の発見および判断を容易にすることができる。
(Effect of the third embodiment)
The image processing apparatus 300 according to the third embodiment of the present invention can display a frame indicating the entire irregular object at high speed even when the detection target object moves and is captured as a moving image. it can. As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator. This is because the heat map generation unit 32 generates a heat map using the heat map generated by the heat map generation unit 32 at the past time in addition to the object region candidate at the current time, and the object region output unit This is because 33 uses this heat map to select an object area from object area candidates. Furthermore, by giving a weight to each time and performing weighting based on the similarity of the current time, a frame indicating the entire irregular object can be displayed at high speed stably with respect to the movement of the detection target object. . As a result, it is possible to facilitate the discovery and determination of irregular objects by the operator.
 上述した本発明の各実施形態において、画像処理装置100、200、300、400は、画像を撮影するカメラ等の装置と一体であってもよいし、カメラ等と情報処理装置(パーソナルコンピュータなど)とを接続し、画像処理の全体動作の制御を情報処理装置側で行っても良い。即ち、情報処理装置側の画像処理用のアプリケーションプログラムを以下のような構成の下で使用してもよい。 In each of the embodiments of the present invention described above, the image processing apparatuses 100, 200, 300, and 400 may be integrated with an apparatus such as a camera that captures an image, or a camera and an information processing apparatus (such as a personal computer). May be connected to control the overall operation of the image processing on the information processing apparatus side. That is, an application program for image processing on the information processing apparatus side may be used under the following configuration.
 (情報処理装置)
 上述した本発明の各実施形態において、図1、7、11等に示す画像処理装置における各構成要素の一部又は全部は、例えば図13に示すような情報処理装置500とプログラムとの任意の組み合わせを用いて実現することも可能である。情報処理装置500は、一例として、以下のような構成を含む。
(Information processing device)
In each of the embodiments of the present invention described above, some or all of the components in the image processing apparatus shown in FIGS. 1, 7, 11 and the like are arbitrary information processing apparatuses 500 and programs as shown in FIG. It can also be realized using a combination. The information processing apparatus 500 includes the following configuration as an example.
  ・CPU(Central Processing Unit)501
  ・ROM(Read Only Memory)502
  ・RAM(Random Access Memory)503
  ・プログラム504および他のデータを格納する記憶装置505
  ・記録媒体506の読み書きを行うドライブ装置507
  ・通信ネットワーク509と接続する通信インターフェース508
  ・データの入出力を行う入出力インターフェース510
  ・各構成要素を接続するバス511
 本願の各実施形態における画像処理装置の各構成要素は、これらの機能を実現するプログラム504をCPU501が取得して実行することで実現される。画像処理装置の各構成要素(例えば、物体領域候補生成部11、ヒートマップ生成部12、物体領域出力部13)の機能を実現するプログラム504は、例えば、予め記憶装置505やRAM503に格納されており、必要に応じてCPU501が読み出す。なお、プログラム504は、通信ネットワーク509を介してCPU501に供給されてもよいし、予め記録媒体506に格納されており、ドライブ装置507が当該プログラムを読み出してCPU501に供給してもよい。
CPU (Central Processing Unit) 501
ROM (Read Only Memory) 502
-RAM (Random Access Memory) 503
A storage device 505 for storing the program 504 and other data
A drive device 507 for reading / writing the recording medium 506
Communication interface 508 connected to the communication network 509
An input / output interface 510 for inputting / outputting data
-Bus 511 connecting each component
Each component of the image processing apparatus in each embodiment of the present application is realized by the CPU 501 acquiring and executing a program 504 that realizes these functions. A program 504 that realizes the functions of each component of the image processing apparatus (for example, the object region candidate generation unit 11, the heat map generation unit 12, and the object region output unit 13) is stored in advance in the storage device 505 or the RAM 503, for example. The CPU 501 reads the data as necessary. Note that the program 504 may be supplied to the CPU 501 via the communication network 509 or may be stored in the recording medium 506 in advance, and the drive device 507 may read the program and supply it to the CPU 501.
 各装置の実現方法には、様々な変形例がある。例えば、画像処理装置は、構成要素毎にそれぞれ別個の情報処理装置とプログラムとの任意の組み合わせにより実現されてもよい。また、画像処理装置が備える複数の構成要素が、一つの情報処理装置500とプログラムとの任意の組み合わせにより実現されてもよい。 There are various modifications to the method of realizing each device. For example, the image processing apparatus may be realized by an arbitrary combination of a separate information processing apparatus and a program for each component. A plurality of components included in the image processing apparatus may be realized by an arbitrary combination of one information processing apparatus 500 and a program.
 また、画像処理装置の各構成要素の一部又は全部は、その他の汎用または専用の回路、プロセッサ等やこれらの組み合わせによって実現される。これらは、単一のチップによって構成されてもよいし、バスを介して接続される複数のチップによって構成されてもよい。 Also, some or all of the constituent elements of the image processing apparatus are realized by other general-purpose or dedicated circuits, processors, or combinations thereof. These may be configured by a single chip or may be configured by a plurality of chips connected via a bus.
 画像処理装置の各構成要素の一部又は全部は、上述した回路等とプログラムとの組み合わせによって実現されてもよい。 Some or all of the components of the image processing apparatus may be realized by a combination of the above-described circuit and the like and a program.
 画像処理装置の各構成要素の一部又は全部が複数の情報処理装置や回路等により実現される場合には、複数の情報処理装置や回路等は、集中配置されてもよいし、分散配置されてもよい。例えば、情報処理装置や回路等は、クライアントアンドサーバシステム、クラウドコンピューティングシステム等、各々が通信ネットワークを介して接続される形態として実現されてもよい。 When some or all of the components of the image processing apparatus are realized by a plurality of information processing apparatuses and circuits, the plurality of information processing apparatuses and circuits may be centrally arranged or distributedly arranged. May be. For example, the information processing apparatus, the circuit, and the like may be realized as a form in which each is connected via a communication network, such as a client and server system and a cloud computing system.
 以上、本実施形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されるものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解し得る様々な変更をすることができる。 As mentioned above, although this invention was demonstrated with reference to this embodiment, this invention is not limited to the said embodiment. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
 本発明は、画像中に含まれる不規則かつ予測不能な形状を伴う非剛体の物体を認識する画像処理装置や、その装置をコンピュータに実現するためのプログラムといった用途に適用可能である。 The present invention can be applied to applications such as an image processing apparatus that recognizes a non-rigid object having an irregular and unpredictable shape included in an image, and a program for realizing the apparatus on a computer.
 100、200、300  画像処理装置
 11、21、31  物体領域候補生成部
 12、22、32  ヒートマップ生成部
 13、23、33  物体領域出力部
500  情報処理装置
501  CPU
503  RAM
504  プログラム
505  記憶装置
506  記録媒体
507  ドライブ装置
508  通信インターフェース
509  通信ネットワーク
510  入出力インターフェース
511  バス
100, 200, 300 Image processing device 11, 21, 31 Object region candidate generation unit 12, 22, 32 Heat map generation unit 13, 23, 33 Object region output unit 500 Information processing device 501 CPU
503 RAM
504 Program 505 Storage device 506 Recording medium 507 Drive device 508 Communication interface 509 Communication network 510 Input / output interface 511 Bus

Claims (10)

  1.  画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成する物体領域候補生成手段と、
     生成された前記物体領域候補を用いて、前記画像中における前記位置に対応するヒートマップを生成するヒートマップ生成手段と、
     生成された前記ヒートマップを用いて、1つ以上の前記物体領域候補から成る、前記検知対象物体の全体を囲う物体領域を出力する物体領域出力手段
    とを備える画像処理装置。
    Object region candidate generation means for generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image;
    Heat map generating means for generating a heat map corresponding to the position in the image using the generated object region candidate;
    An image processing apparatus comprising: an object region output unit configured to output an object region that surrounds the entire detection target object, the object region including one or more object region candidates, using the generated heat map.
  2.  前記ヒートマップ生成手段は、各物体領域候補が前記検知対象物体を含む確からしさを表わす信頼度を用いて、前記ヒートマップを生成する
    請求項1に記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the heat map generation unit generates the heat map using a reliability that represents a probability that each object region candidate includes the detection target object.
  3.  前記物体領域出力手段は、前記ヒートマップを区分した複数の区画毎において、当該区画に紐付けられた前記信頼度が所定の閾値以上である区画の集合を含む物体領域候補を、前記物体領域として出力する
    請求項1または請求項2のいずれかに記載の画像処理装置。
    The object area output means uses, as the object area, an object area candidate including a set of sections whose reliability is equal to or higher than a predetermined threshold for each of a plurality of sections into which the heat map is divided. The image processing apparatus according to claim 1, which outputs the image processing apparatus.
  4.  前記物体領域出力手段は、生成された前記物体領域候補が重畳された前記ヒートマップを用いて、前記物体領域を出力する
    請求項1または請求項3のいずれかに記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the object area output unit outputs the object area using the heat map on which the generated object area candidates are superimposed.
  5.  前記ヒートマップ生成手段は、ある検知対象物体を映す画像において、過去の時刻に生成されたヒートマップと、前記物体領域候補生成手段が生成する現在の時刻に対応する物体領域候補を用いて、前記現在の時刻に対応するヒートマップを生成する
    請求項1または2のいずれかに記載の画像処理装置。
    The heat map generation means uses a heat map generated at a past time and an object region candidate corresponding to the current time generated by the object region candidate generation means in an image showing a certain detection target object, The image processing apparatus according to claim 1, wherein a heat map corresponding to the current time is generated.
  6.  前記ヒートマップ生成手段は、重み付けされた前記過去の時刻に対応するヒートマップを用いて、前記現在の時刻に対応するヒートマップを生成する
    請求項5に記載の画像処理装置。
    The image processing apparatus according to claim 5, wherein the heat map generation unit generates a heat map corresponding to the current time using a weighted heat map corresponding to the past time.
  7.  前記重み付けとして、前記過去の時刻に対応するヒートマップと前記現在の時刻に対応するヒートマップとの一致度合いを示す類似度を用いる
    請求項6に記載の画像処理装置。
    The image processing apparatus according to claim 6, wherein a similarity indicating a degree of coincidence between a heat map corresponding to the past time and a heat map corresponding to the current time is used as the weighting.
  8.  前記検知対象物体が、内視鏡画像に映る病変である
    請求項2、請求項3および請求項5のいずれかに記載の画像処理装置。
    The image processing apparatus according to claim 2, wherein the detection target object is a lesion reflected in an endoscopic image.
  9.  画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成し、
     生成された前記物体領域候補を用いて、前記画像中における前記位置に対応するヒートマップを生成し、
     生成された前記ヒートマップを用いて、1つ以上の前記物体領域候補から成る、前記検知対象物体の全体を囲う物体領域を出力する
    画像処理方法。
    Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image;
    Using the generated object region candidate, generate a heat map corresponding to the position in the image,
    An image processing method for outputting an object region that surrounds the entire detection target object, which is composed of one or more object region candidates, using the generated heat map.
  10.  画像中における検知対象物体の少なくとも一部が存在する位置を示す物体領域候補を複数生成し、
     生成された前記物体領域候補を用いて、前記画像中における前記位置に対応するヒートマップを生成し、
     生成された前記ヒートマップを用いて、1つ以上の前記物体領域候補から成る、前記検知対象物体の全体を囲う物体領域を出力する
    ことをコンピュータに実現させるための画像処理プログラムを格納する記録媒体。
    Generating a plurality of object region candidates indicating positions where at least some of the detection target objects exist in the image;
    Using the generated object region candidate, generate a heat map corresponding to the position in the image,
    A recording medium for storing an image processing program for causing a computer to output an object region surrounding the entire detection target object, which is composed of one or more object region candidates, using the generated heat map .
PCT/JP2018/004691 2018-02-09 2018-02-09 Image processing device, image processing method, and recording medium WO2019155628A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2018/004691 WO2019155628A1 (en) 2018-02-09 2018-02-09 Image processing device, image processing method, and recording medium
JP2019570260A JP6988926B2 (en) 2018-02-09 2018-02-09 Image processing equipment, image processing method and image processing program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2018/004691 WO2019155628A1 (en) 2018-02-09 2018-02-09 Image processing device, image processing method, and recording medium

Publications (1)

Publication Number Publication Date
WO2019155628A1 true WO2019155628A1 (en) 2019-08-15

Family

ID=67548244

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/004691 WO2019155628A1 (en) 2018-02-09 2018-02-09 Image processing device, image processing method, and recording medium

Country Status (2)

Country Link
JP (1) JP6988926B2 (en)
WO (1) WO2019155628A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022162766A1 (en) * 2021-01-27 2022-08-04 オリンパス株式会社 Information processing system, endoscope system, information processing method, and annotation data generation method
WO2023181353A1 (en) * 2022-03-25 2023-09-28 日本電気株式会社 Image processing device, image processing method, and storage medium
WO2024024022A1 (en) * 2022-07-28 2024-02-01 日本電気株式会社 Endoscopic examination assistance device, endoscopic examination assistance method, and recording medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014142760A (en) * 2013-01-23 2014-08-07 Mega Chips Corp Image detection device, control program and image detection method
JP2014191400A (en) * 2013-03-26 2014-10-06 Mega Chips Corp Image detection apparatus, control program, and image detection method
JP2014197353A (en) * 2013-03-29 2014-10-16 株式会社メガチップス Image detection device, control program and image detection method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5489340B2 (en) * 2010-03-12 2014-05-14 パナソニック株式会社 Face masking apparatus and method
JP6303090B2 (en) * 2014-03-24 2018-04-04 アルパイン株式会社 Image processing apparatus and image processing program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014142760A (en) * 2013-01-23 2014-08-07 Mega Chips Corp Image detection device, control program and image detection method
JP2014191400A (en) * 2013-03-26 2014-10-06 Mega Chips Corp Image detection apparatus, control program, and image detection method
JP2014197353A (en) * 2013-03-29 2014-10-16 株式会社メガチップス Image detection device, control program and image detection method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022162766A1 (en) * 2021-01-27 2022-08-04 オリンパス株式会社 Information processing system, endoscope system, information processing method, and annotation data generation method
WO2023181353A1 (en) * 2022-03-25 2023-09-28 日本電気株式会社 Image processing device, image processing method, and storage medium
WO2024024022A1 (en) * 2022-07-28 2024-02-01 日本電気株式会社 Endoscopic examination assistance device, endoscopic examination assistance method, and recording medium

Also Published As

Publication number Publication date
JPWO2019155628A1 (en) 2021-01-28
JP6988926B2 (en) 2022-01-05

Similar Documents

Publication Publication Date Title
US11450146B2 (en) Gesture recognition method, apparatus, and device
US11033251B2 (en) Systems and methods for ultrasound imaging
JP6636154B2 (en) Face image processing method and apparatus, and storage medium
CN109478239B (en) Method for detecting object in image and object detection system
CN111047626B (en) Target tracking method, device, electronic equipment and storage medium
US20180032840A1 (en) Method and apparatus for neural network training and construction and method and apparatus for object detection
KR102046134B1 (en) Neural network training method for utilizing differences between a plurality of images, and method thereof
US20190384967A1 (en) Facial expression detection method, device and system, facial expression driving method, device and system, and storage medium
Wang et al. Afp-net: Realtime anchor-free polyp detection in colonoscopy
US20220076002A1 (en) Action recognition method and apparatus, computer storage medium, and computer device
US11908128B2 (en) Systems and methods to process images for skin analysis and to visualize skin analysis
JP5925068B2 (en) Video processing apparatus, video processing method, and program
KR20200118076A (en) Biometric detection method and device, electronic device and storage medium
WO2019155628A1 (en) Image processing device, image processing method, and recording medium
US10255673B2 (en) Apparatus and method for detecting object in image, and apparatus and method for computer-aided diagnosis
JP6808557B2 (en) Information processing device, its control method and program
JP2014116912A (en) Image processing apparatus, control method for image processing apparatus, and program
KR102530010B1 (en) Apparatus and method for determining disease severity based on medical image
JP6996303B2 (en) Medical image generator
KR20140033907A (en) Method and apparatus for estimating position of head, computer readable storage medium thereof
Chen et al. Pseudo-labeling for small lesion detection on diabetic retinopathy images
KR20150094108A (en) Method for generating saliency map based background location and medium for recording the same
JP2011043922A (en) Device and method for recognizing traveling object, and program
JP7385416B2 (en) Image processing device, image processing system, image processing method, and image processing program
Putro et al. Multiple layered deep learning based real-time face detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18905470

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2019570260

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18905470

Country of ref document: EP

Kind code of ref document: A1