WO2015159585A1 - Image-processing device, imaging device, image-processing method, and program - Google Patents

Image-processing device, imaging device, image-processing method, and program Download PDF

Info

Publication number
WO2015159585A1
WO2015159585A1 PCT/JP2015/054971 JP2015054971W WO2015159585A1 WO 2015159585 A1 WO2015159585 A1 WO 2015159585A1 JP 2015054971 W JP2015054971 W JP 2015054971W WO 2015159585 A1 WO2015159585 A1 WO 2015159585A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
input image
similarity
region
reference image
Prior art date
Application number
PCT/JP2015/054971
Other languages
French (fr)
Japanese (ja)
Inventor
ケツテイ チョウ
直規 葛谷
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Publication of WO2015159585A1 publication Critical patent/WO2015159585A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems

Definitions

  • the present technology relates to an image processing device, an imaging device, an image processing method, and a program. More specifically, the present invention relates to an image processing device, an imaging device, a processing method in these, and a program for causing a computer to execute the method.
  • a surveillance camera that estimates an object region by image processing has been used.
  • a surveillance camera that estimates an object region by a background difference method for obtaining a difference between a background image and an input image has been proposed (see, for example, Patent Document 1).
  • This surveillance camera obtains a difference area in which the pixel value difference is larger than the threshold value in the input image and the background image by the background difference method, and obtains the similarity between these difference areas by normalized cross-correlation matching.
  • the similarity is higher than the threshold, the monitoring camera estimates that the difference area is a disturbance area such as light or shadow, and otherwise the difference area is an object area such as a suspicious object or a suspicious person. Estimated.
  • the above-described conventional technique when noise occurs in the input image, the similarity of the difference area is lowered due to the influence of the noise. For this reason, there is a possibility that the surveillance camera erroneously estimates that the disturbance is a suspicious object or the like. As described above, the above-described surveillance camera has a problem that the area of the object cannot be accurately estimated.
  • This technology was created in view of such a situation, and aims to accurately estimate an object region.
  • a first aspect of the present technology is a comparison region determination unit that determines a comparison region to be compared in each of an input image and a predetermined reference image.
  • An input image feature that acquires a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image as an input image feature amount
  • a reference image feature that acquires a value corresponding to a difference between pixel values of a quantity acquisition unit and a corresponding pixel whose coordinates match the pixel of interest in the reference image and a pixel whose coordinates match the surrounding pixel as a reference image feature quantity Based on the input image feature quantity and the reference image feature quantity, the similarity between the comparison area in the input image and the comparison area in the input image and the comparison area in the reference image is set as the area similarity.
  • An image processing apparatus comprising: a similarity acquisition unit to be obtained; an object estimation unit that estimates the comparison region that is not similar based on the region similarity in the input image as a region of the object; A program for causing a computer to execute the method. This brings about the effect that a comparison region that is not similar is estimated as the region of the object.
  • the input image feature amount acquisition unit may detect a plurality of corners in the comparison region and set the pixel of interest. This brings about the effect that a plurality of corners are detected as the target pixel in the comparison region.
  • the input image feature quantity acquisition unit may generate a random number corresponding to any of the pixels in the comparison region and set the pixel corresponding to the random number as the target pixel. This brings about the effect that the pixel corresponding to the random number is set as the target pixel.
  • the input image feature quantity acquisition unit may extract pixels within a predetermined distance from the pixel of interest as the surrounding pixels. This brings about the effect that pixels within a predetermined distance from the target pixel are extracted as surrounding pixels.
  • the input image feature amount acquisition unit may extract a pixel whose pixel value is not similar to the target pixel from the pixels in the comparison area, and use the extracted pixel as the surrounding pixel. This brings about the effect that pixels whose pixel values are not similar to the target pixel are extracted as surrounding pixels.
  • the object estimation unit acquires the input image feature amount acquired for each of the plurality of target pixels and the corresponding pixels whose coordinates coincide with the target pixel.
  • the local similarity that is the similarity to the reference image feature amount is higher than a predetermined local determination threshold, and the local similarity is higher than the local determination threshold
  • the obtained value may be acquired as the region similarity. Accordingly, an effect is obtained in which a value corresponding to the number of times that the local similarity between the input image feature quantity and the reference image feature quantity is determined to be higher than the local determination threshold is acquired as the area similarity.
  • pixels having similar pixel values in the input image and the reference image may be detected, and an area including the detected pixels may be determined as the comparison area. This brings about the effect that an area composed of pixels whose pixel values are not similar in the input image and the reference image is determined as the comparison area.
  • the comparison area is determined in each of the two input images and the reference image, and a vector from one comparison area to the other comparison area of the two input images is moved. It may be detected as a vector. This brings about the effect that a vector from one comparison area of the two input images to the other comparison area is detected as a movement vector.
  • a second aspect of the present technology includes an imaging unit that captures an input image, a comparison region determination unit that determines a comparison region to be compared in each of the input image and a predetermined reference image, and the input image
  • An input image feature amount acquisition unit that acquires, as an input image feature amount, a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region;
  • a reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinate matches the pixel of interest in a reference image and a pixel whose coordinate matches the surrounding pixel;
  • a similarity acquisition unit that acquires the similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature amount and the reference image feature amount
  • An imaging apparatus including an object estimation unit for estimating the comparison area in the input image is not similar
  • 1 is a block diagram illustrating a configuration example of an imaging apparatus according to a first embodiment. It is a block diagram which shows one structural example of the image process part in 1st Embodiment. It is a block diagram which shows one structural example of the difference area
  • FIG. 3 is a flowchart illustrating an example of an operation of the imaging apparatus according to the first embodiment. It is a flowchart which shows an example of the difference area
  • First embodiment example of obtaining similarity of region from feature amount
  • Second embodiment an example of selecting a pixel of interest randomly when obtaining the similarity of a region from a feature amount
  • Third Embodiment Example in which a movement vector is detected before the similarity of a region is obtained from a feature amount
  • FIG. 1 is a block diagram illustrating a configuration example of the imaging apparatus 100 according to the first embodiment.
  • the imaging apparatus 100 captures an image and includes an imaging lens 110, an imaging element 120, a recording unit 130, a control unit 140, and an image processing unit 200.
  • the imaging lens 110 collects light and guides it to the imaging device 120.
  • the image sensor 120 converts the light from the imaging lens 110 into an electrical signal and captures an image under the control of the control unit 140. Each time the image sensor 120 captures an image, the image sensor 120 supplies the image as an input image to the image processing unit 200 via the signal line 129.
  • the imaging element 120 is an example of an imaging unit described in the claims.
  • the image processing unit 200 estimates an object area in the input image.
  • the image processing unit 200 supplies the estimation result to the control unit 140 via the signal line 209. Further, the image processing unit 200 supplies the input image to the recording unit 130 via the signal line 208.
  • the recording unit 130 records an input image and a reference image. For example, an image captured in advance before the input image is captured at the monitoring target location is used as the reference image.
  • the control unit 140 controls the entire imaging apparatus 100.
  • the control unit 140 generates a control signal for instructing imaging in accordance with a user operation or the like, and supplies the control signal to the image sensor 120 via the signal line 149. Further, the control unit 140 receives the estimation result from the image processing unit 200, and if any region is estimated to be an object region, the control unit 140 sends an alarm signal to that effect to the outside of the imaging device 100 or the like. Output.
  • the image processing unit 200 is provided in the imaging device 100, but may be provided in an image processing device different from the imaging device 100.
  • the imaging apparatus 100 supplies an input image to the image processing apparatus, and the image processing apparatus estimates an object region and supplies an estimation result to the imaging apparatus 200 or the like.
  • FIG. 2 is a block diagram illustrating a configuration example of the image processing unit 200 according to the first embodiment.
  • the image processing unit 200 includes a noise removal unit 210, a difference area detection unit 220, an input image feature amount acquisition unit 230, a reference image feature amount acquisition unit 240, a similarity acquisition unit 250, and an object estimation unit 260.
  • the noise removing unit 210 performs processing for removing noise from the input image. For example, the noise removing unit 210 removes noise by passing a low-pass filter that suppresses high frequency components higher than a predetermined cutoff frequency. This low-pass filter is realized by, for example, an IIR (Infinite Impulse Response) filter or an FIR (Finite Impulse Response) filter.
  • the noise removal unit 210 supplies the input image from which noise has been removed to the recording unit 130, the difference area detection unit 220, and the input image feature amount acquisition unit 230.
  • the difference area detection unit 220 determines an area to be compared in the input image and the reference image. For example, the difference area detection unit 220 detects pixels whose pixel values are not similar in the input image and the reference image, and sets an area including these pixels as a comparison area. First, the difference area detection unit 220 detects the absolute value of the difference between pixel values for each pixel having the same coordinate, generates a difference image composed of these differences, and converts the difference image into a binary image using a predetermined binarization threshold. Convert to value.
  • the difference area detection part 220 is the identification information which identifies the area
  • Labeling process to allocate as a label for example, a 4-connection algorithm that connects pixels that are continuous in the horizontal direction and the vertical direction, an 8-connection algorithm that further connects pixels that are continuous in an oblique direction in addition to those directions, and the like are used. It is done.
  • an area in the input image and the reference image indicated by the labeled area is referred to as a “difference area”.
  • difference areas are compared as areas to be compared in the input image and the reference image.
  • the difference area detection unit 220 supplies the image subjected to the labeling process to the input image feature amount acquisition unit 230 and the reference image feature amount acquisition unit 240 as a labeling image.
  • the difference area detection unit 220 is an example of a comparison area determination unit described in the claims.
  • the input image feature amount acquisition unit 230 acquires a feature amount in a difference area in the input image.
  • the input image feature amount acquisition unit 230 detects a plurality of target pixels in the difference region, obtains a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each target pixel, and according to the difference The obtained value is acquired as a local feature amount.
  • the pixel value compared when obtaining the local feature amount is, for example, a luminance value or a color difference.
  • the input image feature amount acquisition unit 230 supplies each of the local feature amounts to the similarity acquisition unit 250.
  • the reference image feature amount acquisition unit 240 acquires feature amounts in the difference area in the reference image.
  • the reference image feature amount acquisition unit 240 receives the coordinates of the target pixel from the input image feature amount acquisition unit 230, and obtains a difference in pixel value between the corresponding pixel whose coordinates match the target pixel and the surrounding surrounding pixels. Then, the reference image feature amount acquisition unit 240 acquires a value corresponding to the difference as a local feature amount, and supplies it to the similarity acquisition unit 250.
  • the similarity acquisition unit 250 calculates the similarity between the difference areas of the input image and the reference image based on the local feature acquired by the input image feature acquisition unit 230 and the reference image feature acquisition unit 240. Is what you want. For example, the region similarity is higher as the degree of similarity between regions is higher. Details of the region similarity acquisition method will be described later.
  • the similarity acquisition unit 250 supplies the obtained region similarity to the object estimation unit 260.
  • the object estimation unit 260 estimates a region having a region similarity lower than a predetermined region determination threshold (in other words, a region that is not similar) as a region of a suspicious object or a suspicious object that is not in the reference image.
  • the object estimation unit 260 supplies the estimation result to the control unit 140.
  • the image processing unit 200 performs noise removal on the input image, but may perform processing other than noise removal as long as the image quality is improved.
  • the image processing unit 200 may perform contrast enhancement or electronic camera shake correction instead of noise removal.
  • Electronic camera shake correction is also called stabilization processing.
  • the image processing unit 200 may perform contrast enhancement and electronic camera shake correction in addition to noise removal. As described above, by performing the process of improving the image quality, it is possible to improve the estimation accuracy of the object.
  • FIG. 3 is a block diagram illustrating a configuration example of the difference area detection unit 220 according to the first embodiment.
  • the difference area detection unit 220 includes a difference image generation unit 221 and a labeling processing unit 222.
  • the difference image generation unit 221 generates a difference image between the input image and the reference image.
  • the difference image generation unit 221 binarizes the generated difference image and supplies it to the labeling processing unit 222.
  • the labeling processing unit 222 performs a labeling process on the binarized difference image.
  • the labeling processing unit 222 supplies the image subjected to the labeling process to the input image feature amount acquisition unit 230 and the reference image feature amount acquisition unit 240 as a labeling image.
  • FIG. 4 is a diagram illustrating an example of the reference image 500 according to the first embodiment. As shown in the figure, the reference image 500 includes only backgrounds such as houses and trees.
  • FIG. 5 is a diagram illustrating an example of the input image 510 according to the first embodiment.
  • the input image 510 includes subjects 511, 512 and 513 in addition to the background.
  • the subject 511 is, for example, a suspicious object or a suspicious person.
  • the subject 512 is light such as search light, for example.
  • the subject 513 is, for example, a cloud shadow.
  • FIG. 6 is a diagram illustrating an example of the difference image 520 according to the first embodiment.
  • the difference image 520 is obtained by binarizing the difference image between the reference image 500 and the input image 510.
  • the difference image 520 includes difference areas 521, 522, and 523 corresponding to the subjects 511, 512, and 513. These difference areas are areas in the input image 510 where the absolute value of the pixel value difference with respect to the reference image 500 is larger than a predetermined binarization threshold.
  • the imaging apparatus 100 generates a difference image from the entire input image and reference image, but is not limited to this configuration. For example, when performing an electronic zoom or the like, the imaging apparatus 100 may generate a difference image from an extraction range of an input image and a reference image using a part of the image as an extraction range.
  • FIG. 7 is a diagram illustrating an example of a labeling image 530 according to the first embodiment.
  • the labeling image 530 includes difference areas 531, 532, and 533 corresponding to the subjects 511, 512, and 513.
  • a label “1” is assigned to the difference area 531.
  • a label “2” is assigned to the difference area 532, and a label “3” is assigned to the difference area 533, for example.
  • FIG. 8 is a block diagram illustrating a configuration example of the input image feature amount acquisition unit 230 according to the first embodiment.
  • the input image feature amount acquisition unit 230 includes a corner detection unit 231 and a local feature amount acquisition unit 232.
  • the corner detection unit 231 detects a corner as a target pixel in the difference area.
  • the corner detection unit 231 detects an edge using a canny filter or the like in each difference region in the input image, and detects a corner that is an intersection of the detected edges as a target pixel.
  • the corner detection unit 231 supplies the coordinates of the target pixel to the local feature amount acquisition unit 232 and the reference image feature amount acquisition unit 240.
  • the local feature amount acquisition unit 232 acquires a local feature amount in the input image.
  • a value corresponding to the difference between the target pixel and surrounding pixels around the target pixel is obtained as the local feature amount.
  • the local feature amount acquisition unit 232 calculates, for example, a difference obtained by subtracting the pixel value of the target pixel from the pixel value of the surrounding pixel for each of the eight surrounding pixels whose coordinate Euclidean distance is 2 1/2 or less, A value corresponding to the difference is obtained as a local feature amount.
  • the local feature amount acquisition unit 232 generates a local feature vector including the local feature amount for each group including the target pixel and the eight surrounding pixels, and supplies the local feature vector to the similarity acquisition unit 250.
  • the configuration of the reference image feature quantity acquisition unit 240 is the same as that of the local feature quantity acquisition unit 232 except that a local feature vector is generated from the reference image.
  • FIG. 9 is a diagram illustrating an example of a pixel of interest and surrounding pixels in the input image 510 according to the first embodiment.
  • the input image feature amount acquisition unit 230 detects a corner as the target pixel 611. Then, the input image feature amount acquisition unit 230 extracts eight pixels around the pixel of interest 611 as surrounding pixels.
  • a group 612 surrounded by a dotted line is a group including a pixel of interest 611 and surrounding pixels around it.
  • FIG. 10 is a diagram illustrating an example of a pixel of interest and surrounding pixels in the reference image 500 according to the first embodiment.
  • the reference image feature amount acquisition unit 240 extracts a corresponding pixel 601 having the same coordinates as the target pixel 611 and eight surrounding pixels around the corresponding pixel 601.
  • a group 602 surrounded by a dotted line is a group including a corresponding pixel 601 and surrounding pixels around it.
  • FIG. 11 is a diagram for explaining a local feature vector acquisition method according to the first embodiment.
  • the input image feature amount acquisition unit 230 extracts the pixel of interest 611 and eight surrounding pixels. Then, the input image feature amount acquisition unit 230 calculates, for each peripheral pixel, a difference obtained by subtracting the pixel value of the target pixel 611 from the pixel value of the peripheral pixel. For example, when the difference is within a certain range of +30 to ⁇ 30, the input image feature quantity acquisition unit 230 acquires a value of “0” as a local feature quantity. Further, the input image feature quantity acquisition unit 230 acquires the value “1” as the local feature quantity when the difference is greater than +30, and the value “ ⁇ 1” when the difference is less than ⁇ 30. Acquired as a feature value. A local feature vector including these local feature amounts is obtained for each group 612. When the number of surrounding pixels is 8, each of the local feature vectors includes 8 local feature amounts.
  • the pixel value of the target pixel is “65”, and the pixel values of the eight surrounding pixels are “30”, “60”, “65”, “60”, “140”, “60”, “200”, respectively. And “60”.
  • “ ⁇ 1”, “0”, “0”, “0”, “1”, “0”, “1”, and “0” are obtained as local feature amounts.
  • the imaging apparatus 100 obtains a local feature amount corresponding to a difference in pixel value, and obtains a region similarity from the local feature amount, so noise resistance is increased.
  • the difference between them is lower than ⁇ 30. ⁇ 1 ”.
  • the pixel value of the target pixel fluctuates to “67” due to noise that cannot be removed by the noise removing unit 210, the difference is still lower than ⁇ 30, and thus the local feature amount is “ ⁇ 1”. The value does not change. For this reason, the fluctuation
  • FIG. 12 is a diagram showing an example of local feature vectors in the first embodiment.
  • the input image feature quantity acquisition unit 230 extracts a plurality of groups for each difference area corresponding to the label in the input image, and obtains a local feature vector for each of these groups.
  • the reference image feature amount acquisition unit 240 extracts a plurality of groups for each label (difference area) in the reference image, and obtains a local feature vector for each of these groups.
  • FIG. 13 is a diagram illustrating an example of the degree of similarity according to the first embodiment.
  • the similarity acquiring unit 250 obtains the similarity between the local feature acquired in the input image and the local feature acquired in the reference image for each group as the local similarity.
  • the local similarity is obtained by, for example, normalized cross correlation matching.
  • the similarity acquisition unit 250 determines whether or not the local similarity is higher than a predetermined local determination threshold for each group, and sets a value corresponding to the number of times the local similarity is determined to be higher than the local determination threshold as a region. Calculate as similarity. For example, a value obtained by dividing the number of times that the local similarity is higher than the local determination threshold in the difference area by the number of groups in the difference area is obtained for each label (difference area) as the area similarity.
  • the similarity acquisition unit 250 may determine the number of times the local similarity is determined to be higher than the local determination threshold as the local similarity.
  • the corner detection part 231 should just make the number of each attention pixels of a difference area the same. For example, the corner detection unit 231 detects all corners for each difference area, and reduces the target pixel in other areas in accordance with the difference area with the smallest number of detections.
  • the similarity acquisition unit 250 obtains the local similarity by the normalized cross correlation matching, it is not limited to this configuration.
  • the similarity obtaining unit 250 may obtain SAD (Sum of Absolute Differences), which is the sum of absolute values of differences, as the local similarity.
  • the similarity acquisition unit 250 may obtain SSD (SumSof Squared Differences), which is the sum of squares of differences, as the local similarity.
  • FIG. 14 is a diagram illustrating an example of an estimation result in the first embodiment.
  • the object estimation unit 260 estimates a label (difference area) whose area similarity is lower than a predetermined area determination threshold as an area of a suspicious object or a suspicious person.
  • a label whose region similarity is equal to or higher than a predetermined region determination threshold indicates a region where the texture of the region is not changed and the brightness is merely changed as a whole.
  • Such a region is determined as a region where disturbance such as light or shadow has occurred. In other words, it is determined that it is not a suspicious object.
  • FIG. 15 is a flowchart illustrating an example of the operation of the imaging apparatus 100 according to the first embodiment. This operation is executed every time an input image is captured, for example.
  • the imaging apparatus 100 executes a difference area detection process for detecting a difference area between the input image and the reference image (step S910). Then, the imaging apparatus 100 executes a feature amount acquisition process for acquiring a feature amount in the difference area (step S920).
  • the imaging apparatus 100 selects any one of the difference areas where the object is not estimated (step S901), and acquires the area similarity of the difference area based on the feature amount (step S902).
  • the imaging apparatus 100 determines whether or not the region similarity is lower than the region determination threshold (step S903).
  • step S903: Yes If the area similarity is lower than the area determination threshold (step S903: Yes), the imaging apparatus 100 determines that the area is an area such as a suspicious object (step S904). On the other hand, when the region similarity is equal to or greater than the region determination threshold (step S903: No), the imaging apparatus 100 determines that the region is a region such as a disturbance and is not a region such as a suspicious object (step S905). ).
  • step S904 or S905 the imaging apparatus 100 determines whether the object has been estimated in all the difference areas (step S906). If the object is not estimated in all the difference areas (step S906: No), the imaging apparatus 100 returns to step S901. On the other hand, when the object is estimated in all the difference areas (step S906: Yes), the imaging apparatus 100 ends the process on the input image (step S906).
  • FIG. 16 is a flowchart illustrating an example of a difference area detection process according to the first embodiment.
  • the imaging device 100 generates a difference image between the input image and the reference image (step S911). Then, the imaging apparatus 100 performs a labeling process on the difference image to generate a labeling image (step S912). After step S912, the imaging apparatus 100 ends the difference area detection process.
  • FIG. 17 is a flowchart illustrating an example of a feature amount acquisition process according to the first embodiment.
  • the imaging apparatus 100 detects a plurality of corners as the target pixel in the difference area of the input image (step S921). Then, the imaging apparatus 100 extracts a group composed of the pixel of interest and surrounding pixels around it from the input image, and acquires a local feature vector composed of a local feature amount for each group (step S922). In addition, the imaging apparatus 100 extracts a group including a pixel whose coordinates match the target image and surrounding pixels around the target image in the input image, and acquires a local feature vector including a local feature amount for each group (step). S923). After step S923, the imaging apparatus 100 ends the feature amount acquisition process.
  • the region similarity acquired from the feature amount according to the difference between the pixel values of the target pixel and the surrounding pixels It is possible to accurately estimate the region of the object while suppressing the variation of the region similarity due to noise.
  • the imaging apparatus 100 detects the corner as the target pixel, but may select a randomly obtained pixel as the target pixel. In addition, the imaging apparatus 100 extracts pixels within a certain distance from the target pixel as surrounding pixels, but may extract pixels whose pixel values are not similar to the target pixel as surrounding pixels. The imaging apparatus 100 according to the second embodiment selects the randomly obtained pixel as a target pixel, and extracts pixels whose pixel values are not similar to the target pixel as surrounding pixels. Different from form.
  • FIG. 18 is a block diagram illustrating a configuration example of the input image feature quantity acquisition unit 230 according to the second embodiment.
  • the input image feature amount acquisition unit 230 according to the second embodiment includes a random number generation unit 233, a surrounding pixel extraction unit 234, and a local feature amount acquisition unit 235.
  • the random number generation unit 233 generates a random number corresponding to any coordinate in the difference area. For example, the minimum value and the maximum value of the x coordinate of the pixels in the difference area are set to x min and x max, and the minimum value and the maximum value of the y coordinate are set to y min and y max .
  • the random number generation unit 233 generates random numbers x r and y r by the following formula.
  • rand (A) is a function that returns a random number from 0 to A-1 using a linear congruential method or the like.
  • the random number generator 233 again generates a random number. Generate.
  • the random number generation unit 233 supplies the coordinates to the surrounding pixel extraction unit 234 as the coordinates of the target pixel. Then, the random number generation unit 233 repeats generation of random numbers until the number of generated coordinates of the target pixel reaches a certain number.
  • the surrounding pixel extraction unit 234 extracts surrounding pixels whose pixel values are not similar to the target pixel.
  • the surrounding pixel extraction unit 234 obtains a pixel value difference with respect to the target pixel in order from a pixel close to the target pixel among pixels around the target pixel in the input image, and the absolute value of the difference is a predetermined difference threshold (for example, 30) It is determined whether or not it is larger. Then, the surrounding pixel extraction unit 234 extracts, as surrounding pixels corresponding to the target pixel, a pixel that is first determined that the absolute value of the difference is larger than the difference threshold (in other words, the pixel values are not similar).
  • the surrounding pixel extraction unit 234 supplies the coordinates of the surrounding pixel and the pixel pair of the surrounding pixel to the reference image feature amount acquisition unit 240. In addition, the surrounding pixel extraction unit 234 supplies the difference to the local feature amount acquisition unit 235.
  • the local feature amount acquisition unit 235 acquires a local feature amount corresponding to the difference for each pixel pair. For example, the local feature quantity acquisition unit 235 acquires a value of “1” as the local feature quantity when the sign of the difference is + and a value of “0” when the sign of the difference is ⁇ . The local feature amount acquisition unit 235 supplies the local feature amount of each pixel pair to the similarity acquisition unit 250.
  • the input image feature quantity acquisition unit 230 randomly selects a pixel of interest, but may detect a corner as a pixel of interest as in the first embodiment. In this case, every time a corner is detected, surrounding pixels corresponding to the corner are extracted, and a pixel pair is generated.
  • FIG. 19 is a block diagram illustrating a configuration example of the reference image feature amount acquisition unit 240 according to the second embodiment.
  • the reference image feature amount acquisition unit 240 includes a difference acquisition unit 241 and a local feature amount acquisition unit 242.
  • the difference acquisition unit 241 acquires a pixel value difference between a pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixels.
  • the difference acquisition unit 241 supplies the difference to the local feature amount acquisition unit 242.
  • the configuration of the local feature quantity acquisition unit 242 is the same as that of the local feature quantity acquisition unit 235.
  • FIG. 20 is a diagram for explaining a pixel pair extraction method according to the second embodiment.
  • a black pixel is a pixel of interest selected at random.
  • the hatched pixels are surrounding pixels associated with the target pixel.
  • a pixel pair whose both ends are connected by a line segment indicated by an arrow is a pixel pair of a corresponding target pixel and surrounding pixels.
  • the surrounding pixel extraction unit 2344 pays attention to the pixels in order along a spiral path that moves away from the target pixel 711 as the vehicle turns, obtains a pixel value difference with respect to the target pixel 711 for the target pixel, and the absolute value of the difference is It is determined whether or not the difference threshold value is greater. Then, the surrounding pixel extraction unit 234 extracts, as the corresponding surrounding pixel 712, a pixel that is first determined that the absolute value of the difference is larger than the difference threshold.
  • the surrounding pixel extraction unit 234 extracts surrounding pixels by paying attention to the pixels in order along the spiral path.
  • pixels within a certain distance are used as surrounding pixels. It may be extracted.
  • surrounding 8 pixels are extracted as surrounding pixels, and a local feature vector is obtained.
  • FIG. 21 is an enlarged view of a part of the difference area of the reference image in the second embodiment.
  • the number of pixels in the enlarged part is 200 pixels of 20 ⁇ 10. In this portion, it is assumed that there are 139 pixels having a pixel value “40” and 61 pixels having a pixel value “200”.
  • FIG. 22 is an enlarged view of a part of the input image in the second embodiment.
  • the enlarged part is a part corresponding to the part enlarged in the input image in FIG.
  • This part is a part of a cloud shadow or the like, and it is assumed that the average pixel value is about 40 lower than the reference image.
  • the pixel value of the pixel having the same coordinate as the pixel having the pixel value “40” in the reference image is around “0” in the input image, and the pixel value having the same coordinate as the pixel having the pixel value “200” in the reference image.
  • the pixel value is around “160” in the input image.
  • noise is generated in the input image that cannot be completely removed by the noise removal unit 210.
  • the pixel value has a variation of about a variance value “5” due to the influence of noise.
  • the region similarity R is calculated by the following equation.
  • N is the number of pixels in the x direction of the difference area
  • M is the number of pixels in the Y direction.
  • I (i, j) is a pixel value of a pixel in the difference area of the reference image.
  • T (i, j) is a pixel value of a pixel in the difference area of the input image.
  • Equation 6 when the region similarity R is obtained by normalized cross-correlation matching or the like, the region similarity R may be lowered due to the influence of noise.
  • FIG. 23 is a diagram illustrating an example of local feature amounts according to the second embodiment.
  • a in the same figure is an example of the local feature-value of the expansion part of an input image.
  • B in the figure is an example of the local feature amount of the enlarged portion of the reference image.
  • the local feature values for each pixel pair in the input image and the reference image all match. For this reason, the region similarity is the maximum value “1”. Thus, the region similarity is unlikely to decrease due to the influence of noise.
  • FIG. 24 is a flowchart illustrating an example of a feature amount acquisition process according to the second embodiment.
  • the imaging apparatus 100 selects a target pixel at random in the difference area of the input image (step S925). Then, the imaging apparatus 100 pays attention to the pixels in order along the spiral path around the pixel of interest, and extracts surrounding pixels in which the absolute value of the pixel value difference is greater than the difference threshold (step S926).
  • the imaging apparatus 100 acquires the local feature amount of the pixel pair of the target pixel and the surrounding pixels (step S927).
  • the imaging apparatus 100 determines whether or not the number of pixel pairs is equal to or greater than a set value (step S928). When the number of pixel pairs is less than the set value (step S928: No), the imaging apparatus 100 returns to step S925.
  • step S928 when the number of pixel pairs is equal to or larger than the set value (step S928: Yes), the imaging apparatus 100 acquires a local feature amount in the reference image for each pixel pair (step S929). After step S929, the imaging apparatus 100 ends the feature amount acquisition process.
  • the imaging apparatus 100 since the imaging apparatus 100 generates a random number and selects a pixel corresponding to the random number as a target pixel, a difference region (for example, a density gradient that is difficult to detect a corner) is selected. In a small area, the target pixel can be easily selected.
  • a difference region for example, a density gradient that is difficult to detect a corner
  • the movement vector of the object may be further detected on the assumption that the object moves.
  • the imaging apparatus 100 according to the third embodiment is different from the first embodiment in that the movement vector is further detected.
  • FIG. 25 is a block diagram illustrating a configuration example of the difference area detection unit 220 according to the third embodiment.
  • the difference area detection unit 220 of the third embodiment differs from the first embodiment in that it further includes a data buffer 223 and a movement vector detection unit 224.
  • the data buffer 223 holds a labeling image, a detection vector, and a stay period.
  • the detection vector and the stay period will be described later.
  • the labeling processing unit 222 supplies a labeling image to the data buffer 223 and the movement vector detection unit 224.
  • the movement vector detection unit 224 detects a movement vector for each difference area.
  • the movement vector detection unit 224 acquires the current labeling image and the past labeling image and movement vector held in the data buffer 223.
  • the movement vector detection unit 224 pays attention to one of the difference areas in the past input image in order, and searches the current difference area corresponding to the difference area.
  • the movement vector detection unit 224 calculates, as the current movement vector, a vector from the reference coordinates of the focused past difference area to the respective reference coordinates of the current difference area.
  • the reference coordinates are, for example, the center and the center of gravity of the difference area.
  • the movement vector detection unit 224 calculates the area of the past difference area of interest and the area of each of the current difference areas.
  • the movement vector detection unit 224 gives priority to a difference area with a small variation in area and movement vector and acquires it as a corresponding area.
  • the movement vector detection unit 224 obtains an evaluation value for each current difference area by the following formula, and acquires a difference image having the highest evaluation value as a corresponding area.
  • v_k is a movement vector of the difference area corresponding to the current label k.
  • v_k ′ is a movement vector of the past difference region of interest.
  • S_k is the area of the difference area corresponding to the current label k.
  • S_k ′ is the area of the past difference region of interest.
  • F (k) is an evaluation value of the difference area corresponding to the label k.
  • the movement vector is set to an initial value.
  • the movement vector detection unit 224 updates the label of the current difference area to the same label as the corresponding past difference area.
  • the movement vector detection unit 224 counts up the stay period of the updated label by one frame. In the first input image, the stay period of each label is set to an initial value. Then, the movement vector detection unit supplies the labeling image with the updated label to the data buffer 223, the input image feature amount acquisition unit 230, and the reference image feature amount acquisition unit 240, and the movement vector and the stay period are displayed in the data buffer 223 and This is supplied to the control unit 140.
  • the control unit 140 As described above, by associating the past difference area and the current difference area based on the area and the movement vector, it is possible to track the moving body moving at a constant speed.
  • FIG. 26 is a block diagram illustrating a configuration example of the movement vector and the stay period in the third embodiment. As illustrated at the same time, the movement vector detection unit 224 obtains a movement vector and a stay period for each label.
  • FIG. 27 is a flowchart illustrating an example of a difference area detection process according to the third embodiment.
  • the difference area detection process of the third embodiment is different from that of the first embodiment in that steps S913 and S914 are further executed.
  • the difference area detection unit 220 generates a labeling image (step S912) and detects a movement vector for each label (step S913). Further, a stay period is obtained for each label (step S914). After step S914, the difference area detection unit 220 ends the difference area detection process.
  • the imaging apparatus 100 can easily track the moving body in order to obtain the movement vector of the difference area. Thereby, the imaging device 100 can perform a useful process for crime prevention, such as obtaining the stay period of the moving body and reproducing only the input image of the stay period.
  • the processing procedure described in the above embodiment may be regarded as a method having a series of these procedures, and a program for causing a computer to execute these series of procedures or a recording medium storing the program. You may catch it.
  • a recording medium for example, a CD (Compact Disc), an MD (MiniDisc), a DVD (Digital Versatile Disc), a memory card, a Blu-ray disc (Blu-ray (registered trademark) Disc), or the like can be used.
  • this technique can also take the following structures.
  • a comparison area determination unit that determines a comparison area to be compared in each of the input image and the predetermined reference image;
  • An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image An acquisition unit;
  • a reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
  • a similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
  • An image processing apparatus comprising: an object estimation unit configured to estimate the comparison region that is not similar based on the region similarity in
  • the image processing apparatus detects a plurality of corners in the comparison region and sets the target pixel.
  • the input image feature amount acquisition unit generates a random number corresponding to any one of the pixels in the comparison area and sets the pixel corresponding to the random number as the pixel of interest. Processing equipment.
  • the input image feature quantity acquisition unit extracts a pixel within a predetermined distance from the target pixel as the surrounding pixel.
  • the input image feature quantity acquisition unit extracts a pixel whose pixel value is not similar to the target pixel from the pixels in the comparison area, and sets the pixel as the surrounding pixel.
  • the object estimation unit may calculate the input image feature amount acquired for each of the plurality of target pixels and the reference image feature amount acquired for each of the corresponding pixels whose coordinates coincide with the target pixel. Whether the local similarity that is the similarity is higher than a predetermined local determination threshold value is determined for each pixel of interest, and a value corresponding to the number of times that the local similarity is determined to be higher than the local determination threshold value is the region similarity.
  • the image processing apparatus according to any one of (1) to (5), acquired as a degree.
  • the comparison area determination unit determines the comparison area in each of the two input images and the reference image, and calculates a vector from one comparison area to the other comparison area of the two input images.
  • the image processing device according to any one of (1) to (7), wherein the image processing device is detected as a movement vector.
  • an imaging unit that captures an input image;
  • a comparison area determination unit for determining a comparison area to be compared in each of the input image and the predetermined reference image;
  • An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image
  • An acquisition unit A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
  • a similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
  • An imaging apparatus comprising: an object estimation unit configured to estimate the comparison area that is not similar based on the area similarity in the input
  • (10) a comparison region determination procedure in which the comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
  • the input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image.
  • a reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image.
  • a reference image feature acquisition procedure Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
  • An object processing method comprising: an object estimation unit configured to estimate, as an object region, the comparison region that is not similar based on the region similarity in the input image.
  • (11) a comparison region determination procedure in which the comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
  • the input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image.
  • a reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image.
  • a reference image feature acquisition procedure Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
  • DESCRIPTION OF SYMBOLS 100 Image pick-up device 110 Imaging lens 120 Image pick-up element 130 Recording part 140 Control part 200 Image processing part 210 Noise removal part 220 Difference area detection part 221 Difference image generation part 222 Labeling process part 223 Data buffer 224 Movement vector detection part 230 Input image Feature amount acquisition unit 231 Corner detection unit 232, 235, 242 Local feature amount acquisition unit 233 Random number generation unit 234 Surrounding pixel extraction unit 240 Reference image feature amount acquisition unit 241 Difference acquisition unit 250 Similarity acquisition unit 260 Object estimation unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

This invention infers object regions accurately. A comparison-region determination unit determines a comparison region in which to compare an input image and a given reference image. For each of a plurality of pixels of interest in the comparison region of the input image, an input-image feature-quantity acquisition unit acquires an input-image feature quantity consisting of a value that is a function of the differences between the value of that pixel of interest and the values of surrounding pixels surrounding that pixel of interest. A reference-image feature-quantity acquisition unit acquires reference-image feature quantities each consisting of a value that is a function of the differences between the value of a corresponding pixel in the reference image that has the same coordinates as a pixel of interest and the values of pixels that have the same coordinates as the surrounding pixels surrounding said pixel of interest. On the basis of the input-image feature quantities and the reference-image feature quantities, a similarity-degree acquisition unit acquires a region similarity degree consisting of the degree of similarity between the comparison region of the input image and the comparison region of the reference image. An object inference unit infers that comparison regions of the input image that region similarity degrees indicate are not similar to the reference image are object regions.

Description

画像処理装置、撮像装置、画像処理方法およびプログラムImage processing apparatus, imaging apparatus, image processing method, and program
 本技術は、画像処理装置、撮像装置、画像処理方法およびプログラムに関する。詳しくは、物体の領域を推定する画像処理装置、撮像装置、および、これらにおける処理方法ならびに当該方法をコンピュータに実行させるプログラムに関する。 The present technology relates to an image processing device, an imaging device, an image processing method, and a program. More specifically, the present invention relates to an image processing device, an imaging device, a processing method in these, and a program for causing a computer to execute the method.
 従来より、監視や計測などを行う目的で、画像処理により物体の領域を推定する監視カメラが用いられている。例えば、背景画像と入力画像との差分を求める背景差分法により物体の領域を推定する監視カメラが提案されている(例えば、特許文献1参照。)。この監視カメラは、入力画像および背景画像において、画素値の差分が閾値より大きい差分領域を背景差分法により求め、それらの差分領域の類似度を正規化相互相関マッチングにより求める。そして、監視カメラは、類似度が閾値より高い場合に差分領域が光や影などの外乱の領域であると推定し、そうでない場合に差分領域が不審物や不審人物などの物体の領域であると推定する。 Conventionally, for the purpose of monitoring or measuring, a surveillance camera that estimates an object region by image processing has been used. For example, a surveillance camera that estimates an object region by a background difference method for obtaining a difference between a background image and an input image has been proposed (see, for example, Patent Document 1). This surveillance camera obtains a difference area in which the pixel value difference is larger than the threshold value in the input image and the background image by the background difference method, and obtains the similarity between these difference areas by normalized cross-correlation matching. When the similarity is higher than the threshold, the monitoring camera estimates that the difference area is a disturbance area such as light or shadow, and otherwise the difference area is an object area such as a suspicious object or a suspicious person. Estimated.
特開2007-201933号公報JP 2007-201933 A
 しかしながら、上述の従来技術では、入力画像にノイズが生じた際に、そのノイズの影響により差分領域の類似度が低くなってしまう。このため、監視カメラは、外乱を不審物等であると誤って推定するおそれがある。このように、上述の監視カメラでは、物体の領域を正確に推定することができないという問題がある。 However, in the above-described conventional technique, when noise occurs in the input image, the similarity of the difference area is lowered due to the influence of the noise. For this reason, there is a possibility that the surveillance camera erroneously estimates that the disturbance is a suspicious object or the like. As described above, the above-described surveillance camera has a problem that the area of the object cannot be accurately estimated.
 本技術はこのような状況に鑑みて生み出されたものであり、物体の領域を正確に推定することを目的とする。 This technology was created in view of such a situation, and aims to accurately estimate an object region.
 本技術は、上述の問題点を解消するためになされたものであり、その第1の側面は、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、上記入力画像において上記比較領域内の複数の注目画素のそれぞれについて上記注目画素の周囲の周囲画素と上記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、上記基準画像において上記注目画素に座標が一致する対応画素と上記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、上記入力画像内の上記比較領域と上記基準画像内の上記比較領域との類似度を領域類似度として上記入力画像特徴量および上記基準画像特徴量に基づいて取得する類似度取得部と、上記入力画像において上記領域類似度に基づいて類似していない上記比較領域を物体の領域として推定する物体推定部とを具備する画像処理装置、および、これらにおける処理方法ならびに当該方法をコンピュータに実行させるプログラムである。これにより、類似していない比較領域が物体の領域として推定されるという作用をもたらす。 The present technology has been made to solve the above-described problems, and a first aspect of the present technology is a comparison region determination unit that determines a comparison region to be compared in each of an input image and a predetermined reference image. An input image feature that acquires a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image as an input image feature amount A reference image feature that acquires a value corresponding to a difference between pixel values of a quantity acquisition unit and a corresponding pixel whose coordinates match the pixel of interest in the reference image and a pixel whose coordinates match the surrounding pixel as a reference image feature quantity Based on the input image feature quantity and the reference image feature quantity, the similarity between the comparison area in the input image and the comparison area in the input image and the comparison area in the reference image is set as the area similarity. An image processing apparatus comprising: a similarity acquisition unit to be obtained; an object estimation unit that estimates the comparison region that is not similar based on the region similarity in the input image as a region of the object; A program for causing a computer to execute the method. This brings about the effect that a comparison region that is not similar is estimated as the region of the object.
 また、この第1の側面において、上記入力画像特徴量取得部は、上記比較領域において複数のコーナーを検出して上記注目画素としてもよい。これにより、比較領域において複数のコーナーが注目画素として検出されるという作用をもたらす。 Also, in the first aspect, the input image feature amount acquisition unit may detect a plurality of corners in the comparison region and set the pixel of interest. This brings about the effect that a plurality of corners are detected as the target pixel in the comparison region.
 また、この第1の側面において、上記入力画像特徴量取得部は、上記比較領域内の画素のいずれかに対応する乱数を生成して当該乱数に対応する上記画素を上記注目画素としてもよい。これにより、乱数に対応する画素が注目画素とされるという作用をもたらす。 Also, in this first aspect, the input image feature quantity acquisition unit may generate a random number corresponding to any of the pixels in the comparison region and set the pixel corresponding to the random number as the target pixel. This brings about the effect that the pixel corresponding to the random number is set as the target pixel.
 また、この第1の側面において、上記入力画像特徴量取得部は、上記注目画素から所定距離内の画素を抽出して上記周囲画素としてもよい。これにより、注目画素から所定距離内の画素が周囲画素として抽出されるという作用をもたらす。 Further, in the first aspect, the input image feature quantity acquisition unit may extract pixels within a predetermined distance from the pixel of interest as the surrounding pixels. This brings about the effect that pixels within a predetermined distance from the target pixel are extracted as surrounding pixels.
 また、この第1の側面において、上記入力画像特徴量取得部は、上記比較領域内の画素の中から上記注目画素に対して画素値が類似しない画素を抽出して上記周囲画素としてもよい。これにより、注目画素に対して画素値が類似しない画素が周囲画素として抽出されるという作用をもたらす。 Further, in the first aspect, the input image feature amount acquisition unit may extract a pixel whose pixel value is not similar to the target pixel from the pixels in the comparison area, and use the extracted pixel as the surrounding pixel. This brings about the effect that pixels whose pixel values are not similar to the target pixel are extracted as surrounding pixels.
 また、この第1の側面において、上記物体推定部は、上記複数の注目画素の各々について取得された上記入力画像特徴量と上記注目画素に座標が一致する上記対応画素の各々について取得された上記基準画像特徴量との類似度である局所類似度が所定の局所判定閾値より高いか否かを上記注目画素ごとに判定して上記局所類似度が上記局所判定閾値より高いと判定した回数に応じた値を上記領域類似度として取得してもよい。これにより、入力画像特徴量と基準画像特徴量との局所類似度が局所判定閾値より高いと判定された回数に応じた値が領域類似度として取得されるという作用をもたらす。 In the first aspect, the object estimation unit acquires the input image feature amount acquired for each of the plurality of target pixels and the corresponding pixels whose coordinates coincide with the target pixel. Depending on the number of times it is determined for each pixel of interest whether the local similarity that is the similarity to the reference image feature amount is higher than a predetermined local determination threshold, and the local similarity is higher than the local determination threshold The obtained value may be acquired as the region similarity. Accordingly, an effect is obtained in which a value corresponding to the number of times that the local similarity between the input image feature quantity and the reference image feature quantity is determined to be higher than the local determination threshold is acquired as the area similarity.
 また、この第1の側面において、上記入力画像および上記基準画像において画素値が類似しない画素を検出して当該検出した画素からなる領域を上記比較領域として決定してもよい。これにより、入力画像および基準画像において画素値が類似しない画素からなる領域が比較領域として決定されるという作用をもたらす。 Further, in the first aspect, pixels having similar pixel values in the input image and the reference image may be detected, and an area including the detected pixels may be determined as the comparison area. This brings about the effect that an area composed of pixels whose pixel values are not similar in the input image and the reference image is determined as the comparison area.
 また、この第1の側面において、2つの上記入力画像のそれぞれと上記基準画像において上記比較領域を決定して上記2つの入力画像の一方の上記比較領域から他方の上記比較領域へのベクトルを移動ベクトルとして検出してもよい。これにより、2つの入力画像の一方の比較領域から他方の比較領域へのベクトルが移動ベクトルとして検出されるという作用をもたらす。 Further, in the first aspect, the comparison area is determined in each of the two input images and the reference image, and a vector from one comparison area to the other comparison area of the two input images is moved. It may be detected as a vector. This brings about the effect that a vector from one comparison area of the two input images to the other comparison area is detected as a movement vector.
 また、本技術の第2の側面は、入力画像を撮像する撮像部と、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、上記入力画像において上記比較領域内の複数の注目画素のそれぞれについて上記注目画素の周囲の周囲画素と上記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、上記基準画像において上記注目画素に座標が一致する対応画素と上記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、上記入力画像内の上記比較領域と上記基準画像内の上記比較領域との類似度を領域類似度として上記入力画像特徴量および上記基準画像特徴量に基づいて取得する類似度取得部と、上記入力画像において上記領域類似度に基づいて類似していない上記比較領域を物体の領域として推定する物体推定部とを具備する撮像装置である。これにより、類似していない比較領域が物体の領域として推定されるという作用をもたらす。 A second aspect of the present technology includes an imaging unit that captures an input image, a comparison region determination unit that determines a comparison region to be compared in each of the input image and a predetermined reference image, and the input image An input image feature amount acquisition unit that acquires, as an input image feature amount, a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region; A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinate matches the pixel of interest in a reference image and a pixel whose coordinate matches the surrounding pixel; A similarity acquisition unit that acquires the similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature amount and the reference image feature amount An imaging apparatus including an object estimation unit for estimating the comparison area in the input image is not similar based on the region similarity as an area of the object. This brings about the effect that a comparison region that is not similar is estimated as the region of the object.
 本技術によれば、物体の領域を正確に推定することができるという優れた効果を奏し得る。なお、ここに記載された効果は必ずしも限定されるものではなく、本開示中に記載されたいずれかの効果であってもよい。 According to the present technology, an excellent effect that the area of the object can be accurately estimated can be achieved. Note that the effects described here are not necessarily limited, and may be any of the effects described in the present disclosure.
第1の実施の形態における撮像装置の一構成例を示すブロック図である。1 is a block diagram illustrating a configuration example of an imaging apparatus according to a first embodiment. 第1の実施の形態における画像処理部の一構成例を示すブロック図である。It is a block diagram which shows one structural example of the image process part in 1st Embodiment. 第1の実施の形態における差分領域検出部の一構成例を示すブロック図である。It is a block diagram which shows one structural example of the difference area | region detection part in 1st Embodiment. 第1の実施の形態における基準画像の一例を示す図である。It is a figure which shows an example of the reference | standard image in 1st Embodiment. 第1の実施の形態における入力画像の一例を示す図である。It is a figure which shows an example of the input image in 1st Embodiment. 第1の実施の形態における差分画像の一例を示す図である。It is a figure which shows an example of the difference image in 1st Embodiment. 第1の実施の形態におけるラべリング画像の一例を示す図である。It is a figure which shows an example of the labeling image in 1st Embodiment. 第1の実施の形態における入力画像特徴量取得部の一構成例を示すブロック図である。It is a block diagram which shows the example of 1 structure of the input image feature-value acquisition part in 1st Embodiment. 第1の実施の形態における入力画像内の注目画素および周囲画素の一例を示す図である。It is a figure which shows an example of the attention pixel and surrounding pixel in the input image in 1st Embodiment. 第1の実施の形態における基準画像内の注目画素および周囲画素の一例を示す図である。It is a figure which shows an example of the attention pixel and surrounding pixel in the reference | standard image in 1st Embodiment. 第1の実施の形態における局所特徴ベクトルの取得方法を説明するための図である。It is a figure for demonstrating the acquisition method of the local feature vector in 1st Embodiment. 第1の実施の形態における局所特徴ベクトルの一例を示す図である。It is a figure which shows an example of the local feature vector in 1st Embodiment. 第1の実施の形態における類似度の一例を示す図である。It is a figure which shows an example of the similarity in 1st Embodiment. 第1の実施の形態における推定結果の一例を示す図である。It is a figure which shows an example of the estimation result in 1st Embodiment. 第1の実施の形態における撮像装置の動作の一例を示すフローチャートである。3 is a flowchart illustrating an example of an operation of the imaging apparatus according to the first embodiment. 第1の実施の形態における差分領域検出処理の一例を示すフローチャートである。It is a flowchart which shows an example of the difference area | region detection process in 1st Embodiment. 第1の実施の形態における特徴量取得処理の一例を示すフローチャートである。It is a flowchart which shows an example of the feature-value acquisition process in 1st Embodiment. 第2の実施の形態における入力画像特徴量取得部の一構成例を示すブロック図である。It is a block diagram which shows the example of 1 structure of the input image feature-value acquisition part in 2nd Embodiment. 第2の実施の形態における基準画像特徴量取得部の一構成例を示すブロック図である。It is a block diagram which shows the example of 1 structure of the reference | standard image feature-value acquisition part in 2nd Embodiment. 第2の実施の形態における画素対の抽出方法を説明するための図である。It is a figure for demonstrating the extraction method of the pixel pair in 2nd Embodiment. 第2の実施の形態における基準画像の差分領域の一部を拡大した図である。It is the figure which expanded a part of difference area | region of the reference | standard image in 2nd Embodiment. 第2の実施の形態における入力画像の差分領域の一部を拡大した図である。It is the figure which expanded a part of difference area of the input image in 2nd Embodiment. 第2の実施の形態における局所特徴量の一例を示す図である。It is a figure which shows an example of the local feature-value in 2nd Embodiment. 第2の実施の形態における特徴量取得処理の一例を示すフローチャートである。It is a flowchart which shows an example of the feature-value acquisition process in 2nd Embodiment. 第3の実施の形態における差分領域検出部の一構成例を示すブロック図である。It is a block diagram which shows the example of 1 structure of the difference area | region detection part in 3rd Embodiment. 第3の実施の形態における移動ベクトルおよび滞在期間の一構成例を示すブロック図である。It is a block diagram which shows the example of 1 structure of the movement vector in 3rd Embodiment, and a stay period. 第3の実施の形態における差分領域検出処理の一例を示すフローチャートである。It is a flowchart which shows an example of the difference area | region detection process in 3rd Embodiment.
 以下、本技術を実施するための形態(以下、実施の形態と称する)について説明する。説明は以下の順序により行う。
 1.第1の実施の形態(特徴量から領域の類似度を求める例)
 2.第2の実施の形態(特徴量から領域の類似度を求める際にランダムに注目画素を選択する例)
 3.第3の実施の形態(特徴量から領域の類似度を求める前に移動ベクトルを検出する例)
Hereinafter, modes for carrying out the present technology (hereinafter referred to as embodiments) will be described. The description will be made in the following order.
1. First embodiment (example of obtaining similarity of region from feature amount)
2. Second embodiment (an example of selecting a pixel of interest randomly when obtaining the similarity of a region from a feature amount)
3. Third Embodiment (Example in which a movement vector is detected before the similarity of a region is obtained from a feature amount)
 <1.第1の実施の形態>
 [撮像装置の構成例]
 図1は、第1の実施の形態における撮像装置100の一構成例を示すブロック図である。この撮像装置100は、画像を撮像するものであり、撮像レンズ110、撮像素子120、記録部130、制御部140および画像処理部200を備える。
<1. First Embodiment>
[Configuration example of imaging device]
FIG. 1 is a block diagram illustrating a configuration example of the imaging apparatus 100 according to the first embodiment. The imaging apparatus 100 captures an image and includes an imaging lens 110, an imaging element 120, a recording unit 130, a control unit 140, and an image processing unit 200.
 撮像レンズ110は、光を集光して撮像素子120に導くものである。撮像素子120は、制御部140の制御に従って、撮像レンズ110からの光を電気信号に変換して画像を撮像するものである。この撮像素子120は、画像を撮像するたびに、その画像を入力画像として画像処理部200に信号線129を介して供給する。なお、撮像素子120は、特許請求の範囲に記載の撮像部の一例である。 The imaging lens 110 collects light and guides it to the imaging device 120. The image sensor 120 converts the light from the imaging lens 110 into an electrical signal and captures an image under the control of the control unit 140. Each time the image sensor 120 captures an image, the image sensor 120 supplies the image as an input image to the image processing unit 200 via the signal line 129. The imaging element 120 is an example of an imaging unit described in the claims.
 画像処理部200は、入力画像において物体の領域を推定するものである。この画像処理部200は、推定結果を制御部140に信号線209を介して供給する。また、画像処理部200は、入力画像を記録部130に信号線208を介して供給する。 The image processing unit 200 estimates an object area in the input image. The image processing unit 200 supplies the estimation result to the control unit 140 via the signal line 209. Further, the image processing unit 200 supplies the input image to the recording unit 130 via the signal line 208.
 記録部130は、入力画像および基準画像を記録するものである。例えば、監視対象の場所において、入力画像の撮像前に予め撮像しておいた画像が基準画像として用いられる。 The recording unit 130 records an input image and a reference image. For example, an image captured in advance before the input image is captured at the monitoring target location is used as the reference image.
 制御部140は、撮像装置100全体を制御するものである。この制御部140は、ユーザの操作などに従って撮像を指示する制御信号を生成し、撮像素子120に信号線149を介して供給する。また、制御部140は、推定結果を画像処理部200から受け取り、いずれかの領域が物体の領域であると推定されたのであれば、その旨を報知するアラーム信号を撮像装置100の外部などへ出力する。 The control unit 140 controls the entire imaging apparatus 100. The control unit 140 generates a control signal for instructing imaging in accordance with a user operation or the like, and supplies the control signal to the image sensor 120 via the signal line 149. Further, the control unit 140 receives the estimation result from the image processing unit 200, and if any region is estimated to be an object region, the control unit 140 sends an alarm signal to that effect to the outside of the imaging device 100 or the like. Output.
 なお、画像処理部200を撮像装置100内に設ける構成としているが、撮像装置100と異なる画像処理装置に設けてもよい。この構成において撮像装置100は入力画像を画像処理装置に供給し、画像処理装置は、物体の領域を推定して推定結果を撮像装置200などに供給する。 The image processing unit 200 is provided in the imaging device 100, but may be provided in an image processing device different from the imaging device 100. In this configuration, the imaging apparatus 100 supplies an input image to the image processing apparatus, and the image processing apparatus estimates an object region and supplies an estimation result to the imaging apparatus 200 or the like.
 [画像処理部の構成例]
 図2は、第1の実施の形態における画像処理部200の一構成例を示すブロック図である。この画像処理部200は、ノイズ除去部210、差分領域検出部220、入力画像特徴量取得部230、基準画像特徴量取得部240、類似度取得部250および物体推定部260を備える。
[Configuration example of image processing unit]
FIG. 2 is a block diagram illustrating a configuration example of the image processing unit 200 according to the first embodiment. The image processing unit 200 includes a noise removal unit 210, a difference area detection unit 220, an input image feature amount acquisition unit 230, a reference image feature amount acquisition unit 240, a similarity acquisition unit 250, and an object estimation unit 260.
 ノイズ除去部210は、入力画像に対してノイズを除去する処理を行うものである。このノイズ除去部210は、例えば、所定の遮断周波数より高い高周波数成分を抑制するローパスフィルタを通過させることによりノイズを除去する。このローパスフィルタは、例えば、IIR(Infinite Impulse Response)フィルタや、FIR(Finite Impulse Response)フィルタにより実現される。ノイズ除去部210は、ノイズを除去した入力画像を、記録部130、差分領域検出部220および入力画像特徴量取得部230に供給する。 The noise removing unit 210 performs processing for removing noise from the input image. For example, the noise removing unit 210 removes noise by passing a low-pass filter that suppresses high frequency components higher than a predetermined cutoff frequency. This low-pass filter is realized by, for example, an IIR (Infinite Impulse Response) filter or an FIR (Finite Impulse Response) filter. The noise removal unit 210 supplies the input image from which noise has been removed to the recording unit 130, the difference area detection unit 220, and the input image feature amount acquisition unit 230.
 差分領域検出部220は、入力画像および基準画像において比較すべき領域を決定するものである。この差分領域検出部220は、例えば、入力画像および基準画像において、画素値が類似しない画素を検出して、それらの画素からなる領域を比較領域とする。まず、差分領域検出部220は、座標が同一の画素ごとに画素値の差分の絶対値を検出して、これらの差分からなる差分画像を生成し、所定の二値化閾値により差分画像を二値化する。 The difference area detection unit 220 determines an area to be compared in the input image and the reference image. For example, the difference area detection unit 220 detects pixels whose pixel values are not similar in the input image and the reference image, and sets an area including these pixels as a comparison area. First, the difference area detection unit 220 detects the absolute value of the difference between pixel values for each pixel having the same coordinate, generates a difference image composed of these differences, and converts the difference image into a binary image using a predetermined binarization threshold. Convert to value.
 そして、差分領域検出部220は、その二値化した差分画像において、二値化閾値より差分絶対値が大きいことを示す画素値の画素が連続する領域のそれぞれに、その領域を識別する識別情報をラベルとして割り振るラべリング処理を行う。このラべリング処理において、例えば、水平方向および垂直方向に連続する画素を連結する4連結のアルゴリズムや、それらの方向に加えて斜め方向に連続する画素をさらに連結する8連結のアルゴリズムなどが用いられる。以下、ラベルが付された領域の示す、入力画像および基準画像内の領域を「差分領域」と称する。これらの差分領域が、入力画像および基準画像において、比較すべき領域として比較される。差分領域検出部220は、ラべリング処理を行った画像をラべリング画像として入力画像特徴量取得部230および基準画像特徴量取得部240に供給する。 And the difference area detection part 220 is the identification information which identifies the area | region in each of the area | region where the pixel of the pixel value which shows that a difference absolute value is larger than a binarization threshold value in the binarized difference image. Labeling process to allocate as a label. In this labeling process, for example, a 4-connection algorithm that connects pixels that are continuous in the horizontal direction and the vertical direction, an 8-connection algorithm that further connects pixels that are continuous in an oblique direction in addition to those directions, and the like are used. It is done. Hereinafter, an area in the input image and the reference image indicated by the labeled area is referred to as a “difference area”. These difference areas are compared as areas to be compared in the input image and the reference image. The difference area detection unit 220 supplies the image subjected to the labeling process to the input image feature amount acquisition unit 230 and the reference image feature amount acquisition unit 240 as a labeling image.
 なお、差分領域検出部220は、特許請求の範囲に記載の比較領域決定部の一例である。 The difference area detection unit 220 is an example of a comparison area determination unit described in the claims.
 入力画像特徴量取得部230は、入力画像内の差分領域において特徴量を取得するものである。この入力画像特徴量取得部230は、差分領域内の複数の注目画素を検出し、注目画素の各々について注目画素の周囲の周囲画素と注目画素との画素値の差分を求め、その差分に応じた値を局所特徴量として取得する。ここで、局所特徴量を求める際に比較される画素値は、例えば、輝度値や色差である。入力画像特徴量取得部230は、局所特徴量のそれぞれを類似度取得部250に供給する。 The input image feature amount acquisition unit 230 acquires a feature amount in a difference area in the input image. The input image feature amount acquisition unit 230 detects a plurality of target pixels in the difference region, obtains a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each target pixel, and according to the difference The obtained value is acquired as a local feature amount. Here, the pixel value compared when obtaining the local feature amount is, for example, a luminance value or a color difference. The input image feature amount acquisition unit 230 supplies each of the local feature amounts to the similarity acquisition unit 250.
 基準画像特徴量取得部240は、基準画像内の差分領域において特徴量を取得するものである。この基準画像特徴量取得部240は、入力画像特徴量取得部230から注目画素の座標を受け取り、注目画素に座標が一致する対応画素と、その周囲の周囲画素との画素値の差分を求める。そして、基準画像特徴量取得部240は、その差分に応じた値を局所特徴量として取得し、類似度取得部250に供給する。 The reference image feature amount acquisition unit 240 acquires feature amounts in the difference area in the reference image. The reference image feature amount acquisition unit 240 receives the coordinates of the target pixel from the input image feature amount acquisition unit 230, and obtains a difference in pixel value between the corresponding pixel whose coordinates match the target pixel and the surrounding surrounding pixels. Then, the reference image feature amount acquisition unit 240 acquires a value corresponding to the difference as a local feature amount, and supplies it to the similarity acquisition unit 250.
 類似度取得部250は、入力画像特徴量取得部230および基準画像特徴量取得部240により取得された局所特徴量に基づいて、入力画像および基準画像のそれぞれの差分領域の類似度を領域類似度として求めるものである。この領域類似度は、例えば、領域の類似性の度合いが高いほど高い値であるものとする。領域類似度の取得方法の詳細については後述する。類似度取得部250は、求めた領域類似度を物体推定部260に供給する。 The similarity acquisition unit 250 calculates the similarity between the difference areas of the input image and the reference image based on the local feature acquired by the input image feature acquisition unit 230 and the reference image feature acquisition unit 240. Is what you want. For example, the region similarity is higher as the degree of similarity between regions is higher. Details of the region similarity acquisition method will be described later. The similarity acquisition unit 250 supplies the obtained region similarity to the object estimation unit 260.
 物体推定部260は、領域類似度が所定の領域判定閾値より低い(言い換えれば、類似していない)領域を、基準画像にない不審物や不審物体の領域として推定するものである。物体推定部260は、推定結果を制御部140に供給する。 The object estimation unit 260 estimates a region having a region similarity lower than a predetermined region determination threshold (in other words, a region that is not similar) as a region of a suspicious object or a suspicious object that is not in the reference image. The object estimation unit 260 supplies the estimation result to the control unit 140.
 なお、画像処理部200は、入力画像に対してノイズ除去を行っているが、画質を向上させる処理であれば、ノイズ除去以外の処理を行ってもよい。例えば、画像処理部200は、ノイズ除去の代わりに、コントラストの強調や、電子式の手ブレ補正を行ってもよい。電子式手ブレ補正は、スタビライズ処理とも呼ばれる。あるいは、画像処理部200は、ノイズ除去に加えて、コントラストの強調や電子式の手ブレ補正を行ってもよい。このように、画質を向上させる処理を行うことにより、物体の推定精度を向上させることができる。 The image processing unit 200 performs noise removal on the input image, but may perform processing other than noise removal as long as the image quality is improved. For example, the image processing unit 200 may perform contrast enhancement or electronic camera shake correction instead of noise removal. Electronic camera shake correction is also called stabilization processing. Alternatively, the image processing unit 200 may perform contrast enhancement and electronic camera shake correction in addition to noise removal. As described above, by performing the process of improving the image quality, it is possible to improve the estimation accuracy of the object.
 [差分領域検出部の構成例]
 図3は、第1の実施の形態における差分領域検出部220の一構成例を示すブロック図である。この差分領域検出部220は、差分画像生成部221およびラべリング処理部222を備える。
[Configuration example of difference area detection unit]
FIG. 3 is a block diagram illustrating a configuration example of the difference area detection unit 220 according to the first embodiment. The difference area detection unit 220 includes a difference image generation unit 221 and a labeling processing unit 222.
 差分画像生成部221は、入力画像および基準画像の差分画像を生成するものである。差分画像生成部221は、生成した差分画像を二値化してラべリング処理部222に供給する。 The difference image generation unit 221 generates a difference image between the input image and the reference image. The difference image generation unit 221 binarizes the generated difference image and supplies it to the labeling processing unit 222.
 ラべリング処理部222は、二値化された差分画像に対してラべリング処理を行うものである。ラべリング処理部222は、ラべリング処理を行った画像をラべリング画像として入力画像特徴量取得部230および基準画像特徴量取得部240に供給する。 The labeling processing unit 222 performs a labeling process on the binarized difference image. The labeling processing unit 222 supplies the image subjected to the labeling process to the input image feature amount acquisition unit 230 and the reference image feature amount acquisition unit 240 as a labeling image.
 図4は、第1の実施の形態における基準画像500の一例を示す図である。同図に示すように、基準画像500は、家や樹木などの背景のみを含む。 FIG. 4 is a diagram illustrating an example of the reference image 500 according to the first embodiment. As shown in the figure, the reference image 500 includes only backgrounds such as houses and trees.
 図5は、第1の実施の形態における入力画像510の一例を示す図である。同図に示すように、入力画像510には、背景の他、被写体511、512および513が含まれる。被写体511は、例えば、不審物や不審人物である。被写体512は、例えば、サーチライトなどの光である。被写体513は、例えば、雲の影である。 FIG. 5 is a diagram illustrating an example of the input image 510 according to the first embodiment. As shown in the figure, the input image 510 includes subjects 511, 512 and 513 in addition to the background. The subject 511 is, for example, a suspicious object or a suspicious person. The subject 512 is light such as search light, for example. The subject 513 is, for example, a cloud shadow.
 図6は、第1の実施の形態における差分画像520の一例を示す図である。この差分画像520は、基準画像500および入力画像510の差分画像を二値化したものである。同図に示すように、差分画像520は、被写体511、512および513に対応する差分領域521、522および523を含む。これらの差分領域は、入力画像510において、基準画像500に対する画素値の差分の絶対値が所定の二値化閾値より大きい領域である。 FIG. 6 is a diagram illustrating an example of the difference image 520 according to the first embodiment. The difference image 520 is obtained by binarizing the difference image between the reference image 500 and the input image 510. As shown in the figure, the difference image 520 includes difference areas 521, 522, and 523 corresponding to the subjects 511, 512, and 513. These difference areas are areas in the input image 510 where the absolute value of the pixel value difference with respect to the reference image 500 is larger than a predetermined binarization threshold.
 なお、撮像装置100は、入力画像および基準画像のそれぞれの全体から差分画像を生成しているが、この構成に限定されない。例えば、撮像装置100は、電子ズームなどを行う際に、画像の一部を抽出範囲とし、入力画像および基準画像の抽出範囲から差分画像を生成してもよい。 The imaging apparatus 100 generates a difference image from the entire input image and reference image, but is not limited to this configuration. For example, when performing an electronic zoom or the like, the imaging apparatus 100 may generate a difference image from an extraction range of an input image and a reference image using a part of the image as an extraction range.
 図7は、第1の実施の形態におけるラべリング画像530の一例を示す図である。このラべリング画像530は、被写体511、512および513に対応する差分領域531、532および533を含む。差分領域531には、例えば、「1」のラベルが割り当てられる。また、差分領域532には、例えば、「2」のラベルが割り当てられ、差分領域533には、例えば、「3」のラベルが割り当てられる。 FIG. 7 is a diagram illustrating an example of a labeling image 530 according to the first embodiment. The labeling image 530 includes difference areas 531, 532, and 533 corresponding to the subjects 511, 512, and 513. For example, a label “1” is assigned to the difference area 531. Further, for example, a label “2” is assigned to the difference area 532, and a label “3” is assigned to the difference area 533, for example.
 図8は、第1の実施の形態における入力画像特徴量取得部230の一構成例を示すブロック図である。この入力画像特徴量取得部230は、コーナー検出部231および局所特徴量取得部232を備える。 FIG. 8 is a block diagram illustrating a configuration example of the input image feature amount acquisition unit 230 according to the first embodiment. The input image feature amount acquisition unit 230 includes a corner detection unit 231 and a local feature amount acquisition unit 232.
 コーナー検出部231は、差分領域においてコーナーを注目画素として検出するものである。このコーナー検出部231は、入力画像内の差分領域のそれぞれにおいて、キャニーフィルタなどを用いてエッジを検出し、検出したエッジの交点であるコーナーを注目画素として検出する。コーナー検出部231は、注目画素の座標を局所特徴量取得部232および基準画像特徴量取得部240に供給する。 The corner detection unit 231 detects a corner as a target pixel in the difference area. The corner detection unit 231 detects an edge using a canny filter or the like in each difference region in the input image, and detects a corner that is an intersection of the detected edges as a target pixel. The corner detection unit 231 supplies the coordinates of the target pixel to the local feature amount acquisition unit 232 and the reference image feature amount acquisition unit 240.
 局所特徴量取得部232は、入力画像において局所特徴量を取得するものである。ここで、注目画素と、その注目画素の周囲の周囲画素との差分に応じた値が局所特徴量として求められる。局所特徴量取得部232は、例えば、座標のユークリッド距離が21/2以下である8個の周囲画素のそれぞれについて、周囲画素の画素値から注目画素の画素値を引いた差分を算出し、差分に応じた値を局所特徴量として求める。そして、局所特徴量取得部232は、注目画素と8個の周囲画素とからなるグループごとに、局所特徴量からなる局所特徴ベクトルを生成して類似度取得部250に供給する。 The local feature amount acquisition unit 232 acquires a local feature amount in the input image. Here, a value corresponding to the difference between the target pixel and surrounding pixels around the target pixel is obtained as the local feature amount. The local feature amount acquisition unit 232 calculates, for example, a difference obtained by subtracting the pixel value of the target pixel from the pixel value of the surrounding pixel for each of the eight surrounding pixels whose coordinate Euclidean distance is 2 1/2 or less, A value corresponding to the difference is obtained as a local feature amount. Then, the local feature amount acquisition unit 232 generates a local feature vector including the local feature amount for each group including the target pixel and the eight surrounding pixels, and supplies the local feature vector to the similarity acquisition unit 250.
 なお、基準画像特徴量取得部240の構成は、基準画像から局所特徴ベクトルを生成する点以外は、局所特徴量取得部232と同様である。 The configuration of the reference image feature quantity acquisition unit 240 is the same as that of the local feature quantity acquisition unit 232 except that a local feature vector is generated from the reference image.
 図9は、第1の実施の形態における入力画像510内の注目画素および周囲画素の一例を示す図である。差分領域511、512および513のそれぞれにおいて、入力画像特徴量取得部230は、コーナーを注目画素611として検出する。そして、入力画像特徴量取得部230は、その注目画素611の周囲の8画素を周囲画素として抽出する。同図において、点線で囲まれたグループ612は、注目画素611と、その周囲の周囲画素とからなるグループである。 FIG. 9 is a diagram illustrating an example of a pixel of interest and surrounding pixels in the input image 510 according to the first embodiment. In each of the difference areas 511, 512, and 513, the input image feature amount acquisition unit 230 detects a corner as the target pixel 611. Then, the input image feature amount acquisition unit 230 extracts eight pixels around the pixel of interest 611 as surrounding pixels. In the figure, a group 612 surrounded by a dotted line is a group including a pixel of interest 611 and surrounding pixels around it.
 図10は、第1の実施の形態における基準画像500内の注目画素および周囲画素の一例を示す図である。基準画像特徴量取得部240は、注目画素611と座標が同一の対応画素601と、その対応画素601の周囲の8個の周囲画素とを抽出する。同図において、点線で囲まれたグループ602は、対応画素601と、その周囲の周囲画素とからなるグループである。 FIG. 10 is a diagram illustrating an example of a pixel of interest and surrounding pixels in the reference image 500 according to the first embodiment. The reference image feature amount acquisition unit 240 extracts a corresponding pixel 601 having the same coordinates as the target pixel 611 and eight surrounding pixels around the corresponding pixel 601. In the figure, a group 602 surrounded by a dotted line is a group including a corresponding pixel 601 and surrounding pixels around it.
 図11は、第1の実施の形態における局所特徴ベクトルの取得方法を説明するための図である。入力画像特徴量取得部230は、注目画素611と周囲の8個の周囲画素とを抽出する。そして、入力画像特徴量取得部230は、周囲画素ごとに、その周囲画素の画素値から注目画素611の画素値を引いた差分を算出する。例えば、差分が+30乃至-30の一定の範囲内である場合には、入力画像特徴量取得部230は、「0」の値を局所特徴量として取得する。また、入力画像特徴量取得部230は、差分が+30より大きい場合には、「1」の値を局所特徴量として取得し、差分が-30より小さい場合には「-1」の値を局所特徴量として取得する。これらの局所特徴量からなる局所特徴ベクトルが、グループ612ごとに求められる。周囲画素が8個である場合には、局所特徴ベクトルのそれぞれは、8個の局所特徴量を含む。 FIG. 11 is a diagram for explaining a local feature vector acquisition method according to the first embodiment. The input image feature amount acquisition unit 230 extracts the pixel of interest 611 and eight surrounding pixels. Then, the input image feature amount acquisition unit 230 calculates, for each peripheral pixel, a difference obtained by subtracting the pixel value of the target pixel 611 from the pixel value of the peripheral pixel. For example, when the difference is within a certain range of +30 to −30, the input image feature quantity acquisition unit 230 acquires a value of “0” as a local feature quantity. Further, the input image feature quantity acquisition unit 230 acquires the value “1” as the local feature quantity when the difference is greater than +30, and the value “−1” when the difference is less than −30. Acquired as a feature value. A local feature vector including these local feature amounts is obtained for each group 612. When the number of surrounding pixels is 8, each of the local feature vectors includes 8 local feature amounts.
 例えば、注目画素の画素値が「65」で、8個の周囲画素の画素値がそれぞれ「30」、「60」、「65」、「60」、「140」、「60」、「200」および「60」である場合を考える。この場合、局所特徴量として、「-1」、「0」、「0」、「0」、「1」、「0」、「1」および「0」が求められる。 For example, the pixel value of the target pixel is “65”, and the pixel values of the eight surrounding pixels are “30”, “60”, “65”, “60”, “140”, “60”, “200”, respectively. And “60”. In this case, “−1”, “0”, “0”, “0”, “1”, “0”, “1”, and “0” are obtained as local feature amounts.
 図11に例示したように、撮像装置100は、画素値の差分に応じた局所特徴量を求め、その局所特徴量から領域類似度を求めているため、ノイズ耐性が高くなる。例えば、ノイズのない入力画像内の注目画素の画素値が「65」で、ある周囲画素の画素値が「30」である場合、それらの差分が-30より低いため、局所特徴量は、「-1」になる。ここで、ノイズ除去部210で除去することができないほどのノイズにより注目画素の画素値が「67」に変動しても、差分は依然として-30より低いため、局所特徴量は、「-1」のままで値が変わらない。このため、領域類似度の変動を抑制することができる。 As illustrated in FIG. 11, the imaging apparatus 100 obtains a local feature amount corresponding to a difference in pixel value, and obtains a region similarity from the local feature amount, so noise resistance is increased. For example, when the pixel value of the target pixel in the input image without noise is “65” and the pixel value of a certain surrounding pixel is “30”, the difference between them is lower than −30. −1 ”. Here, even if the pixel value of the target pixel fluctuates to “67” due to noise that cannot be removed by the noise removing unit 210, the difference is still lower than −30, and thus the local feature amount is “−1”. The value does not change. For this reason, the fluctuation | variation of area | region similarity can be suppressed.
 図12は、第1の実施の形態における局所特徴ベクトルの一例を示す図である。入力画像特徴量取得部230は、入力画像においてラベルに対応する差分領域ごとに複数のグループを抽出し、それらのグループのそれぞれについて、局所特徴ベクトルを求める。また、基準画像特徴量取得部240は、基準画像においてラベル(差分領域)ごとに複数のグループを抽出し、それらのグループのそれぞれについて、局所特徴ベクトルを求める。 FIG. 12 is a diagram showing an example of local feature vectors in the first embodiment. The input image feature quantity acquisition unit 230 extracts a plurality of groups for each difference area corresponding to the label in the input image, and obtains a local feature vector for each of these groups. Further, the reference image feature amount acquisition unit 240 extracts a plurality of groups for each label (difference area) in the reference image, and obtains a local feature vector for each of these groups.
 図13は、第1の実施の形態における類似度の一例を示す図である。類似度取得部250は、入力画像において取得された局所特徴量と基準画像において取得された局所特徴量との類似度を局所類似度としてグループごとに求める。局所類似度は、例えば、正規化相互相関マッチングにより求められる。そして、類似度取得部250は、局所類似度が所定の局所判定閾値より高いか否かをグループごとに判定し、局所類似度が局所判定閾値より高いと判定された回数に応じた値を領域類似度として求める。例えば、差分領域において局所類似度が局所判定閾値より高いと判定された回数を、差分領域内のグループ数により除した値が領域類似度としてラベル(差分領域)ごとに求められる。 FIG. 13 is a diagram illustrating an example of the degree of similarity according to the first embodiment. The similarity acquiring unit 250 obtains the similarity between the local feature acquired in the input image and the local feature acquired in the reference image for each group as the local similarity. The local similarity is obtained by, for example, normalized cross correlation matching. Then, the similarity acquisition unit 250 determines whether or not the local similarity is higher than a predetermined local determination threshold for each group, and sets a value corresponding to the number of times the local similarity is determined to be higher than the local determination threshold as a region. Calculate as similarity. For example, a value obtained by dividing the number of times that the local similarity is higher than the local determination threshold in the difference area by the number of groups in the difference area is obtained for each label (difference area) as the area similarity.
 なお、類似度取得部250は、局所類似度が局所判定閾値より高いと判定された回数そのものを局所類似度として求めてもよい。この場合には、コーナー検出部231は、差分領域のそれぞれの注目画素の個数を同一にすればよい。例えば、コーナー検出部231は、差分領域ごとに全てのコーナーを検出し、最も検出数の少ない差分領域に合わせて、他の領域の注目画素を削減する。 Note that the similarity acquisition unit 250 may determine the number of times the local similarity is determined to be higher than the local determination threshold as the local similarity. In this case, the corner detection part 231 should just make the number of each attention pixels of a difference area the same. For example, the corner detection unit 231 detects all corners for each difference area, and reduces the target pixel in other areas in accordance with the difference area with the smallest number of detections.
 また、類似度取得部250は、正規化相互相関マッチングにより局所類似度を求めているが、この構成に限定されない。例えば、類似度取得部250は、差の絶対値の和であるSAD(Sum of Absolute Differences)を局所類似度として求めてもよい。また、類似度取得部250は、差の二乗和であるSSD(Sum of Squared Differences)を局所類似度として求めてもよい。 Further, although the similarity acquisition unit 250 obtains the local similarity by the normalized cross correlation matching, it is not limited to this configuration. For example, the similarity obtaining unit 250 may obtain SAD (Sum of Absolute Differences), which is the sum of absolute values of differences, as the local similarity. The similarity acquisition unit 250 may obtain SSD (SumSof Squared Differences), which is the sum of squares of differences, as the local similarity.
 図14は、第1の実施の形態における推定結果の一例を示す図である。物体推定部260は、領域類似度が所定の領域判定閾値より低いラベル(差分領域)を不審物や不審人物の領域であると推定する。一方、領域類似度が所定の領域判定閾値以上のラベルは、その領域のテクスチャに変動がなく、明るさが全体的に変わったにすぎない領域を示す。このような領域は、光や影などの外乱が生じた領域と判断される。言い換えれば、不審物等でないものと判断される。 FIG. 14 is a diagram illustrating an example of an estimation result in the first embodiment. The object estimation unit 260 estimates a label (difference area) whose area similarity is lower than a predetermined area determination threshold as an area of a suspicious object or a suspicious person. On the other hand, a label whose region similarity is equal to or higher than a predetermined region determination threshold indicates a region where the texture of the region is not changed and the brightness is merely changed as a whole. Such a region is determined as a region where disturbance such as light or shadow has occurred. In other words, it is determined that it is not a suspicious object.
 [撮像装置の動作例]
 図15は、第1の実施の形態における撮像装置100の動作の一例を示すフローチャートである。この動作は、例えば、入力画像が撮像されるたびに実行される。
[Operation example of imaging device]
FIG. 15 is a flowchart illustrating an example of the operation of the imaging apparatus 100 according to the first embodiment. This operation is executed every time an input image is captured, for example.
 まず、撮像装置100は、入力画像および基準画像の差分領域を検出する差分領域検出処理を実行する(ステップS910)。そして、撮像装置100は、差分領域において特徴量を取得する特徴量取得処理を実行する(ステップS920)。 First, the imaging apparatus 100 executes a difference area detection process for detecting a difference area between the input image and the reference image (step S910). Then, the imaging apparatus 100 executes a feature amount acquisition process for acquiring a feature amount in the difference area (step S920).
 また、撮像装置100は、物体の推定を行っていない差分領域のいずれかを選択し(ステップS901)、その差分領域の領域類似度を特徴量に基づいて取得する(ステップS902)。撮像装置100は、領域類似度が領域判定閾値より低いか否かを判断する(ステップS903)。 Further, the imaging apparatus 100 selects any one of the difference areas where the object is not estimated (step S901), and acquires the area similarity of the difference area based on the feature amount (step S902). The imaging apparatus 100 determines whether or not the region similarity is lower than the region determination threshold (step S903).
 領域類似度が領域判定閾値より低い場合には(ステップS903:Yes)、撮像装置100は、その領域を不審物等の領域であると判断する(ステップS904)。一方、領域類似度が領域判定閾値以上である場合には(ステップS903:No)、撮像装置100は、その領域は、外乱などの領域であり、不審物等の領域でないと判断する(ステップS905)。 If the area similarity is lower than the area determination threshold (step S903: Yes), the imaging apparatus 100 determines that the area is an area such as a suspicious object (step S904). On the other hand, when the region similarity is equal to or greater than the region determination threshold (step S903: No), the imaging apparatus 100 determines that the region is a region such as a disturbance and is not a region such as a suspicious object (step S905). ).
 ステップS904またはS905の後、撮像装置100は、全ての差分領域で物体の推定を行ったか否かを判断する(ステップS906)。全ての差分領域で物体の推定を行っていない場合には(ステップS906:No)、撮像装置100は、ステップS901に戻る。一方、全ての差分領域で物体の推定を行った場合には(ステップS906:Yes)、撮像装置100は、入力画像に対する処理を終了する(ステップS906)。 After step S904 or S905, the imaging apparatus 100 determines whether the object has been estimated in all the difference areas (step S906). If the object is not estimated in all the difference areas (step S906: No), the imaging apparatus 100 returns to step S901. On the other hand, when the object is estimated in all the difference areas (step S906: Yes), the imaging apparatus 100 ends the process on the input image (step S906).
 図16は、第1の実施の形態における差分領域検出処理の一例を示すフローチャートである。撮像装置100は、入力画像および基準画像の差分画像を生成する(ステップS911)。そして、撮像装置100は、差分画像に対してラべリング処理を行って、ラべリング画像を生成する(ステップS912)。ステップS912の後、撮像装置100は、差分領域検出処理を終了する。 FIG. 16 is a flowchart illustrating an example of a difference area detection process according to the first embodiment. The imaging device 100 generates a difference image between the input image and the reference image (step S911). Then, the imaging apparatus 100 performs a labeling process on the difference image to generate a labeling image (step S912). After step S912, the imaging apparatus 100 ends the difference area detection process.
 図17は、第1の実施の形態における特徴量取得処理の一例を示すフローチャートである。撮像装置100は、入力画像の差分領域において、複数のコーナーを注目画素として検出する(ステップS921)。そして、撮像装置100は、入力画像において、注目画素と、その周囲の周囲画素とからなるグループを抽出し、グループごとに局所特徴量からなる局所特徴ベクトルを取得する(ステップS922)。また、撮像装置100は、入力画像において、注目画像に座標が一致する画素と、その周囲の周囲画素とからなるグループを抽出し、グループごとに局所特徴量からなる局所特徴ベクトルを取得する(ステップS923)。ステップS923の後、撮像装置100は、特徴量取得処理を終了する。 FIG. 17 is a flowchart illustrating an example of a feature amount acquisition process according to the first embodiment. The imaging apparatus 100 detects a plurality of corners as the target pixel in the difference area of the input image (step S921). Then, the imaging apparatus 100 extracts a group composed of the pixel of interest and surrounding pixels around it from the input image, and acquires a local feature vector composed of a local feature amount for each group (step S922). In addition, the imaging apparatus 100 extracts a group including a pixel whose coordinates match the target image and surrounding pixels around the target image in the input image, and acquires a local feature vector including a local feature amount for each group (step). S923). After step S923, the imaging apparatus 100 ends the feature amount acquisition process.
 このように、本技術の第1の実施の形態によれば、注目画素と周囲画素との画素値の差分に応じた特徴量から取得した領域類似度に基づいて物体の領域を推定するため、ノイズによる領域類似度の変動を抑制して、物体の領域を正確に推定することができる。 As described above, according to the first embodiment of the present technology, in order to estimate the region of the object based on the region similarity acquired from the feature amount according to the difference between the pixel values of the target pixel and the surrounding pixels, It is possible to accurately estimate the region of the object while suppressing the variation of the region similarity due to noise.
 <2.第2の実施の形態>
 第1の実施の形態では、撮像装置100は、コーナーを注目画素として検出していたが、ランダムに求めた画素を注目画素として選択してもよい。また、撮像装置100は、注目画素から一定距離内の画素を周囲画素として抽出していたが、注目画素に対して画素値が類似していない画素を周囲画素として抽出してもよい。第2の実施の形態の撮像装置100は、ランダムに求めた画素を注目画素として選択し、注目画素に対して画素値が類似していない画素を周囲画素として抽出する点において第1の実施の形態と異なる。
<2. Second Embodiment>
In the first embodiment, the imaging apparatus 100 detects the corner as the target pixel, but may select a randomly obtained pixel as the target pixel. In addition, the imaging apparatus 100 extracts pixels within a certain distance from the target pixel as surrounding pixels, but may extract pixels whose pixel values are not similar to the target pixel as surrounding pixels. The imaging apparatus 100 according to the second embodiment selects the randomly obtained pixel as a target pixel, and extracts pixels whose pixel values are not similar to the target pixel as surrounding pixels. Different from form.
 図18は、第2の実施の形態における入力画像特徴量取得部230の一構成例を示すブロック図である。第2の実施の形態の入力画像特徴量取得部230は、乱数生成部233、周囲画素抽出部234および局所特徴量取得部235を備える。 FIG. 18 is a block diagram illustrating a configuration example of the input image feature quantity acquisition unit 230 according to the second embodiment. The input image feature amount acquisition unit 230 according to the second embodiment includes a random number generation unit 233, a surrounding pixel extraction unit 234, and a local feature amount acquisition unit 235.
 乱数生成部233は、差分領域内のいずれかの座標に対応する乱数を生成するものである。例えば、差分領域内の画素のx座標の最小値および最大値をxminおよびxmaxとし、y座標の最小値および最大値をyminおよびymaxとする。乱数生成部233は、次の式により乱数xrおよびyrを生成する。
  w=xmax-xmin                         ・・・式1
  h=ymax-ymin                         ・・・式2
  xr=rand(w)                       ・・・式3
  yr=rand(h)                       ・・・式4
式3および式4において、rand(A)は、線形合同法などを使用して、0乃至A-1の乱数を返す関数である。
The random number generation unit 233 generates a random number corresponding to any coordinate in the difference area. For example, the minimum value and the maximum value of the x coordinate of the pixels in the difference area are set to x min and x max, and the minimum value and the maximum value of the y coordinate are set to y min and y max . The random number generation unit 233 generates random numbers x r and y r by the following formula.
w = x max -x min Equation 1
h = y max −y min Expression 2
x r = rand (w) Equation 3
y r = rand (h) Expression 4
In Equations 3 and 4, rand (A) is a function that returns a random number from 0 to A-1 using a linear congruential method or the like.
 乱数生成部233は、(xmin、ymin)を基準とする相対座標(xr、yr)が差分領域内の座標でない場合、または、既に生成した座標である場合には、再度乱数を生成する。一方、相対座標(xr、yr)が差分領域内の新規の座標である場合に、乱数生成部233は、その座標を注目画素の座標として周囲画素抽出部234に供給する。そして、乱数生成部233は、生成した注目画素の座標の個数が一定の個数に達するまで、乱数の生成を繰り返す。 If the relative coordinates (x r , y r ) based on (x min , y min ) are not coordinates in the difference area, or if they are already generated coordinates, the random number generator 233 again generates a random number. Generate. On the other hand, when the relative coordinates (x r , y r ) are new coordinates in the difference area, the random number generation unit 233 supplies the coordinates to the surrounding pixel extraction unit 234 as the coordinates of the target pixel. Then, the random number generation unit 233 repeats generation of random numbers until the number of generated coordinates of the target pixel reaches a certain number.
 周囲画素抽出部234は、注目画素に対して画素値が類似していない周囲画素を抽出するものである。この周囲画素抽出部234は、入力画像において、注目画素の周囲の画素のうち、注目画素に近い画素から順に注目画素に対する画素値の差分を求め、差分の絶対値が所定の差分閾値(例えば、30)より大きいか否かを判断する。そして、周囲画素抽出部234は、差分の絶対値が差分閾値より大きい(言い換えれば、画素値が類似していない)と最初に判断した画素を、注目画素に対応する周囲画素として抽出する。周囲画素抽出部234は、周囲画素を抽出するたびに、周囲画素および周囲画素の画素対のそれぞれの座標を基準画像特徴量取得部240に供給する。また、周囲画素抽出部234は、差分を局所特徴量取得部235に供給する。 The surrounding pixel extraction unit 234 extracts surrounding pixels whose pixel values are not similar to the target pixel. The surrounding pixel extraction unit 234 obtains a pixel value difference with respect to the target pixel in order from a pixel close to the target pixel among pixels around the target pixel in the input image, and the absolute value of the difference is a predetermined difference threshold (for example, 30) It is determined whether or not it is larger. Then, the surrounding pixel extraction unit 234 extracts, as surrounding pixels corresponding to the target pixel, a pixel that is first determined that the absolute value of the difference is larger than the difference threshold (in other words, the pixel values are not similar). Each time the surrounding pixel is extracted, the surrounding pixel extraction unit 234 supplies the coordinates of the surrounding pixel and the pixel pair of the surrounding pixel to the reference image feature amount acquisition unit 240. In addition, the surrounding pixel extraction unit 234 supplies the difference to the local feature amount acquisition unit 235.
 局所特徴量取得部235は、差分に応じた局所特徴量を画素対ごとに取得するものである。例えば、局所特徴量取得部235は、差分の符号が+である場合に「1」の値を、-である場合に「0」の値を局所特徴量として取得する。局所特徴量取得部235は、画素対のそれぞれの局所特徴量を類似度取得部250に供給する。 The local feature amount acquisition unit 235 acquires a local feature amount corresponding to the difference for each pixel pair. For example, the local feature quantity acquisition unit 235 acquires a value of “1” as the local feature quantity when the sign of the difference is + and a value of “0” when the sign of the difference is −. The local feature amount acquisition unit 235 supplies the local feature amount of each pixel pair to the similarity acquisition unit 250.
 なお、入力画像特徴量取得部230は、ランダムに注目画素を選択しているが、第1の実施の形態と同様にコーナーを注目画素として検出してもよい。この場合には、コーナーが検出されるたびにコーナーに対応する周囲画素が抽出され、画素対が生成される。 Note that the input image feature quantity acquisition unit 230 randomly selects a pixel of interest, but may detect a corner as a pixel of interest as in the first embodiment. In this case, every time a corner is detected, surrounding pixels corresponding to the corner are extracted, and a pixel pair is generated.
 図19は、第2の実施の形態における基準画像特徴量取得部240の一構成例を示すブロック図である。この基準画像特徴量取得部240は、差分取得部241および局所特徴量取得部242を備える。差分取得部241、基準画像において注目画素に座標が一致する画素と周囲画素に座標が一致する画素との画素値の差分を取得するものである。差分取得部241は、その差分を局所特徴量取得部242に供給する。局所特徴量取得部242の構成は、局所特徴量取得部235と同様である。 FIG. 19 is a block diagram illustrating a configuration example of the reference image feature amount acquisition unit 240 according to the second embodiment. The reference image feature amount acquisition unit 240 includes a difference acquisition unit 241 and a local feature amount acquisition unit 242. The difference acquisition unit 241 acquires a pixel value difference between a pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixels. The difference acquisition unit 241 supplies the difference to the local feature amount acquisition unit 242. The configuration of the local feature quantity acquisition unit 242 is the same as that of the local feature quantity acquisition unit 235.
 図20は、第2の実施の形態における画素対の抽出方法を説明するための図である。同図において、黒色の画素は、ランダムに選択された注目画素である。また、斜線を引いた画素は、注目画素に対応付けられた周囲画素である。両端が矢印の線分により接続された画素対は、対応する注目画素および周囲画素の画素対である。 FIG. 20 is a diagram for explaining a pixel pair extraction method according to the second embodiment. In the figure, a black pixel is a pixel of interest selected at random. Also, the hatched pixels are surrounding pixels associated with the target pixel. A pixel pair whose both ends are connected by a line segment indicated by an arrow is a pixel pair of a corresponding target pixel and surrounding pixels.
 周囲画素抽出部234は、例えば、旋回するにつれて注目画素711から遠ざかる渦巻き状の経路に沿って順に画素に着目し、着目した画素について注目画素711に対する画素値の差分を求め、差分の絶対値が差分閾値より大きいか否かを判断する。そして、周囲画素抽出部234は、差分の絶対値が差分閾値より大きいと最初に判断した画素を、対応する周囲画素712として抽出する。 The surrounding pixel extraction unit 234, for example, pays attention to the pixels in order along a spiral path that moves away from the target pixel 711 as the vehicle turns, obtains a pixel value difference with respect to the target pixel 711 for the target pixel, and the absolute value of the difference is It is determined whether or not the difference threshold value is greater. Then, the surrounding pixel extraction unit 234 extracts, as the corresponding surrounding pixel 712, a pixel that is first determined that the absolute value of the difference is larger than the difference threshold.
 なお、周囲画素抽出部234は、渦巻き状の経路に沿って順に画素に着目して周囲画素を抽出しているが、第1の実施の形態と同様に、一定距離内の画素を周囲画素として抽出してもよい。この場合には、ランダムに選択された注目画素ごとに、例えば、その周囲の8画素が周囲画素として抽出され、局所特徴量ベクトルが求められる。 Note that the surrounding pixel extraction unit 234 extracts surrounding pixels by paying attention to the pixels in order along the spiral path. However, as in the first embodiment, pixels within a certain distance are used as surrounding pixels. It may be extracted. In this case, for each pixel of interest selected at random, for example, surrounding 8 pixels are extracted as surrounding pixels, and a local feature vector is obtained.
 図21は、第2の実施の形態における基準画像の差分領域の一部を拡大した図である。拡大した部分の画素数は、20×10の200画素である。また、この部分において、画素値「40」の画素は139個であり、画素値「200」の画素は61個であるものとする。 FIG. 21 is an enlarged view of a part of the difference area of the reference image in the second embodiment. The number of pixels in the enlarged part is 200 pixels of 20 × 10. In this portion, it is assumed that there are 139 pixels having a pixel value “40” and 61 pixels having a pixel value “200”.
 図22は、第2の実施の形態における入力画像の一部を拡大した図である。拡大した部分は、図21において入力画像で拡大した部分に対応する部分である。この部分は、雲の影などの一部であり、基準画像に対して、画素値が平均して40前後、低くなっているものとする。この結果、基準画像内の画素値「40」の画素と同じ座標の画素の画素値は、入力画像において「0」前後となり、基準画像内の画素値「200」の画素と同じ座標の画素の画素値は、入力画像において「160」前後となる。 FIG. 22 is an enlarged view of a part of the input image in the second embodiment. The enlarged part is a part corresponding to the part enlarged in the input image in FIG. This part is a part of a cloud shadow or the like, and it is assumed that the average pixel value is about 40 lower than the reference image. As a result, the pixel value of the pixel having the same coordinate as the pixel having the pixel value “40” in the reference image is around “0” in the input image, and the pixel value having the same coordinate as the pixel having the pixel value “200” in the reference image. The pixel value is around “160” in the input image.
 ここで、入力画像において、ノイズ除去部210で完全に除去することができないほどのノイズが生じた場合を想定する。ノイズの影響により、画素値に分散値「5」程度となるばらつきが生じたものとする。この結果、例えば、「0」の画素が95個、画素値「2」の画素が13個、画素値「4」の画素が31個になる。また、画素値「155」の画素が24個、画素値「160」の画素が26個、画素値「162」の画素が11個になる。 Here, it is assumed that noise is generated in the input image that cannot be completely removed by the noise removal unit 210. It is assumed that the pixel value has a variation of about a variance value “5” due to the influence of noise. As a result, for example, there are 95 pixels of “0”, 13 pixels of pixel value “2”, and 31 pixels of pixel value “4”. Also, there are 24 pixels with the pixel value “155”, 26 pixels with the pixel value “160”, and 11 pixels with the pixel value “162”.
 仮に特徴量を求めず、正規化相互相関マッチングにより、入力画像および基準画像の拡大部分の領域類似度を求めた場合、領域類似度Rは次の式により算出される。
Figure JPOXMLDOC01-appb-M000001
上式において、Nは差分領域のx方向の画素数であり、MはY方向の画素数である。I(i,j)は、基準画像の差分領域内の画素の画素値である。T(i,j)は、入力画像の差分領域内の画素の画素値である。
If the feature similarity is not obtained and the region similarity of the enlarged portion of the input image and the reference image is obtained by normalized cross-correlation matching, the region similarity R is calculated by the following equation.
Figure JPOXMLDOC01-appb-M000001
In the above equation, N is the number of pixels in the x direction of the difference area, and M is the number of pixels in the Y direction. I (i, j) is a pixel value of a pixel in the difference area of the reference image. T (i, j) is a pixel value of a pixel in the difference area of the input image.
 式5に画素値を実際に入力すると、次の式が得られる。
Figure JPOXMLDOC01-appb-M000002
When the pixel value is actually input to Equation 5, the following equation is obtained.
Figure JPOXMLDOC01-appb-M000002
 式6に例示するように、領域類似度Rを正規化相互相関マッチングなどにより求めると、ノイズの影響により、その領域類似度Rが低下してしまうおそれがある。 As exemplified in Equation 6, when the region similarity R is obtained by normalized cross-correlation matching or the like, the region similarity R may be lowered due to the influence of noise.
 図23は、第2の実施の形態における局所特徴量の一例を示す図である。同図におけるaは、入力画像の拡大部分の局所特徴量の一例である。同図におけるbは、基準画像の拡大部分の局所特徴量の一例である。同図に例示するように、入力画像および基準画像において画素対ごとの局所特徴量は全て一致する。このため、領域類似度は、最大値「1」となる。このように、ノイズの影響により、領域類似度が低下しにくくなる。 FIG. 23 is a diagram illustrating an example of local feature amounts according to the second embodiment. A in the same figure is an example of the local feature-value of the expansion part of an input image. B in the figure is an example of the local feature amount of the enlarged portion of the reference image. As illustrated in the figure, the local feature values for each pixel pair in the input image and the reference image all match. For this reason, the region similarity is the maximum value “1”. Thus, the region similarity is unlikely to decrease due to the influence of noise.
 図24は、第2の実施の形態における特徴量取得処理の一例を示すフローチャートである。撮像装置100は、入力画像の差分領域において、ランダムに注目画素を選択する(ステップS925)。そして、撮像装置100は、注目画素の周囲において渦巻き状の経路に沿って画素に順に着目して、画素値の差分の絶対値が差分閾値より大きい周囲画素を抽出する(ステップS926)。撮像装置100は、注目画素および周囲画素の画素対の局所特徴量を取得する(ステップS927)。撮像装置100は、画素対の個数が設定値以上であるか否かを判断する(ステップS928)。画素対の個数が設定値未満である場合に(ステップS928:No)撮像装置100は、ステップS925に戻る。 FIG. 24 is a flowchart illustrating an example of a feature amount acquisition process according to the second embodiment. The imaging apparatus 100 selects a target pixel at random in the difference area of the input image (step S925). Then, the imaging apparatus 100 pays attention to the pixels in order along the spiral path around the pixel of interest, and extracts surrounding pixels in which the absolute value of the pixel value difference is greater than the difference threshold (step S926). The imaging apparatus 100 acquires the local feature amount of the pixel pair of the target pixel and the surrounding pixels (step S927). The imaging apparatus 100 determines whether or not the number of pixel pairs is equal to or greater than a set value (step S928). When the number of pixel pairs is less than the set value (step S928: No), the imaging apparatus 100 returns to step S925.
 一方、画素対の個数が設定値以上である場合に(ステップS928:Yes)撮像装置100は、基準画像における局所特徴量を画素対ごとに取得する(ステップS929)。ステップS929の後、撮像装置100は、特徴量取得処理を終了する。 On the other hand, when the number of pixel pairs is equal to or larger than the set value (step S928: Yes), the imaging apparatus 100 acquires a local feature amount in the reference image for each pixel pair (step S929). After step S929, the imaging apparatus 100 ends the feature amount acquisition process.
 このように、第2の実施の形態によれば、撮像装置100は、乱数を生成し、乱数に対応する画素を注目画素として選択するため、コーナーを検出しにくい差分領域(例えば、濃度勾配の小さい領域)において、注目画素を容易に選択することができる。 As described above, according to the second embodiment, since the imaging apparatus 100 generates a random number and selects a pixel corresponding to the random number as a target pixel, a difference region (for example, a density gradient that is difficult to detect a corner) is selected. In a small area, the target pixel can be easily selected.
 <3.第3の実施の形態>
 第1の実施の形態では、物体が移動することを想定していなかったが、物体が移動することを想定して、物体の移動ベクトルをさらに検出してもよい。第3の実施の形態の撮像装置100は、移動ベクトルをさらに検出する点において第1の実施の形態と異なる。
<3. Third Embodiment>
In the first embodiment, it is not assumed that the object moves. However, the movement vector of the object may be further detected on the assumption that the object moves. The imaging apparatus 100 according to the third embodiment is different from the first embodiment in that the movement vector is further detected.
 図25は、第3の実施の形態における差分領域検出部220の一構成例を示すブロック図である。第3の実施の形態の差分領域検出部220は、データバッファ223および移動ベクトル検出部224をさらに備える点において第1の実施の形態と異なる。 FIG. 25 is a block diagram illustrating a configuration example of the difference area detection unit 220 according to the third embodiment. The difference area detection unit 220 of the third embodiment differs from the first embodiment in that it further includes a data buffer 223 and a movement vector detection unit 224.
 データバッファ223は、ラべリング画像、検出ベクトルおよび滞在期間を保持するものである。検出ベクトルおよび滞在期間については後述する。 The data buffer 223 holds a labeling image, a detection vector, and a stay period. The detection vector and the stay period will be described later.
 第3の実施の形態のラべリング処理部222は、ラべリング画像をデータバッファ223および移動ベクトル検出部224に供給する。 The labeling processing unit 222 according to the third embodiment supplies a labeling image to the data buffer 223 and the movement vector detection unit 224.
 移動ベクトル検出部224は、差分領域ごとに、移動ベクトルを検出するものである。この移動ベクトル検出部224は、現在のラべリング画像と、データバッファ223に保持された過去のラべリング画像および移動ベクトルとを取得する。移動ベクトル検出部224は、過去の入力画像における差分領域のいずれかに順に着目し、その差分領域に対応する現在の差分領域を探索する。例えば、移動ベクトル検出部224は、着目した過去の差分領域の基準座標から、現在の差分領域のそれぞれの基準座標までのベクトルを現在の移動ベクトルとして算出する。ここで、基準座標は、例えば、差分領域の中心や重心である。また、移動ベクトル検出部224は、着目した過去の差分領域の面積と、現在の差分領域のそれぞれの面積とを算出する。そして、移動ベクトル検出部224は、面積や移動ベクトルの変動が少ない差分領域を優先して、対応する領域として取得する。例えば、移動ベクトル検出部224は、次の式により現在の差分領域ごとに評価値を求め、評価値が最も高い差分画像を対応する領域として取得する。 The movement vector detection unit 224 detects a movement vector for each difference area. The movement vector detection unit 224 acquires the current labeling image and the past labeling image and movement vector held in the data buffer 223. The movement vector detection unit 224 pays attention to one of the difference areas in the past input image in order, and searches the current difference area corresponding to the difference area. For example, the movement vector detection unit 224 calculates, as the current movement vector, a vector from the reference coordinates of the focused past difference area to the respective reference coordinates of the current difference area. Here, the reference coordinates are, for example, the center and the center of gravity of the difference area. In addition, the movement vector detection unit 224 calculates the area of the past difference area of interest and the area of each of the current difference areas. Then, the movement vector detection unit 224 gives priority to a difference area with a small variation in area and movement vector and acquires it as a corresponding area. For example, the movement vector detection unit 224 obtains an evaluation value for each current difference area by the following formula, and acquires a difference image having the highest evaluation value as a corresponding area.
  Fv(k)=|v_k-v_k’|                 ・・・式7
Figure JPOXMLDOC01-appb-M000003
  F(k)=Fv(k)-Fs(k)               ・・・式9
上式において、v_kは現在のラベルkに対応する差分領域の移動ベクトルである。v_k’は、着目した過去の差分領域の移動ベクトルである。S_kは、現在のラベルkに対応する差分領域の面積である。S_k’は、着目した過去の差分領域の面積である。F(k)は、ラベルkに対応する差分領域の評価値である。なお、最初の入力画像においては、移動ベクトルは初期値に設定される。
Fv (k) = | v_k−v_k ′ |
Figure JPOXMLDOC01-appb-M000003
F (k) = Fv (k) −Fs (k) Equation 9
In the above equation, v_k is a movement vector of the difference area corresponding to the current label k. v_k ′ is a movement vector of the past difference region of interest. S_k is the area of the difference area corresponding to the current label k. S_k ′ is the area of the past difference region of interest. F (k) is an evaluation value of the difference area corresponding to the label k. In the first input image, the movement vector is set to an initial value.
 移動ベクトル検出部224は、現在の差分領域のラベルを、対応する過去の差分領域と同じラベルに更新する。また、移動ベクトル検出部224は、更新したラベルの滞在期間を1フレーム分、カウントアップする。なお、最初の入力画像においては、各ラベルの滞在期間は初期値に設定される。そして、移動ベクトル検出部は、ラベルを更新したラべリング画像をデータバッファ223、入力画像特徴量取得部230および基準画像特徴量取得部240に供給し、移動ベクトルおよび滞在期間をデータバッファ223および制御部140に供給する。このように、過去の差分領域と、現在の差分領域とを、面積や移動ベクトルに基づいて対応付けることにより、一定の速度で移動する移動体を追跡することができる。 The movement vector detection unit 224 updates the label of the current difference area to the same label as the corresponding past difference area. The movement vector detection unit 224 counts up the stay period of the updated label by one frame. In the first input image, the stay period of each label is set to an initial value. Then, the movement vector detection unit supplies the labeling image with the updated label to the data buffer 223, the input image feature amount acquisition unit 230, and the reference image feature amount acquisition unit 240, and the movement vector and the stay period are displayed in the data buffer 223 and This is supplied to the control unit 140. As described above, by associating the past difference area and the current difference area based on the area and the movement vector, it is possible to track the moving body moving at a constant speed.
 図26は、第3の実施の形態における移動ベクトルおよび滞在期間の一構成例を示すブロック図である。同時に例示するように、移動ベクトル検出部224は、ラベルごとに移動ベクトルや滞在期間を求める。 FIG. 26 is a block diagram illustrating a configuration example of the movement vector and the stay period in the third embodiment. As illustrated at the same time, the movement vector detection unit 224 obtains a movement vector and a stay period for each label.
 図27は、第3の実施の形態における差分領域検出処理の一例を示すフローチャートである。第3の実施の形態の差分領域検出処理は、ステップS913およびS914をさらに実行する点において第1の実施の形態と異なる。差分領域検出部220は、ラべリング画像を生成し(ステップS912)、ラベルごとに移動ベクトルを検出する(ステップS913)。また、ラベルごとに滞在期間を求める(ステップS914)。ステップS914の後、差分領域検出部220は、差分領域検出処理を終了する。 FIG. 27 is a flowchart illustrating an example of a difference area detection process according to the third embodiment. The difference area detection process of the third embodiment is different from that of the first embodiment in that steps S913 and S914 are further executed. The difference area detection unit 220 generates a labeling image (step S912) and detects a movement vector for each label (step S913). Further, a stay period is obtained for each label (step S914). After step S914, the difference area detection unit 220 ends the difference area detection process.
 このように、第3の実施の形態によれば、撮像装置100は、差分領域の移動ベクトルを求めるため、移動体を容易に追跡することができる。これにより、撮像装置100は、移動体の滞在期間などを求め、滞在期間の入力画像のみを再生するなど、防犯に有用な処理を行うことができる。 As described above, according to the third embodiment, the imaging apparatus 100 can easily track the moving body in order to obtain the movement vector of the difference area. Thereby, the imaging device 100 can perform a useful process for crime prevention, such as obtaining the stay period of the moving body and reproducing only the input image of the stay period.
 なお、上述の実施の形態は本技術を具現化するための一例を示したものであり、実施の形態における事項と、特許請求の範囲における発明特定事項とはそれぞれ対応関係を有する。同様に、特許請求の範囲における発明特定事項と、これと同一名称を付した本技術の実施の形態における事項とはそれぞれ対応関係を有する。ただし、本技術は実施の形態に限定されるものではなく、その要旨を逸脱しない範囲において実施の形態に種々の変形を施すことにより具現化することができる。 The above-described embodiment shows an example for embodying the present technology, and the matters in the embodiment and the invention-specific matters in the claims have a corresponding relationship. Similarly, the invention specific matter in the claims and the matter in the embodiment of the present technology having the same name as this have a corresponding relationship. However, the present technology is not limited to the embodiment, and can be embodied by making various modifications to the embodiment without departing from the gist thereof.
 また、上述の実施の形態において説明した処理手順は、これら一連の手順を有する方法として捉えてもよく、また、これら一連の手順をコンピュータに実行させるためのプログラム乃至そのプログラムを記憶する記録媒体として捉えてもよい。この記録媒体として、例えば、CD(Compact Disc)、MD(MiniDisc)、DVD(Digital Versatile Disc)、メモリカード、ブルーレイディスク(Blu-ray(登録商標)Disc)等を用いることができる。 Further, the processing procedure described in the above embodiment may be regarded as a method having a series of these procedures, and a program for causing a computer to execute these series of procedures or a recording medium storing the program. You may catch it. As this recording medium, for example, a CD (Compact Disc), an MD (MiniDisc), a DVD (Digital Versatile Disc), a memory card, a Blu-ray disc (Blu-ray (registered trademark) Disc), or the like can be used.
 なお、ここに記載された効果は必ずしも限定されるものではなく、本開示中に記載されたいずれかの効果であってもよい。 It should be noted that the effects described here are not necessarily limited, and may be any of the effects described in the present disclosure.
 なお、本技術は以下のような構成もとることができる。
(1)入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、
 前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、
 前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、
 前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得部と、
 前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定部と
を具備する画像処理装置。
(2)前記入力画像特徴量取得部は、前記比較領域において複数のコーナーを検出して前記注目画素とする
前記(1)記載の画像処理装置。
(3)前記入力画像特徴量取得部は、前記比較領域内の画素のいずれかに対応する乱数を生成して当該乱数に対応する前記画素を前記注目画素とする
前記(1)に記載の画像処理装置。
(4)前記入力画像特徴量取得部は、前記注目画素から所定距離内の画素を抽出して前記周囲画素とする
前記(1)から(3)のいずれかに記載の画像処理装置。
(5)前記入力画像特徴量取得部は、前記比較領域内の画素の中から前記注目画素に対して画素値が類似しない画素を抽出して前記周囲画素とする
前記(1)から(2)のいずれかに記載の画像処理装置。
(6)前記物体推定部は、前記複数の注目画素の各々について取得された前記入力画像特徴量と前記注目画素に座標が一致する前記対応画素の各々について取得された前記基準画像特徴量との類似度である局所類似度が所定の局所判定閾値より高いか否かを前記注目画素ごとに判定して前記局所類似度が前記局所判定閾値より高いと判定した回数に応じた値を前記領域類似度として取得する
前記(1)から(5)のいずれかに記載の画像処理装置。
(7)前記入力画像および前記基準画像において画素値が類似しない画素を検出して当該検出した画素からなる領域を前記比較領域とする
前記(1)から(6)のいずれかに記載の画像処理装置。
(8)前記比較領域決定部は、2つの前記入力画像のそれぞれと前記基準画像において前記比較領域を決定して前記2つの入力画像の一方の前記比較領域から他方の前記比較領域へのベクトルを移動ベクトルとして検出する
前記(1)から(7)のいずれかに記載の画像処理装置。
(9)入力画像を撮像する撮像部と、
 入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、
 前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、
 前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、
 前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得部と、
 前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定部と
を具備する撮像装置。
(10)比較領域決定部が、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定手順と、
 入力画像特徴量取得部が、前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得手順と、
 基準画像特徴量取得部が、前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得手順と、
 類似度取得部が、前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得手順と、
 物体推定部が、前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定手順と
を具備する画像処理方法。
(11)比較領域決定部が、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定手順と、
 入力画像特徴量取得部が、前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得手順と、
 基準画像特徴量取得部が、前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得手順と、
 類似度取得部が、前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得手順と、
 物体推定部が、前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定手順と
をコンピュータに実行させるためのプログラム。
In addition, this technique can also take the following structures.
(1) a comparison area determination unit that determines a comparison area to be compared in each of the input image and the predetermined reference image;
An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image An acquisition unit;
A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
A similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
An image processing apparatus comprising: an object estimation unit configured to estimate the comparison region that is not similar based on the region similarity in the input image as an object region.
(2) The image processing apparatus according to (1), wherein the input image feature amount acquisition unit detects a plurality of corners in the comparison region and sets the target pixel.
(3) The image according to (1), wherein the input image feature amount acquisition unit generates a random number corresponding to any one of the pixels in the comparison area and sets the pixel corresponding to the random number as the pixel of interest. Processing equipment.
(4) The image processing device according to any one of (1) to (3), wherein the input image feature quantity acquisition unit extracts a pixel within a predetermined distance from the target pixel as the surrounding pixel.
(5) The input image feature quantity acquisition unit extracts a pixel whose pixel value is not similar to the target pixel from the pixels in the comparison area, and sets the pixel as the surrounding pixel. An image processing apparatus according to any one of the above.
(6) The object estimation unit may calculate the input image feature amount acquired for each of the plurality of target pixels and the reference image feature amount acquired for each of the corresponding pixels whose coordinates coincide with the target pixel. Whether the local similarity that is the similarity is higher than a predetermined local determination threshold value is determined for each pixel of interest, and a value corresponding to the number of times that the local similarity is determined to be higher than the local determination threshold value is the region similarity The image processing apparatus according to any one of (1) to (5), acquired as a degree.
(7) The image processing according to any one of (1) to (6), wherein a pixel whose pixel value is not similar in the input image and the reference image is detected and an area including the detected pixel is used as the comparison area. apparatus.
(8) The comparison area determination unit determines the comparison area in each of the two input images and the reference image, and calculates a vector from one comparison area to the other comparison area of the two input images. The image processing device according to any one of (1) to (7), wherein the image processing device is detected as a movement vector.
(9) an imaging unit that captures an input image;
A comparison area determination unit for determining a comparison area to be compared in each of the input image and the predetermined reference image;
An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image An acquisition unit;
A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
A similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
An imaging apparatus comprising: an object estimation unit configured to estimate the comparison area that is not similar based on the area similarity in the input image as an object area.
(10) a comparison region determination procedure in which the comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
The input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image. Input image feature amount acquisition procedure to be acquired as a quantity;
A reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image. A reference image feature acquisition procedure;
Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
An object processing method, comprising: an object estimation unit configured to estimate, as an object region, the comparison region that is not similar based on the region similarity in the input image.
(11) a comparison region determination procedure in which the comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
The input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image. Input image feature amount acquisition procedure to be acquired as a quantity;
A reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image. A reference image feature acquisition procedure;
Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
A program for causing an object estimation unit to cause a computer to execute an object estimation procedure for estimating a comparison area that is not similar based on the area similarity in the input image as an object area.
 100 撮像装置
 110 撮像レンズ
 120 撮像素子
 130 記録部
 140 制御部
 200 画像処理部
 210 ノイズ除去部
 220 差分領域検出部
 221 差分画像生成部
 222 ラべリング処理部
 223 データバッファ
 224 移動ベクトル検出部
 230 入力画像特徴量取得部
 231 コーナー検出部
 232、235、242 局所特徴量取得部
 233 乱数生成部
 234 周囲画素抽出部
 240 基準画像特徴量取得部
 241 差分取得部
 250 類似度取得部
 260 物体推定部
DESCRIPTION OF SYMBOLS 100 Image pick-up device 110 Imaging lens 120 Image pick-up element 130 Recording part 140 Control part 200 Image processing part 210 Noise removal part 220 Difference area detection part 221 Difference image generation part 222 Labeling process part 223 Data buffer 224 Movement vector detection part 230 Input image Feature amount acquisition unit 231 Corner detection unit 232, 235, 242 Local feature amount acquisition unit 233 Random number generation unit 234 Surrounding pixel extraction unit 240 Reference image feature amount acquisition unit 241 Difference acquisition unit 250 Similarity acquisition unit 260 Object estimation unit

Claims (11)

  1.  入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、
     前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、
     前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、
     前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得部と、
     前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定部と
    を具備する画像処理装置。
    A comparison area determination unit for determining a comparison area to be compared in each of the input image and the predetermined reference image;
    An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image An acquisition unit;
    A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
    A similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
    An image processing apparatus comprising: an object estimation unit configured to estimate the comparison region that is not similar based on the region similarity in the input image as an object region.
  2.  前記入力画像特徴量取得部は、前記比較領域において複数のコーナーを検出して前記注目画素とする
    請求項1記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the input image feature amount acquisition unit detects a plurality of corners in the comparison region as the target pixel.
  3.  前記入力画像特徴量取得部は、前記比較領域内の画素のいずれかに対応する乱数を生成して当該乱数に対応する前記画素を前記注目画素とする
    請求項1記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the input image feature amount acquisition unit generates a random number corresponding to any one of the pixels in the comparison region and sets the pixel corresponding to the random number as the target pixel.
  4.  前記入力画像特徴量取得部は、前記注目画素から所定距離内の画素を抽出して前記周囲画素とする
    請求項1記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the input image feature amount acquisition unit extracts pixels within a predetermined distance from the target pixel as the surrounding pixels.
  5.  前記入力画像特徴量取得部は、前記比較領域内の画素の中から前記注目画素に対して画素値が類似しない画素を抽出して前記周囲画素とする
    請求項1記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the input image feature amount acquisition unit extracts a pixel whose pixel value is not similar to the target pixel from the pixels in the comparison region, and sets the pixel as the surrounding pixel.
  6.  前記物体推定部は、前記複数の注目画素の各々について取得された前記入力画像特徴量と前記注目画素に座標が一致する前記対応画素の各々について取得された前記基準画像特徴量との類似度である局所類似度が所定の局所判定閾値より高いか否かを前記注目画素ごとに判定して前記局所類似度が前記局所判定閾値より高いと判定した回数に応じた値を前記領域類似度として取得する
    請求項1記載の画像処理装置。
    The object estimation unit is based on a similarity between the input image feature amount acquired for each of the plurality of target pixels and the reference image feature amount acquired for each of the corresponding pixels whose coordinates coincide with the target pixel. Whether or not a certain local similarity is higher than a predetermined local determination threshold is determined for each pixel of interest, and a value corresponding to the number of times the local similarity is determined to be higher than the local determination threshold is acquired as the region similarity The image processing apparatus according to claim 1.
  7.  前記比較領域決定部は、前記入力画像および前記基準画像において画素値が類似しない画素を検出して当該検出した画素からなる領域を前記比較領域とする
    請求項1記載の画像処理装置。
    The image processing apparatus according to claim 1, wherein the comparison area determination unit detects pixels whose pixel values are not similar in the input image and the reference image, and sets an area including the detected pixels as the comparison area.
  8.  前記比較領域決定部は、2つの前記入力画像のそれぞれと前記基準画像において前記比較領域を決定して前記2つの入力画像の一方の前記比較領域から他方の前記比較領域へのベクトルを移動ベクトルとして検出する
    請求項1記載の画像処理装置。
    The comparison area determination unit determines the comparison area in each of the two input images and the reference image, and uses a vector from one comparison area to the other comparison area of the two input images as a movement vector. The image processing apparatus according to claim 1 for detection.
  9.  入力画像を撮像する撮像部と、
     入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定部と、
     前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得部と、
     前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得部と、
     前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得部と、
     前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定部と
    を具備する撮像装置。
    An imaging unit that captures an input image;
    A comparison area determination unit for determining a comparison area to be compared in each of the input image and the predetermined reference image;
    An input image feature value that acquires, as an input image feature value, a value corresponding to a pixel value difference between surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image An acquisition unit;
    A reference image feature amount acquisition unit that acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel in the reference image and a pixel whose coordinates match the surrounding pixel;
    A similarity acquisition unit that acquires a similarity between the comparison region in the input image and the comparison region in the reference image as a region similarity based on the input image feature and the reference image feature;
    An imaging apparatus comprising: an object estimation unit configured to estimate the comparison area that is not similar based on the area similarity in the input image as an object area.
  10.  比較領域決定部が、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定手順と、
     入力画像特徴量取得部が、前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得手順と、
     基準画像特徴量取得部が、前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得手順と、
     類似度取得部が、前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得手順と、
     物体推定部が、前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定手順と
    を具備する画像処理方法。
    A comparison region determination procedure in which a comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
    The input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image. Input image feature amount acquisition procedure to be acquired as a quantity;
    A reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image. A reference image feature acquisition procedure;
    Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
    An object processing method, comprising: an object estimation unit configured to estimate, as an object region, the comparison region that is not similar based on the region similarity in the input image.
  11.  比較領域決定部が、入力画像と所定の基準画像とのそれぞれにおいて比較すべき比較領域を決定する比較領域決定手順と、
     入力画像特徴量取得部が、前記入力画像において前記比較領域内の複数の注目画素のそれぞれについて前記注目画素の周囲の周囲画素と前記注目画素との画素値の差分に応じた値を入力画像特徴量として取得する入力画像特徴量取得手順と、
     基準画像特徴量取得部が、前記基準画像において前記注目画素に座標が一致する対応画素と前記周囲画素に座標が一致する画素との画素値の差分に応じた値を基準画像特徴量として取得する基準画像特徴量取得手順と、
     類似度取得部が、前記入力画像内の前記比較領域と前記基準画像内の前記比較領域との類似度を領域類似度として前記入力画像特徴量および前記基準画像特徴量に基づいて取得する類似度取得手順と、
     物体推定部が、前記入力画像において前記領域類似度に基づいて類似していない前記比較領域を物体の領域として推定する物体推定手順と
    をコンピュータに実行させるためのプログラム。
    A comparison region determination procedure in which a comparison region determination unit determines a comparison region to be compared in each of the input image and the predetermined reference image;
    The input image feature amount acquisition unit inputs a value corresponding to a difference between pixel values of surrounding pixels around the target pixel and the target pixel for each of the plurality of target pixels in the comparison region in the input image. Input image feature amount acquisition procedure to be acquired as a quantity;
    A reference image feature amount acquisition unit acquires, as a reference image feature amount, a value corresponding to a difference in pixel value between a corresponding pixel whose coordinates match the target pixel and a pixel whose coordinates match the surrounding pixel in the reference image. A reference image feature acquisition procedure;
    Similarity acquired by the similarity acquisition unit based on the input image feature quantity and the reference image feature quantity as the similarity between the comparison area in the input image and the comparison area in the reference image Acquisition procedure;
    A program for causing an object estimation unit to cause a computer to execute an object estimation procedure for estimating a comparison area that is not similar based on the area similarity in the input image as an object area.
PCT/JP2015/054971 2014-04-16 2015-02-23 Image-processing device, imaging device, image-processing method, and program WO2015159585A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2014084182 2014-04-16
JP2014-084182 2014-04-16

Publications (1)

Publication Number Publication Date
WO2015159585A1 true WO2015159585A1 (en) 2015-10-22

Family

ID=54323803

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/054971 WO2015159585A1 (en) 2014-04-16 2015-02-23 Image-processing device, imaging device, image-processing method, and program

Country Status (1)

Country Link
WO (1) WO2015159585A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112801891A (en) * 2021-01-11 2021-05-14 Tcl华星光电技术有限公司 Display screen picture detection method and display screen picture detection system
CN113516145A (en) * 2020-12-01 2021-10-19 阿里巴巴集团控股有限公司 Image processing and vehicle information providing method, apparatus and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003271972A (en) * 2002-03-14 2003-09-26 Mitsubishi Electric Corp Image processor, method for processing image, and program
JP2003346156A (en) * 2002-05-23 2003-12-05 Nippon Telegr & Teleph Corp <Ntt> Object detection device, object detection method, program, and recording medium
JP2005293033A (en) * 2004-03-31 2005-10-20 Secom Co Ltd Image processor and intruder detection device
JP2006338187A (en) * 2005-05-31 2006-12-14 Secom Co Ltd Monitoring device
JP2007180933A (en) * 2005-12-28 2007-07-12 Secom Co Ltd Image sensor
JP2012238175A (en) * 2011-05-11 2012-12-06 Canon Inc Information processing device, information processing method, and program

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003271972A (en) * 2002-03-14 2003-09-26 Mitsubishi Electric Corp Image processor, method for processing image, and program
JP2003346156A (en) * 2002-05-23 2003-12-05 Nippon Telegr & Teleph Corp <Ntt> Object detection device, object detection method, program, and recording medium
JP2005293033A (en) * 2004-03-31 2005-10-20 Secom Co Ltd Image processor and intruder detection device
JP2006338187A (en) * 2005-05-31 2006-12-14 Secom Co Ltd Monitoring device
JP2007180933A (en) * 2005-12-28 2007-07-12 Secom Co Ltd Image sensor
JP2012238175A (en) * 2011-05-11 2012-12-06 Canon Inc Information processing device, information processing method, and program

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516145A (en) * 2020-12-01 2021-10-19 阿里巴巴集团控股有限公司 Image processing and vehicle information providing method, apparatus and storage medium
CN113516145B (en) * 2020-12-01 2024-02-27 阿里巴巴集团控股有限公司 Image processing and vehicle information providing method, apparatus and storage medium
CN112801891A (en) * 2021-01-11 2021-05-14 Tcl华星光电技术有限公司 Display screen picture detection method and display screen picture detection system
CN112801891B (en) * 2021-01-11 2023-10-03 Tcl华星光电技术有限公司 Display screen picture detection method and display screen picture detection system

Similar Documents

Publication Publication Date Title
Mosleh et al. Automatic inpainting scheme for video text detection and removal
CN110008795B (en) Image target tracking method and system and computer readable recording medium
JP5457606B2 (en) Image processing method and apparatus
JP5821457B2 (en) Image processing apparatus, image processing apparatus control method, and program for causing computer to execute the method
CN109727275B (en) Object detection method, device, system and computer readable storage medium
JP4533836B2 (en) Fluctuating region detection apparatus and method
JP6075190B2 (en) Image processing method and apparatus
JP5779089B2 (en) Edge detection apparatus, edge detection program, and edge detection method
JP2013012820A (en) Image processing apparatus, image processing apparatus control method, and program for causing computer to execute the method
JP6924064B2 (en) Image processing device and its control method, and image pickup device
JP2012038318A (en) Target detection method and device
Kurnianggoro et al. Dense optical flow in stabilized scenes for moving object detection from a moving camera
JP2016505185A (en) Image processor having edge selection function
US20120121189A1 (en) Image processing device, image processing method, program, and integrated circuit
JP6720845B2 (en) Image processing apparatus, image processing method and program
WO2015159585A1 (en) Image-processing device, imaging device, image-processing method, and program
JP6221283B2 (en) Image processing apparatus, image processing method, and image processing program
JP2018133042A (en) Left object detector
JP2021111228A (en) Learning device, learning method, and program
KR101829386B1 (en) Apparatus and method for detecting target
JP5539565B2 (en) Imaging apparatus and subject tracking method
CN112991419B (en) Parallax data generation method, parallax data generation device, computer equipment and storage medium
JP2009076094A (en) Moving object monitoring device
CN111275045B (en) Image main body recognition method and device, electronic equipment and medium
KR101284200B1 (en) Video processing apparatus and method for detecting smoke from video

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15779589

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15779589

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP