WO2012161291A1 - Site separation location extraction device, program, and method - Google Patents

Site separation location extraction device, program, and method Download PDF

Info

Publication number
WO2012161291A1
WO2012161291A1 PCT/JP2012/063403 JP2012063403W WO2012161291A1 WO 2012161291 A1 WO2012161291 A1 WO 2012161291A1 JP 2012063403 W JP2012063403 W JP 2012063403W WO 2012161291 A1 WO2012161291 A1 WO 2012161291A1
Authority
WO
WIPO (PCT)
Prior art keywords
person
change
separation position
part separation
amount
Prior art date
Application number
PCT/JP2012/063403
Other languages
French (fr)
Japanese (ja)
Inventor
康史 平川
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Priority to JP2013516448A priority Critical patent/JP6090162B2/en
Publication of WO2012161291A1 publication Critical patent/WO2012161291A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition

Definitions

  • the present invention relates to a part separation position extraction device, a program, and a method.
  • a monitor or the like can use information such as the color, shape, and type of clothes worn by a person in order to discover a suspicious person from a camera installed for monitoring purposes and express the characteristics of the person. Monitors can also use information such as the color, shape, and type of clothes worn by a person to collect marketing information to understand the trends of customers coming to clothing stores. .
  • Patent Document 1 discloses an example of a method for searching for an image showing a person wearing a specific clothes.
  • Patent Document 2 discloses an example of an image search method for searching for an image similar to an input image.
  • an object of the present invention is to provide a part separation position extraction apparatus, program, and method capable of separating and extracting a region such as a person or an object on an image.
  • Feature change model accumulating means for storing, as a model, a change amount of a calculated feature amount value with respect to a change in a position on a vertical axis with respect to the predetermined axis, and the change amount for a person region of an external input image.
  • a feature amount calculated for each line segment where a plurality of axes parallel to a predetermined axis on the image and a person region included in the image intersect The change amount is calculated for a human region of the input image from the outside in a computer having a feature change model accumulating unit that stores a change amount of the value of the value with respect to a change in position on the vertical axis with respect to the predetermined axis as a model.
  • the change amount is compared with the change amount stored in the feature change model accumulating means to extract a model having a similar change amount, and the value of the change amount of the extracted model is not less than a predetermined value.
  • a program storage medium for storing a part separation position extraction program for executing a part separation position extraction processing step for outputting a peak position as a part. Further, according to the present invention, for each person region on the image, a feature amount calculated for each line segment where a plurality of axes parallel to a predetermined axis on the image and a person region included in the image intersect Is stored as a model with respect to a change in position on the vertical axis with respect to the predetermined axis, the change amount is calculated for a human region of an input image from the outside, and the change amount and the feature change are calculated.
  • a separation position extraction method is provided.
  • Each unit constituting the device of each embodiment includes a control unit, a memory, a program loaded in the memory, a storage unit such as a hard disk for storing the program, a network connection interface, and the like. Realized by a combination of And unless there is particular notice, the realization method and apparatus are not limited.
  • the control unit includes a CPU (Central Processing Unit) and the like, and controls the entire apparatus by operating an operating system, and reads programs and data from a recording medium mounted on a drive device to a memory, for example.
  • Various processes are executed according to the above.
  • the recording medium is, for example, an optical disk, a flexible disk, a magnetic optical disk, an external hard disk, a semiconductor memory, or the like, and records a computer program so that the computer can read it.
  • the computer program may be downloaded from an external computer (not shown) connected to the communication network.
  • the communication network may be the Internet, a LAN (Local Area Network), a public line network, a wireless communication network, or a network configured by a combination thereof.
  • Each unit constituting the device of each embodiment is configured by a combination of hardware such as a logic circuit, software, and the like.
  • the block diagram used in the description of each embodiment shows a functional unit configuration, not a hardware unit configuration. These functional blocks are realized by any combination of hardware and software.
  • each embodiment may be described as being realized by one physically coupled device, but the means for realizing the same is not limited thereto. That is, the components, devices, systems, and the like described in each embodiment may be realized by a plurality of devices by connecting two or more physically separated devices in a wired or wireless manner.
  • the means for realizing it is not limited to this. That is, each component, device, system, and the like of each embodiment may be realized by arbitrarily combining hardware and software so as to be realized by one physically coupled device.
  • the part separation position extraction apparatus 1 includes a part separation position output processing unit 2.
  • the part separation position output processing unit 2 includes a part separation position extraction processing unit 101 and a feature change model storage unit 102, and the feature change model storage unit 102 is connected to the part separation position extraction processing unit 101.
  • the feature change model accumulating unit 102 has an image feature in a predetermined axial direction defined on an image showing a person, for example, a vertical direction of the person (an axial direction connecting a head and a foot).
  • the part separation position output processing unit 2 extracts a person area indicating a person area from image information (also simply referred to as an image) input from the outside using an existing method, and determines the size of the person area in the vertical direction.
  • the size of the person area is normalized by enlarging or reducing to a predetermined size.
  • the part separation position output processing unit 2 scans the human area in the horizontal direction, that is, pixel values on a line segment where the axis parallel to the horizontal axis intersects the human area (for example, the type of pixel color)
  • Image feature quantity indicating the feature of the clothing at each position on the axis in the vertical direction of the person is calculated on the basis of the numerical value representing the intensity, brightness, brightness, etc. (see Equation 1 described later).
  • the vertical direction is an axial direction connecting the head and feet of a person in the image
  • the part separation position output processing unit 2 may calculate from the image using a known method, or the vertical direction of the camera in advance.
  • the camera is set so that the direction coincides with the vertical direction
  • the part separation position output unit 2 may set the vertical direction of the image as the vertical direction.
  • the part separation position output processing unit 2 calculates a feature change amount indicating a change in the image feature amount that occurs in the vertical direction of the person region, and associates the feature change amount with the type of clothes worn by the person. Information on clothes worn by persons on the image is given in advance.
  • the part separation position output processing unit 2 obtains each feature change amount for each clothing (see FIG. 12).
  • the part separation position output processing unit 2 uses the peak position of each feature change amount as the part separation position, and uses the information associating the part separation position and the obtained change amount as the above feature change model. Store in the model storage unit 102.
  • the feature change amounts of each classification may be similar. For example, there is a high possibility that the feature change amount obtained from clothes such as sweaters and jeans is similar to the feature change amount obtained from clothes such as Y-shirts and slacks. Therefore, the part separation position output processing unit 2 may combine similar feature change amounts into one. For example, the part separation position output processing unit 2 may combine feature change amounts using an average of each feature change amount when combining similar feature change amounts into one feature change amount.
  • one representative feature change amount may be selected from similar feature change amounts based on a predetermined criterion.
  • the part separation position output processing unit 2 may use another method for calculating a feature change amount that is representative of a plurality of feature change amounts.
  • the part separation position output processing unit 2 uses the degree of similarity calculated using the number of part separation positions and the difference distance between the part separation positions when collecting clothes having similar part separation positions, or DP (Dynamic Programming) matching.
  • the degree of similarity calculated using a method such as the above can be used, but other methods for calculating the degree of similarity may be used.
  • the part separation position output processing unit 2 uses the pixel value obtained by scanning the image included in the person region in the horizontal direction as a method for calculating the image feature amount indicating the feature of the person's clothes. The method may be used. That is, the part separation position output processing unit 2 extracts a pixel value of a certain pixel included in the person area, and calculates an image feature amount based on the pixel value and the pixel values of pixels around the pixel. Then, a new image feature amount may be calculated by scanning the image feature amount in the horizontal direction.
  • the part separation position output processing unit 2 considers a plurality of pixels as one pixel instead of one pixel, and scans an average value of pixel values of the plurality of pixels in the horizontal direction to obtain an image feature amount. It may be calculated.
  • the part separation position output processing unit 2 assigns a clothing name and a body part name (part generic name information) to the section (part separation section) divided by the part separation position, and is characterized as a feature change model.
  • the change model storage unit 102 may store the change model. For example, as shown in FIG.
  • the part separation position output processing unit 2 refers to, for example, a storage unit (not shown) in which the section length and the part generic information are stored in association with each other, and each part separation section includes a face, a T-shirt, a skirt, You may give the part generic name information called a leg.
  • the part separation position output processing unit 2 may further add the name of part generic information that summarizes clothes having similar lengths when calculating the feature change amount included in the feature change model.
  • the part separation position output processing unit 2 may use T-shirts, jerseys, blousons, knits and the like as “short upper body clothes (upper body 1)”, and long coats, rain feathers, etc. as “long upper body”. Clothing (upper body 2) ”, mini skirts, shorts and half pants“ short lower body clothing (lower body 1) ”, lower body clothing such as seven-part trousers and long skirts that can be seen near the ankle, etc.
  • the part separation position output processing unit 2 may add not only the part separation section but also part generic name information about the entire clothing to the feature change model. For example, a combination of a jacket and slacks may be given part generic information “suits”. By the way, depending on clothes, the gender that can wear the clothes may be determined. Therefore, the part separation position output processing unit 2 may add information (partial sex information) such as male, female, and unisex that is determined for each clothing to the feature change model.
  • the part separation position output processing unit 2 may also store this part sex information in the feature change model storage unit 102.
  • the part separation position output processing unit 2 refers to, for example, a storage unit (not shown) in which part generic information and part sex information are stored in association with each other. It is possible to give a woman as part sex information to the skirt.
  • the part separation position output processing unit 2 may further add attribute information such as a person's sex, age, height, race, hairstyle (such as length) to the feature change model together with the part sex information.
  • the attribute information includes “male / female” for gender, “length” and “style” for hairstyle, “age group” for age, and “skin color” for race.
  • the part separation position output processing unit 2 can extract the attribute information from a person shown in the image using a known technique.
  • the part separation position output processing unit 2 may also store these attribute information in the feature change model storage unit 102.
  • the part separation position output processing unit 2 can add to the feature change model if there is other attribute information and can store it in the feature change model storage unit 102.
  • the accumulated part sex information, part generic name information, and attribute information can be used for narrowing down feature change models extracted from the feature change model storage unit 102 when matching feature change models to be described later.
  • the part sex information, the part generic name information, and the attribute information are collectively referred to as a person attribute.
  • the part sex information, part generic information, attribute information, part separation section length (part separation section length), ID (IDentification) for indicating a correspondence relationship between feature change amounts, and the like are, for example, an association list as shown in FIG.
  • Information may be stored in the feature change model storage unit 102 as information.
  • the saved format may be another data format.
  • the feature change model storage unit 102 may store the feature change model for each direction of the person.
  • the part separation position output processing unit 2 detects, for example, that a person is facing in the horizontal direction or the rear direction by a known technique, calculates a feature change amount and separation position information for each person's direction, and models each direction.
  • the feature change model accumulating unit 102 stores feature change models constructed for the front direction, the horizontal direction, and the rear direction for suits that look different from the front direction, the horizontal direction, and the rear direction. You may accumulate according to direction.
  • the above-described process has been described as the process in which the part separation position output processing unit 2 stores the feature change model in the feature change model storage unit 102, but is not limited thereto.
  • a feature change model generated in advance by an external device (not shown) or the like may be stored in the feature change model storage unit 102.
  • the feature change model storage unit 102 has the feature change amount of the image feature amount or the feature change amount and the part separation position at each position on the vertical axis of the person as shown in FIG.
  • the associated feature change model is accumulated.
  • the feature change model storage unit 102 may further store person attributes (partial sex information, part generic name information, attribute information) and part separation section length. Further, the feature change model storage unit 102 may store a probability distribution indicating the frequency of occurrence of a location where the feature change amount of the image feature amount on the vertical axis of the person is equal to or greater than a threshold value.
  • the part separation position extraction processing unit 101 determines whether the image information, the person area information (information indicating the position of the area in which the person is shown in the image information), and the clothes of the person area can be separated. Using the part separation possibility determination information to be determined, the feature change amount of the person region is calculated as shown in the right graph of FIG. 12, and the feature change amount and the feature change accumulated in the feature change model storage unit 102 are calculated. The similarity with the feature change amount of the model is calculated.
  • the part separation position extraction processing unit 101 can calculate the degree of similarity using a method such as DP (Dynamic Programming) matching, using the number of part separation positions and the difference distance between the part separation positions. Other similarity calculation methods may be used.
  • the part separation possibility determination information is information indicating whether or not a person can be separated according to how the person in the image is captured.
  • a value that determines whether separation is possible is described.
  • the part separation possibility determination information is a binary value indicating whether or not it is possible, or a numerical value for performing threshold processing.
  • the part separation availability information is calculated by an external device (not shown).
  • the external device can determine how to capture a person showing that the person area is shown in the whole body in the image or the lower body is partially hidden, and can determine whether or not the part can be separated based on the way the image is taken.
  • a specific example of the part separation availability information will be described with reference to FIG. The left figure of FIG.
  • the external device can calculate a region as shown in the right diagram of FIG. 11 by a known method based on, for example, the position of the person's feet.
  • the upper end of the desk is an area where only the upper body is shown (black area)
  • the floor is an area where the whole body is shown (shaded area)
  • the wall is an area where a person's feet cannot exist (white area) ).
  • the part separation possibility determination information is “Yes” if there is a region in the image where there is a possibility of part separation, and “No” if there is no region.
  • the part separation position extraction processing unit 101 outputs the part separation position of the input person region based on the part separation position of the feature change model having the highest feature change amount. At this time, the part separation position extraction processing unit 101 can also receive and output information such as part generic name information, part sex information, attribute information, and part separation section length from the feature change model storage unit 102. For example, the part separation position extraction processing unit 101, when the feature change amount of clothes worn by a person in the input image is similar to a plurality of feature change amounts stored in the feature change model storage unit 102, Site generic information can be output.
  • the part separation position extraction processing unit 101 determines whether or not part separation is possible from the person region on the image based on the part separation possibility determination information (step 1001).
  • the part separation position extraction processing unit 101 determines whether or not separation is possible based on the part separation possibility determination information.
  • the part separation position extraction processing unit 101 normalizes the person region (step 1002).
  • the reason for performing the normalization is to suppress the variation of the feature change amount due to the change in the person area size.
  • the part separation position extraction processing unit 101 performs an enlargement or reduction process of an image while maintaining a vertical / horizontal aspect ratio so that the lengths of the human regions in the vertical direction are aligned with the reference value.
  • the enlargement / reduction processing method may be a general image enlargement / reduction algorithm, such as the nearest neighbor method, the bilinear method, the bicubic method, or the like.
  • the part separation position extraction processing unit 101 calculates an image feature amount in the horizontal direction of the person region.
  • the part separation position extraction processing unit 101 scans pixel values of an image in the range indicated by the human region information in the horizontal direction (x-axis direction) and uses a function such as a pixel value projected in the vertical direction (y-axis direction). It is obtained and used as an image feature amount (step 1003). Specifically, the part separation position extraction processing unit 101 obtains an image feature amount on a line segment where the person region and the x axis intersect, for example, as in (Equation 1).
  • I (x, y) is a pixel value at coordinates (x, y) on the input image (for example, a value calculated from a three-dimensional vector of R (Red), G (Green), and B (Blue)).
  • the part separation position extraction processing unit 101 calculates the change amount (feature change amount) from the image feature amount. That is, the part separation position extraction processing unit 101 calculates the feature change amount of the image feature amount projected in the vertical direction (step 1004). For example, the part separation position extraction processing unit 101 determines the position of the image feature value calculated for each line segment where a plurality of axes parallel to the horizontal axis intersect with the person area on the vertical axis. A feature change amount with respect to the change is calculated.
  • the feature change amount is obtained as in (Equation 2), for example.
  • B is an upper limit of the range of the image feature amount used for obtaining the feature change amount.
  • the part separation position extraction processing unit 101 calculates the similarity between the calculated feature change amount and each feature change amount included in the feature change model stored in the feature change model storage unit 102 in advance.
  • a feature change model having a feature change amount is extracted, and a peak position (part separation position) of the feature change amount of the model is output (step 1005).
  • the part separation position extraction processing unit 101 outputs the part separation position of the feature change amount having the highest similarity.
  • the part separation position extraction processing unit 101 extracts, for example, a value in which D (y) in (Equation 2) is a peak equal to or greater than a threshold as a method for calculating the similarity, and the position of the peak is determined as the part separation position. Let it be a candidate (part separation candidate position). The part separation position extraction processing unit 101 then determines the number of part separation positions and the difference distance between the part separation positions based on the part separation positions accumulated in the feature change model accumulation unit 102 and the extracted part separation candidate positions. Or the similarity of the feature change amount can be calculated by using a technique such as DP matching.
  • the part separation position extraction processing unit 101 obtains a product of the feature change amount of (Equation 2) and the feature change amount of the feature change model stored in the feature change model storage unit 102, and uses the product distribution.
  • the number of peaks that are greater than or equal to the threshold value and the height of the peak may be detected, and the similarity of the feature change amount may be determined based on the number of peaks and the peak height.
  • the part separation position extraction processing unit 101 may determine the similarity of the feature change amount using other methods. According to the present embodiment, by outputting the peak position of a feature change model in which the change amount of the image feature quantity is similar to the clothes worn by the person on the input image, the parts of the clothes are separated. And can be extracted.
  • the feature change amount and its peak position extracted from the input image may not always accurately indicate the separation position of the part due to image noise or the like.
  • the part separation position extraction processing unit 101 is obtained from the input image using the feature change model whose part separation position is already known. Since the feature change amount similar to the feature change amount is specified and the part separation position of the feature change amount is output, an accurate part separation position can be output. Further, the part separation position output in this way is output from the part separation position output processing unit 2 and input to the clothing region feature extraction processing unit 3 that extracts the clothing region feature amount, for example, as shown in FIG. May be.
  • the clothing region feature extraction processing unit 3 causes the image information input to the clothing region feature extraction processing unit 3, person region information indicating a person region, and the part separation position. It is possible to obtain a clothing region for each part using, and to extract a feature amount of the clothing region using pixel information included in the clothing region.
  • a feature amount of the clothing region for example, a histogram indicating colors / edges (patterns) from the clothing region may be used, or feature amounts indicating other colors / patterns may be used. The feature amount enables searching for a specific person, searching for a person using a clothing area, and the like.
  • FIG. 3 is a block diagram showing an example of another configuration of the part separation position output processing unit 2 shown in FIG.
  • the part separation position output processing unit 2 in the present embodiment includes a part separation position extraction processing unit 107, a feature change model storage unit 102, a person specifying information extraction / collation processing unit 301, and a person attribute storage unit 201. Including.
  • the feature change model accumulation unit 102 accumulates not only the feature change amount or the part separation position but also the above-described personal attributes (part generic name information, part sex information, attribute information) and the part separation section. That is, the feature change model stores the feature change model and the person attribute in association with each other.
  • the person attribute storage unit 201 information for specifying a person, for example, a person individual feature amount used for specifying a person by face recognition, and a person attribute (part sex information, part generic information, attribute information) are associated with each other. Accumulated.
  • the part generic information is a part generic name that summarizes similar clothes when calculating a probability distribution of a part separation position included in a feature change model (T-shirt, skirt, etc.) and a feature change model.
  • Information name upper body clothes (upper body 1), etc.), part generic information (suits) considering the entire clothes.
  • the person attribute storage unit 201 is given a name with the same granularity as the part generic name information stored in the feature change model storage unit 102.
  • the person identification information extraction / collation processing unit 301 extracts a person's face area based on the input image information and person area information, and uses the existing method to identify the individual individual feature amount used for face collation from the face area. Is calculated.
  • the person specifying information extraction / collation processing unit 301 compares this individual individual feature amount with the individual individual feature amount stored in the individual attribute storage unit 201, and determines the person attribute corresponding to the most similar individual individual feature amount as the person. Extracted from the attribute storage unit 201 and output to the part separation position extraction processing unit 107.
  • Various techniques have been proposed for the face area extraction technique and the face recognition technique, but the technique used by the person specifying information extraction / collation processing unit 301 is not particularly limited.
  • the individual individual feature amount is not limited to the feature amount extracted from the face region, and may be a feature amount such as a person's body shape, gait feature, iris information, fingerprint information, ID information assigned to each person.
  • the person identification information extraction / collation processing unit 301 may have a determination function such as gender, age, height, race, hairstyle, and the like. Accordingly, the person specifying information extraction / collation processing unit 301 may obtain the individual individual feature amount based on information such as sex, age, height, race, and hairstyle extracted from the person region.
  • the part separation position extraction processing unit 107 has a function of extracting a part separation position similar to the function of the part separation position extraction processing unit 101 of FIG. 1 described in the first embodiment.
  • the part separation position processing unit 107 acquires the person attribute stored in the person attribute storage unit 201 from the person specifying information extraction / collation processing unit 301 and extracts a person attribute similar to the person attribute from the feature change model storage unit 102.
  • the part separation position processing unit 107 extracts a feature change amount corresponding to a person attribute similar to the extracted person attribute from the feature change model storage unit 102. Accordingly, the part separation position extraction processing unit 107 can narrow down the feature change amount stored in the feature change model storage unit 102. The part separation position extraction processing unit 107 improves collation accuracy and processing speed compared to the part separation position information extraction unit 101 of FIG. Next, the operation of the second embodiment will be described in detail with reference to FIG. From the part separation possibility determination process in step 1001 to the calculation of the feature change amount by projection of the feature change amount in the person vertical direction in step 1004, the part separation position extraction processing unit 107 performs the same process as in FIG.
  • the person identification information extraction / collation processing unit 301 collates the individual individual feature amount extracted from the person region with the individual individual feature amount accumulated in the person attribute accumulation unit 201, and is similar to the person attribute accumulation unit 201. A person attribute associated with a person individual feature amount is extracted.
  • the person specifying information extraction / collation processing unit 301 outputs the extracted person attribute to the part separation position extraction processing unit 107.
  • the part separation position extraction processing unit 107 collates the received person attribute with the person attribute included in the feature change model stored in the feature change model storage unit 102, and calculates the feature change amount associated with the similar person attribute. Extract. That is, the part separation position extraction processing unit 107 narrows down the feature change model (step 1006).
  • FIG. 5 is a block diagram illustrating an example of another configuration of the part separation position output processing unit 2 of FIG.
  • the part separation position output processing unit 2 in the present embodiment includes a part separation position extraction processing unit 105, a feature change model storage unit 102, and a horizontal direction feature change model storage unit 103.
  • the site separation position extraction processing unit 105 includes a vertical direction feature change amount matching processing unit 210, a horizontal direction feature change amount matching processing unit 211, and a site separation position integration processing unit 212.
  • the feature change model storage unit 102 is connected to the vertical feature change amount matching processing unit 210.
  • the horizontal feature change model storage unit 103 is connected to the horizontal feature change amount matching processing unit 211.
  • the part separation position integration processing unit 212 is connected to the vertical direction feature change amount matching processing unit 210 and the horizontal direction feature change amount matching processing unit 211.
  • the feature change model storage unit 102 is the same as that in FIG.
  • the horizontal feature change model storage unit 103 stores feature change amounts with known part separation positions. However, the feature change amounts are not the vertical direction of the person but the horizontal direction. Extracted with respect to direction. In other words, the horizontal direction feature change model accumulation unit 103 scans from the pixel value of the image of the person area in the vertical direction (y-axis direction) of the person, and calculates from the pixel value projected onto the axis in the horizontal direction (x-axis direction). The feature change amount (horizontal feature change amount) is stored.
  • the horizontal direction feature change model accumulating unit 103 includes, in addition to the horizontal direction feature change amount, a part separation section, a person attribute (part sex information, part generic information, attribute information), etc., in the same manner as the feature change amount calculation in the vertical direction. You may save it. At that time, the horizontal direction feature change model accumulating unit 103 may store such information in a list format as shown in FIG. As an example of the part separation section stored in the horizontal direction feature change model storage unit 103, for example, an image of a person wearing a cardigan on a T-shirt and wearing jeans (assuming that the cardigan is open before the cardigan) (See FIG. 15).
  • the cardigan fabric is sandwiched between the T-shirt fabric from the left and right, and the jeans are located at the lower end of the T-shirt.
  • part generic information such as cardigan, T-shirt, jeans, T-shirt, and cardigan is given to the part separation section.
  • the part separation position output processing unit 2 assigns part generic information including clothes having similar horizontal feature change amounts to the feature change model. Also good.
  • the part separation position output processing unit 2 may add not only the part separation section but also part generic name information considering the entire clothing to the feature change model.
  • the process of giving the part generic information as described above is the same as the method described in the feature change model in the vertical direction of the first embodiment.
  • the site separation position output processing unit 2 is determined by men and women determined for each clothes.
  • part sex information such as unisex to the horizontal feature change model.
  • the horizontal direction feature change model accumulating unit 103 may store this part sex information.
  • the part separation position output processing unit 2 can add part sex information of unisex to the feature change model of the horizontal direction feature change model storage unit 103 from clothes information such as jeans, T-shirts, and cardigans.
  • the part separation position output processing unit 2 may also add attribute information such as age, height, race, hairstyle (style, length, etc.) to the feature change model. .
  • the part separation position output processing unit 2 may add to the feature change model if there is attribute information other than the information described above.
  • the horizontal direction feature change model storage unit 103 includes the part sex information, the part generic information, the attribute information, the part separation section length, the ID for indicating the correspondence between the feature change amounts, the part separation position, and the clothes. You may save the information etc. which linked
  • the vertical feature change amount matching processing unit 210 obtains image information, person area information, and part separation possibility determination information from the outside. Then, when it is determined that the part separation is possible from the part separation possibility determination information, the vertical direction feature change amount matching processing unit 210 extracts the feature change amount based on the input image information and person area information. .
  • the vertical feature change amount matching processing unit 210 calculates the similarity between the feature change amount and the feature change amount of the feature change model stored in the feature change model storage unit 102, and has the highest similarity.
  • the part separation position on the vertical axis is output. Similar to the vertical direction feature change amount matching processing unit 210, the horizontal direction feature change amount matching processing unit 211 acquires image information, person area information, and part separation possibility determination information from the outside. When it is determined that the part separation is possible from the part separation possibility determination information, the horizontal direction feature change amount matching processing unit 211 extracts the part separation position. However, the horizontal feature change amount matching processing unit 211 extracts a part separation position in the horizontal direction of the person, not in the vertical direction of the person.
  • the horizontal direction feature change amount matching processing unit 211 calculates the horizontal direction feature change amount, and the calculated horizontal direction feature change amount and the horizontal direction feature change amount extracted from the horizontal direction feature change model storage unit 103. And the part separation position of the horizontal feature change model having the highest horizontal feature change amount is output.
  • the part separation position integration processing unit 212 acquires the vertical part separation position extracted by the vertical feature change amount matching processing unit 210 and the horizontal part separation position extracted by the horizontal direction feature change amount matching processing unit 211. Then, those pieces of information are integrated to output the part separation position of the entire person.
  • the part separation position integration processing the part separation position integration processing unit 212 generates information for dividing the person region by combining the part separation positions in each direction.
  • the part separation position integration processing unit 212 extracts the positions of the head, upper body, and lower body areas from the part separation section, the part generic information, and the person area that are stored as the part separation positions in the vertical direction, and the horizontal direction Using this part separation position, an arm region (a clothing region when layered) that is a part of the upper body can be extracted from the upper body region of the person.
  • the part separation position integration processing unit 212 determines the position of each part region in the horizontal and vertical directions, and thus can extract more detailed part separation positions. Further, the part separation position integration processing unit 212 can increase the accuracy of the clothes name and part sex information of the person region by using the part separation positions in the horizontal and vertical directions.
  • FIG. 6 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 105 illustrated in FIG. In order to extract the part separation position from the person region, in the third embodiment (see FIG.
  • the vertical direction feature change amount matching processing unit 210 and the horizontal direction feature change amount matching processing unit 211 are arranged in parallel. A feature change amount matching process was performed.
  • the vertical direction feature change matching processing unit 210 and the horizontal direction feature change matching processing 213 are connected in series. The purpose of connecting in series is that the horizontal feature change amount matching process 213 extracts a horizontal part separation position for each part divided in the vertical direction extracted by the vertical direction feature change amount matching processing unit 210. It is. Cases where the number of occurrences of feature changes in the vertical direction of a person is greater than the number of occurrences of feature changes in the horizontal direction of the person, so that the part separation position in the horizontal direction is more difficult to extract than the part separation position in the vertical direction Can occur.
  • the part separation position extraction processing unit 105 first specifies the part separation position in the vertical direction as in the configuration of FIG. 6, and extracts the horizontal feature change amount for each part in the vertical direction, thereby achieving higher accuracy.
  • the site separation position can be extracted. At this time, it is not necessary to extract the horizontal direction feature change amount for all the parts in the vertical direction.
  • the part separation position extraction processing unit 105 may perform the horizontal feature change amount extraction when the upper body of the person is layered and the lower body is pants. On the other hand, the part separation position extraction processing unit 105 may not calculate the horizontal feature change amount for an area that does not need to be further divided, such as a head part or a shoe part.
  • the site separation position extraction processing unit 105 in the present embodiment includes a vertical direction feature change amount matching processing unit 210, a horizontal direction feature change amount matching processing unit 213, and a site separation position integration processing unit 212.
  • the part separation position integration processing unit 212 is connected to the vertical direction feature variation matching processing unit 210 and the horizontal direction feature variation matching processing unit 213.
  • the part separation position integration processing unit 212 includes information such as the vertical part separation position of the vertical feature change amount matching processing unit 210 and information such as the horizontal part separation position of the horizontal feature change amount matching processing unit 213. To generate and output information such as part separation positions in the entire person area.
  • the horizontal direction feature amount matching processing unit 213 is connected to the vertical direction feature amount matching processing unit 210 and the part separation position integration processing unit 212. Image information, person region information, and a vertical feature change model are input to the vertical feature change amount matching processing unit 210 from the outside. Also, the horizontal part separation position generated by the vertical direction feature variation matching processor 210 is input to the horizontal feature variation matching processor 213.
  • the horizontal feature change amount matching processing unit 213 extracts a horizontal feature change amount for each vertical part separation section based on the vertical part separation position, and the horizontal feature change amount and the input horizontal feature change amount are extracted.
  • the horizontal feature change model included in the direction feature change model is collated, and a horizontal feature change model including a similar feature change amount is searched from the horizontal feature change model storage unit 103.
  • the horizontal feature change amount matching processing unit 213 outputs the horizontal part separation position of the feature change model including the similar feature change amount to the part separation position integration processing unit 212.
  • the part separation position integration processing unit 212 integrates the vertical part separation position of the vertical direction feature change amount matching processing unit 210 and the horizontal part separation position of the horizontal direction feature change amount matching processing unit 213, and Generate and output site separation information.
  • FIG. 7 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
  • the part separation position extraction processing unit 101 of this embodiment includes a person vertical direction extraction processing unit 207 and a feature change amount matching processing unit 206, and the feature change amount matching processing unit 206 is connected to a feature change model storage unit 102 (not shown).
  • the feature change model storage unit 102 is the same as that shown in FIG.
  • the person vertical direction extraction processing unit 207 receives image information and person area information from the outside.
  • the person vertical direction extraction processing unit 207 estimates the vertical direction of the person area from the input image information and person area information.
  • There are various methods for estimating the vertical direction For example, a head detector or a foot detector generated by learning in advance is used to detect the head and feet, and a method for estimating the vertical direction from a vector connecting the two points of the head and feet, or represents the head or feet 2D coordinates on the image are converted into 3D space coordinates using the camera calibration information, and 3D space coordinates of a point on a straight line passing through the point and parallel to the vertical direction are obtained and converted back to 2D.
  • a head detector or a foot detector generated by learning in advance is used to detect the head and feet
  • a method for estimating the vertical direction from a vector connecting the two points of the head and feet, or represents the head or feet 2D coordinates on the image are converted into 3D space coordinates using the camera calibration information, and 3D space coordinate
  • the person vertical direction extraction processing unit 207 converts the image coordinates from two-dimensional to three-dimensional based on the position information of the space to be photographed and information such as camera parameters. Convert the foot position of the two-dimensional coordinates to the foot position of the three-dimensional coordinates by using a transformation matrix that projects in reverse, and set the three-dimensional coordinates set with an appropriate height from the foot positions and reversely transform to two dimensions.
  • the vertical direction of the person on the two-dimensional space that is, the image can be extracted.
  • the vertical direction of the person can be automatically set even in a region where the y-axis and the vertical direction do not coincide with each other, such as a region around the lens, due to lens distortion or the angle of view.
  • the separation position can be acquired.
  • the feature change amount matching processing unit 206 has a function of extracting a part separation position in the same manner as the part separation position extraction processing unit 101 in FIG. Similarly to the first embodiment, the feature change amount matching processing unit 206 performs the first embodiment based on the information in the vertical direction estimated for each person area acquired from the person vertical direction extraction processing unit 207. Similar to the form, the image feature amount of the person region is calculated.
  • FIG. 8 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
  • the part separation position extraction processing unit 101 in this embodiment includes a person orientation determination unit 209 and a feature change amount matching processing unit 208, and the feature change amount matching processing unit 208 is connected to a feature change model storage unit 102 (not shown). ing.
  • the person orientation determination unit 209 is connected to the feature change amount matching processing unit 208.
  • a feature change model storage unit 102 (not shown) stores feature change models for each direction of a person.
  • the direction of a person (front, back, sideways, etc.) is expressed by an angle, a vector, or the like based on a specific direction.
  • a coordinate system is defined on the basis of the direction in which the person is facing the front toward the camera, and the direction of the person is expressed by an angle or a vector direction. These angles and vector information are defined as person orientation information.
  • the feature change model storage unit 102 stores feature change models classified according to person orientation information.
  • the person orientation determination unit 209 determines the direction in which the person is facing based on the image information and the person area information, and outputs the person orientation information.
  • There are various methods for determining the direction of a person but the direction of the face that the person faces by face recognition is set to the direction that the person is facing, or the direction of travel by optical flow using images input in time series
  • the person orientation determination unit 209 needs to know the camera coordinate system.
  • the person orientation determination unit 209 is configured to display the rear side when the person moves away from a straight line with respect to the camera orientation.
  • the direction of the person can be estimated as a front direction when approaching linearly and as a landscape direction when crossing.
  • the feature change amount matching processing unit 208 has a function of extracting a part separation position in the same manner as the part separation position extraction processing unit 101 in FIG.
  • the feature change amount matching processing unit 208 compares the feature change amount for the input image with the feature change amount of the feature change model stored in the feature change model storage unit 102 before the person orientation determination unit 209.
  • the person orientation information obtained from the above and the person orientation information that classifies the feature change model are collated.
  • the feature change amount matching processing unit 208 compares, for example, the angle of person orientation information and the similarity between vectors. For example, when the person orientation information is an angle, the feature change amount matching processing unit 208 may compare the similarity based on the difference between the angles, and may compare the similarity based on an inner product if the information is a vector. However, other methods for comparing similarities may be used. Thereafter, the feature change amount matching processing unit 208 selects the person orientation information with the most similar person orientation information, selects the feature change model corresponding to the person orientation information, and compares the feature change amounts. Accordingly, the feature change amount matching processing unit 208 can perform matching between feature change amounts of the same or similar person orientation.
  • FIG. 9 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
  • the part separation position extraction processing unit 101 in the present embodiment includes a part separation position accumulation unit 203, a person region information accumulation unit 204, and a feature change amount matching processing unit 205, and the feature change amount matching processing unit 205 is not illustrated.
  • the feature change model storage unit 102 is connected.
  • the part separation position accumulating unit 203 accumulates the feature change amount (or part separation position) calculated by the feature change amount matching processing unit 205 together with time stamp information at the time of image information acquisition.
  • the part separation position accumulating unit 203 outputs the past feature change amount (or part separation position) and time stamp information that have already been accumulated to the feature change amount collation processing unit 205.
  • the person area information accumulation unit 204 accumulates time stamp information at the time of image information acquisition and person area information.
  • the person area information storage unit 204 outputs the past person area information that has already been stored to the feature change amount matching processing unit 205 together with the time stamp information.
  • the feature change amount matching processing unit 205 receives externally input image information to which time stamp information is added, person region information, and part separation possibility determination information.
  • the feature change amount matching processing unit 205 receives a feature change model from a feature change model storage unit 102 (not shown).
  • the feature change amount matching processing unit 205 is close to the time stamp information of the input image information (the time indicated by the time stamp information and a time within a predetermined time), and the position and size of the person area are close ( In order to collate a past person area (with an error within a predetermined range), it is connected to the person area information storage unit 204.
  • the feature change amount matching processing unit 205 is connected to the part separation position accumulating part 203 in order to extract a part separation position at a time close to the time stamp information.
  • the feature change amount matching processing unit 205 can extract the past person regions at a time close to the time stamp information. It may be regarded as a person area of the same person. This is because it can be assumed that a person who appears in an image that has not passed much time will not change its position or size abruptly.
  • the feature change amount matching processing unit 205 calculates an optical flow that calculates the change in the luminance information of the image and estimates the movement amount of the pixel. It may be used.
  • the feature change amount matching processing unit 205 can regard a person area with a small amount of movement as a person area of the same person. Further, the feature change amount matching processing unit 205 can also determine whether or not the person area is the same person by extracting the pixel value of the person area range from the image information and looking at the same color and luminance. . However, in this case, the person area information storage unit 204 stores not only the person area information but also the image information together with the time stamp information. The feature change amount matching processing unit 205 extracts a person area regarded as the person area of the same person from the person area information storage unit 204, and time stamp information at a time close to the time stamp information given to the person area information.
  • the amount of feature change (or part separation position) to which is added is extracted from the part separation position accumulation unit 203.
  • This feature change amount (or part separation position) is considered to represent the feature change amount (or part separation position) of the same person.
  • the feature change amount matching processing unit 205 extracts the feature change amount extracted from the extracted feature change amount (or part separation position) and the image information and person area information currently input to the feature change amount matching processing unit 205.
  • the similarity with the part separation candidate position calculated in the same manner as in the first embodiment can be calculated, and the most similar part separation position can be extracted.
  • the feature change amount matching processing unit 205 outputs the part separation position by the same processing as in the first embodiment.
  • the part separation position extraction processing unit 101 uses the average position of the plurality of part separation positions or the time stamp is new.
  • a part separation position can be created using a weighted average position that gives a greater weight to the part separation position.
  • the part separation position extraction processing unit 101 can create a part separation position by obtaining a probability distribution (for example, a Gaussian distribution) indicating the frequency of the part separation positions extracted from a plurality of images. Thereby, the part separation position caused by the noise of the image information can be removed, and the matching accuracy of the part separation position is improved.
  • the person region information storage unit 204 stores an ID for specifying a person in association with the part separation position.
  • the feature change amount collation processing unit 205 extracts a person area to which a time stamp that is similar to the time stamp given to the input image is given from the person area information storage unit 204, and determines the size and position of the person area. From this, it is determined whether or not the same person area (step 1007). When it is determined in step 1007 that the person area is the same person (step 1007, Yes), the feature change amount matching processing unit 205 is given time stamp information at a time close to the time stamp information from the part separation position accumulating unit 203.
  • the amount of feature change (or part separation position) is extracted.
  • the feature change amount matching processing unit 205 obtains a feature change amount based on the image information input to the feature change amount matching processing unit 205 and the person area (or described in step 1005 of the first embodiment).
  • the part separation candidate position is extracted from the feature change amount by the technique).
  • the feature change amount matching processing unit 205 determines the similarity between the feature change amount (or part separation candidate position) and the feature change amount (or part separation position) extracted from the part separation position storage unit 203 in the first embodiment. Calculation is performed by the method described in the embodiment, and a part separation position is extracted (step 1008).
  • the feature change amount matching processing unit 205 outputs the part separation position by the same process as in the first embodiment.
  • the part separation position extraction apparatus 1 includes a feature change model accumulation unit 102 and a part separation position extraction processing unit 101, as shown in FIG. Since these configurations and operations are as described above, detailed description thereof will be omitted.
  • the feature change model storage unit 102 For each person area of the plurality of images, the feature change model storage unit 102 includes a plurality of axes parallel to a predetermined axis (for example, a horizontal axis and an x axis) defined on the image, and a person area included in each image. , And a feature change amount corresponding to a change in a position on a vertical axis (for example, a vertical axis or a y-axis) of a value of an image feature amount defined on each line segment intersecting with a predetermined axis is stored as a model.
  • a predetermined axis for example, a horizontal axis and an x axis
  • the part separation position extraction processing unit 101 calculates a feature change amount for a human region on an image input from the outside, and calculates the feature change amount and the feature change amount stored in the feature change model storage unit 102. The comparison is performed, the similar change amount is specified, and the peak position where the specified change value is a predetermined value or more is output. According to the present embodiment, by outputting the peak position of a feature change model in which the change amount of the image feature quantity is similar to the clothes worn by the person on the input image, the parts of the clothes are separated. And can be extracted. While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

Provided are a site separation location extraction device, program, and method, with which it is possible to separate and extract a region in an image, such as a person or an object. The site separation location is extracted by: a feature quantity being computed from a person region which is obtained from an image; a value of a person in which the feature quantity is projected in a prescribed axial direction being extracted as a feature change quantity; and the feature change quantity and a feature change quantity for a pre-accumulated known site separation location being compared.

Description

部位分離位置抽出装置、プログラム、方法Site separation position extraction apparatus, program, and method
 本発明は部位分離位置抽出装置、プログラム、方法に関する。 The present invention relates to a part separation position extraction device, a program, and a method.
 監視者等は、監視目的として設置されたカメラから不審者を発見し、その人物の特徴を表現するために、人物が身に着けている服装の色や形状、種類などの情報を活用できる。また、監視者等は、服飾品の販売店へ来る客の傾向を把握するためのマーケティング情報を収集するために、人物が身に着けている服装の色や形状、種類などの情報を活用できる。
 この分野に関し、特許文献1には、特定の服を着ている人物が写っている画像の検索方法の一例が開示されている。また、特許文献2には、入力画像と類似する画像を検索する画像検索方法の一例が開示されている。
A monitor or the like can use information such as the color, shape, and type of clothes worn by a person in order to discover a suspicious person from a camera installed for monitoring purposes and express the characteristics of the person. Monitors can also use information such as the color, shape, and type of clothes worn by a person to collect marketing information to understand the trends of customers coming to clothing stores. .
In this field, Patent Document 1 discloses an example of a method for searching for an image showing a person wearing a specific clothes. Patent Document 2 discloses an example of an image search method for searching for an image similar to an input image.
特開2008−139941号公報JP 2008-139951 A 特開2007−026386号公報JP 2007-026386 A
 しかし、上述の方法は、画像上の人物や物体等の領域を分離して抽出することができない。その結果、上述の方法は、画像情報から、人物が身に着けている服装等を分離して抽出することもできない。
 本発明の目的は、上記課題を解決するため、画像上の人物や物体等の領域を分離して抽出することができる部位分離位置抽出装置、プログラム、方法を提供することである。
However, the above-described method cannot separate and extract regions such as persons and objects on the image. As a result, the above-described method cannot separate and extract clothes worn by a person from image information.
In order to solve the above problems, an object of the present invention is to provide a part separation position extraction apparatus, program, and method capable of separating and extracting a region such as a person or an object on an image.
 上記目的を達成するため、本発明は、画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納する特徴変化モデル蓄積手段と、外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記モデルの変化量との比較を行ない、類似する変化量をもつモデルを前記特徴変化モデル蓄積手段から抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出処理手段と、を備える部位分離位置抽出装置を提供する。
 また、本発明は、画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納する特徴変化モデル蓄積手段を備えるコンピュータに、外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記特徴変化モデル蓄積手段に格納された前記変化量との比較を行ない、類似する変化量をもつモデルを抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出処理ステップを実行させる部位分離位置抽出プログラムを格納するプログラム記憶媒体を提供する。
 また、本発明は、画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納し、外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記特徴変化モデル蓄積手段に格納された前記変化量との比較を行ない、類似する変化量をもつモデルを抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出方法を提供する。
To achieve the above object, according to the present invention, for each person area on the image, each line segment in which a plurality of axes parallel to a predetermined axis on the image intersects the person area included in the image. Feature change model accumulating means for storing, as a model, a change amount of a calculated feature amount value with respect to a change in a position on a vertical axis with respect to the predetermined axis, and the change amount for a person region of an external input image. Calculating, comparing the amount of change with the amount of change of the model, extracting a model having a similar amount of change from the feature change model accumulating means, and the value of the amount of change of the extracted model is a predetermined value or more And a part separation position extraction processing means for outputting a peak position which is a part of the part separation position extraction device.
Further, according to the present invention, for each person region on the image, a feature amount calculated for each line segment where a plurality of axes parallel to a predetermined axis on the image and a person region included in the image intersect The change amount is calculated for a human region of the input image from the outside in a computer having a feature change model accumulating unit that stores a change amount of the value of the value with respect to a change in position on the vertical axis with respect to the predetermined axis as a model. The change amount is compared with the change amount stored in the feature change model accumulating means to extract a model having a similar change amount, and the value of the change amount of the extracted model is not less than a predetermined value. Provided is a program storage medium for storing a part separation position extraction program for executing a part separation position extraction processing step for outputting a peak position as a part.
Further, according to the present invention, for each person region on the image, a feature amount calculated for each line segment where a plurality of axes parallel to a predetermined axis on the image and a person region included in the image intersect Is stored as a model with respect to a change in position on the vertical axis with respect to the predetermined axis, the change amount is calculated for a human region of an input image from the outside, and the change amount and the feature change are calculated. A part that compares the amount of change stored in the model storage means, extracts a model having a similar amount of change, and outputs a peak position where the value of the amount of change of the extracted model is a predetermined value or more A separation position extraction method is provided.
 本発明によれば、画像上の人物や物体等の領域を分離して抽出することができる部位分離位置抽出装置、プログラム、方法を提供することができる。 According to the present invention, it is possible to provide a part separation position extraction apparatus, program, and method that can separate and extract a region such as a person or an object on an image.
第1の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 1st Embodiment. 第1の実施の形態の動作の一例を示すフローチャート図である。It is a flowchart figure which shows an example of operation | movement of 1st Embodiment. 第2の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 2nd Embodiment. 第2の実施の形態の動作の一例を示すフローチャート図である。It is a flowchart figure which shows an example of operation | movement of 2nd Embodiment. 第3の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 3rd Embodiment. 第4の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 4th Embodiment. 第5の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 5th Embodiment. 第6の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 6th Embodiment. 第7の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 7th Embodiment. 第7の実施の形態の動作の一例を示すフローチャート図である。It is a flowchart figure which shows an example of operation | movement of 7th Embodiment. 部位分離可否判定情報の一例を説明する図である。It is a figure explaining an example of site | part separation availability determination information. 人物領域の鉛直方向における特徴変化量と部位分離位置の一例を示す図である。It is a figure which shows an example of the characteristic variation | change_quantity and site | part separation position in the vertical direction of a person area. 部位分離位置出力処理部と服装領域特徴抽出処理部の一例を示す図である。It is a figure which shows an example of a site | part separation position output process part and a clothing area | region feature extraction process part. 特徴変化モデル蓄積部に蓄積される関連付けリスト情報の一例である。It is an example of the association list information stored in the feature change model storage unit. 特徴変化モデル蓄積部に蓄積される関連付けリスト情報の一例である。It is an example of the association list information stored in the feature change model storage unit. 水平方向特徴変化モデル蓄積部103に格納される部位分離区間の一例を示す図である。It is a figure which shows an example of the site | part separation area stored in the horizontal direction characteristic change model storage part 103. FIG. 第8の実施の形態の構成の一例を示すブロック図である。It is a block diagram which shows an example of a structure of 8th Embodiment.
 以下、本発明の実施の形態について、図面を用いて説明する。すべての図面において、同様な構成要素には同様の符号を付し、適宜説明を省略する。
 なお、各実施形態の装置等を構成する各部は、制御部、メモリ、メモリにロードされたプログラム、プログラムを格納するハードディスク等の記憶ユニット、ネットワーク接続用インターフェースなどからなり、ハードウェアとソフトウェアの任意の組合せによって実現される。そして特に断りのない限り、その実現方法、装置は限定されない。
 また、制御部はCPU(Central Processing Unit)などからなり、オペレーティングシステムを動作させて装置等の全体を制御するとともに、例えばドライブ装置などに装着された記録媒体からメモリにプログラムやデータを読み出し、これに従って各種の処理を実行する。
 記録媒体は、例えば光ディスク、フレキシブルディスク、磁気光ディスク、外付けハードディスク、半導体メモリ等であって、コンピュータプログラムをコンピュータが読み取り可能に記録する。また、コンピュータプログラムは、通信網に接続されている図示しない外部コンピュータからダウンロードされても良い。ここで、特に断りの無い限り、通信網は、インターネット、LAN(Local Area Network)、公衆回線網、無線通信網、または、これらの組み合わせ等によって構成されるネットワーク等であって良い。
 各実施形態の装置等を構成する各部は、論理回路等のハードウェアやソフトウェア等の組み合わせによって構成される。各実施形態の説明において利用するブロック図は、ハードウェア単位の構成ではなく、機能単位の構成を示している。これらの機能ブロックはハードウェア、ソフトウェアの任意の組み合わせによって実現される。また、これらの図においては、各実施形態の構成部は物理的に結合した一つの装置により実現されるよう記載されている場合もあるが、その実現手段はこれに限定されない。すなわち、各実施形態で説明される構成部、装置、システム等は、二つ以上の物理的に分離した装置を有線または無線で接続することにより複数の装置により実現されてもよい。また、各構成部が物理的に分離した2つ以上装置として記載されている場合もあるが、その実現手段はこれに限定されない。すなわち、物理的に結合した一つの装置により実現されるようにハードウェア、ソフトウェアを任意に組み合わせることにより各実施形態の各構成部、装置、システム等が実現されてもよい。
 (第1の実施の形態)
まず、第1の実施の形態について図1を用いて説明する。
 図1は、本発明の第1の実施の形態で、部位分離位置抽出装置1の一例を示すブロック図である。部位分離位置抽出装置1には部位分離位置出力処理部2が含まれる。部位分離位置出力処理部2には、部位分離位置抽出処理部101と特徴変化モデル蓄積部102が含まれ、部位分離位置抽出処理部101に特徴変化モデル蓄積部102が接続されている。
 特徴変化モデル蓄積部102は、図12のように、人物が写っている画像上で定義される所定軸方向、例えば、人物の鉛直方向(頭部と足部とを結ぶ軸方向)の画像特徴量(後述)の変化量(特徴変化量)、あるいは、当該変化量と、服装や体の部位の分離位置(頭頂、Tシャツ上端、Tシャツ下端、スカート上端、スカート下端頭部などの、服装や体の部位の境目の箇所)を示す部位分離位置と、を関連付けた特徴変化モデルを蓄積する。画像特徴量と特徴変化量については後述する。
 次に、部位分離位置出力処理部2が、特徴変化モデル蓄積部102に特徴変化モデルを蓄積する方法について説明する。部位分離位置出力処理部2は、外部から入力された画像情報(単に画像ともいう)から人物の領域を示す人物領域を既存の方法を用いて抽出し、その人物領域のサイズを、鉛直方向の所定のサイズに拡大ないし縮小することで人物領域のサイズを正規化する。また、部位分離位置出力処理部2は、人物領域を水平方向にスキャンする、すなわち、水平方向軸に平行な軸と人物領域とが交わった線分上の画素値(例えば、画素の色の種類や強さ、明るさを表す数値等)を基に、人物の鉛直方向の軸上の各位置における服装の特徴を示す画像特徴量を算出する(後述の式1参照)。なお、鉛直方向は、画像の人物の頭部と足部とを結ぶ軸方向であり、部位分離位置出力処理部2が公知の方法を用いて画像から算出しても良いし、予めカメラの垂直方向が鉛直方向と一致するようにカメラが設定されており、部位分離位置出力部2は画像の垂直方向を鉛直方向としても良い。
 そして、部位分離位置出力処理部2は、人物領域の鉛直方向で起こる画像特徴量の変化を示す特徴変化量を算出し、該特徴変化量を、人物が着ている服装の種類と対応付ける。なお、画像上の人物が着ている服装の情報は予め与えられている。
 部位分離位置出力処理部2は、服装ごとに、各特徴変化量を求める(図12参照)。また、部位分離位置出力処理部2は、各特徴変化量のピークの位置を部位分離位置とし、部位分離位置と、求めた変化量とを関連付けた情報を、上述の特徴変化モデルとして、特徴変化モデル蓄積部102に蓄積する。
 ところで、服装の種類ごとに特徴変化量が分類された際、各分類の特徴変化量同士が類似することがある。例えば、セーターとジーンズという服装から求めた特徴変化量と、Yシャツとスラックスという服装から求めた特徴変化量は類似する可能性が高い。そのため、部位分離位置出力処理部2は、類似する特徴変化量を1つにまとめても良い。例えば、部位分離位置出力処理部2は、類似する特徴変化量を1つの特徴変化量にまとめる際に、各特徴変化量の平均を用いて特徴変化量をまとめても良いし、特徴変化量同士が殆ど同じなら、類似する特徴変化量の中から代表とする特徴変化量を所定の基準で1つ選択しても良い。また、部位分離位置出力処理部2は、複数の特徴変化量の代表となる特徴変化量を算出する他の手法を用いてもかまわない。
 部位分離位置出力処理部2は、部位分離位置が類似する服装を集める際に、部位分離位置の数や部位分離位置間の差分距離を用いて算出した類似度を用いたり、DP(DynamicProgramming)マッチング等の手法を用いて算出した類似度を用いたりすることができるが、その他の類似度算出手法を用いても良い。
 また、部位分離位置出力処理部2は、人物の服装の特徴を示す画像特徴量の算出方法として、人物領域に含まれる画像を水平方向にスキャンすることで得た画素値を利用する代わりに以下の方法を用いても良い。すなわち、部位分離位置出力処理部2は、人物領域に含まれる、ある画素の画素値を抽出し、その画素値と、その画素の周辺にある画素の画素値とを基に画像特徴量を算出し、その画像特徴量を水平方向にスキャンすることで新たな画像特徴量を算出しても良い。例えば、部位分離位置出力処理部2は、1画素ではなく、複数画素をまとめて1つの画素とみなして、それらの複数画素の画素値の平均値を水平方向にスキャンすることで画像特徴量を算出しても良い。
 なお、部位分離位置出力処理部2は、部位分離位置で分けられた区間(部位分離区間)に対して服装の名称や体の部位の名称(部位総称情報)を付与し、特徴変化モデルとして特徴変化モデル蓄積部102に蓄積しても良い。例えば、図12のようにTシャツにスカート、ストッキング、靴を身に着けている女性は、少なくとも頭頂、Tシャツ上端、Tシャツ下端、スカート上端、スカート下端、靴の間に部位分離位置が存在する。したがって、部位分離位置出力処理部2は、例えば、区間長と部位総称情報とが対応付けられて記憶された不図示記憶部を参照して、各々の部位分離区間に顔、Tシャツ、スカート、脚という部位総称情報を付与してもよい。
 ところで、種類が異なる服装であっても、鉛直方向の服装の丈の長さが同程度の服装は多く存在する。そのため、類似する特徴変化量を1つにまとめる場合に、種類が異なる服装の特徴変化量がまとめられて同一の特徴変化モデルが算出される可能性が高い。そこで、部位分離位置出力処理部2は、特徴変化モデルに含まれる特徴変化量を算出する際に、丈の長さが類似する服装をまとめた部位総称情報の名称をさらに加えてもよい。例えば、部位分離位置出力処理部2は、Tシャツ、ジャージ、ブルゾン、ニットなどを“短い上半身の服装(上半身1)”、という分類名にしても良いし、ロングコート、雨合羽などを“長い上半身の服装(上半身2)”、ミニスカート、ショートパンツ、ハーフパンツを“短い下半身の服装(下半身1)”、七部丈のズボンやロングスカートなど足首付近などが見える下半身の服装を“中間丈の下半身の服装(下半身2)”、足首が見えないような長いズボン、ロングスカートなどを“長い下半身の服装(下半身3)”という分類名にしてもよい。さらに、部位分離位置出力処理部2は、部位分離区間だけでなく、服装全体についての部位総称情報を特徴変化モデルに付与してもよい。例えばジャケットと、スラックスとの組み合わせには、“スーツ類”という部位総称情報が付与されても良い。
 ところで、服装によっては、その服装を身につけられる性別が定まっている場合がある。そのため、部位分離位置出力処理部2は、服装ごとに定まっている男性、女性、男女兼用などの情報(部位性別情報)を特徴変化モデルに付与しても良い。部位分離位置出力処理部2は、この部位性別情報も、特徴変化モデル蓄積部102に保存しても良い。部位分離位置出力処理部2は、例えば、スカートを身に着ける人物は一般に女性であるため、例えば、部位総称情報と部位性別情報とが対応付けられて記憶された不図示記憶部を参照して、スカートに部位性別情報として女性を付与することが可能である。
 部位分離位置出力処理部2は、さらに、人物の性別、年齢、身長、人種、髪型(長さなど)などの属性情報を、部位性別情報と共に特徴変化モデルに付与しても良い。例えば、属性情報は、性別なら“男・女”、髪型なら“長さ”や“スタイル”、年齢なら“年齢層”、人種なら“肌の色”などである。なお、部位分離位置出力処理部2は、公知の技術を用いて、画像に写っている人物からこれらの属性情報を抽出することが可能である。
 部位分離位置出力処理部2は、これらの属性情報も、特徴変化モデル蓄積部102に蓄積しても良い。また、部位分離位置出力処理部2は、上述した情報以外にその他の属性情報があれば、特徴変化モデルに付与することができ、特徴変化モデル蓄積部102に蓄積できる。
 これら蓄積された部位性別情報、部位総称情報、属性情報は、後述する特徴変化モデルのマッチングの際に、特徴変化モデル蓄積部102から取り出す特徴変化モデルの絞込みに用いることができる。部位性別情報、部位総称情報、属性情報をまとめて人物属性と呼ぶ。
 部位性別情報、部位総称情報、属性情報、部位分離区間の長さ(部位分離区間長)、特徴変化量の対応関係を示すためのID(IDentification)等は、例えば、図14のような関連付けリスト情報として特徴変化モデル蓄積部102に保存されても良い。保存される形式は、他のデータ形式でも良い。
 特徴変化モデル蓄積部102は、特徴変化モデルを、人物の方向別に蓄積しても良い。部位分離位置出力処理部2は、公知の技術によって、例えば人物が横方向や後ろを向いていることを検出し、各人物の方向別に特徴変化量と分離位置情報とを算出し、方向別にモデル化して特徴変化モデル蓄積部102に蓄積しても良い。特徴変化モデル蓄積部102は、例えば、前方向、横方向、後ろ方向からの見え方が異なるスーツ類に対して、前方向、横方向、後ろ方向のそれぞれに対して構築された特徴変化モデルを方向別に蓄積してもよい。
 なお、上述の処理は、部位分離位置出力処理部2が特徴変化モデル蓄積部102に特徴変化モデルを蓄積する処理として説明したがそれに限定されない。例えば、図示しない外部装置等によって予め生成された特徴変化モデルが特徴変化モデル蓄積部102に格納されても良い。
 以上のように、特徴変化モデル蓄積部102には、図12のように人物の鉛直方向の軸上の各位置における画像特徴量の特徴変化量、あるいは、特徴変化量と、部位分離位置とが関連付けられた特徴変化モデルが蓄積される。また、上述のように、特徴変化モデル蓄積部102には、人物属性(部位性別情報、部位総称情報、属性情報)、部位分離区間長がさらに格納されてもよい。また、特徴変化モデル蓄積部102には、人物の鉛直方向軸上の画像特徴量の特徴変化量が閾値以上の箇所が出現する頻度を示す確率分布が格納されてもよい。
 部位分離位置抽出処理部101は、画像情報と、人物領域情報(画像情報の中で人物が写っている領域の箇所を示す情報)と、その人物領域の服装が分離可能であるか否かを判定する部位分離可否判定情報と、を用いて、図12の右側グラフのように、人物領域の特徴変化量を算出し、当該特徴変化量と、特徴変化モデル蓄積部102に蓄積された特徴変化モデルの特徴変化量との類似度を算出する。部位分離位置抽出処理部101は、部位分離位置の数や部位分離位置間の差分距離を利用したり、DP(Dynamic Programming)マッチングといった手法を用いてこれらの類似度を算出することができるが、その他の類似度算出手法を用いても良い。
 部位分離可否判定情報とは、画像中の人物の写り方によって人物の部位分離が可能か否かを示す情報である。部位分離可否判定情報には、分離の可否を決める値が記述されている。例えば、部位分離可否判定情報は、可または否を表す2値、または、閾値処理を行うための数値などである。部位分離可否情報は図示しない外部装置によって算出される。該外部装置は、例えば、画像中において人物領域が全身写る、または、下半身が一部隠れる、などを示す人物の写り方を求め、その写り方によって部位分離可否の判定を算出することができる。部位分離可否情報の具体例を図11を用いて説明する。図11の左図は入力画像であり、人物が3人おり、1人は机の奥に立っているため上半身しか見えない状況を示している。この場合、外部装置は、例えば人物の足元位置を基に公知の方法によって、図11の右図のような領域を算出できる。図11の右図は、机の上端は上半身しか写らない領域(黒い領域)であり、床面は全身が写る領域(斜線領域)であり、壁は人物の足元位置が存在できない領域(白い領域)であることを示している。このように、部位分離可否判定情報は、画像中に部位分離できる可能性がある領域が存在すれば“可”となり、存在しなければ“否”となる。
 部位分離位置抽出処理部101は、最も類似度が高い特徴変化量をもつ特徴変化モデルの部位分離位置を基に、入力された人物領域の部位分離位置を出力する。この際、部位分離位置抽出処理部101は、特徴変化モデル蓄積部102から、部位総称情報、部位性別情報、属性情報、部位分離区間長などの情報を併せて受け取り、出力することもできる。例えば、部位分離位置抽出処理部101は、入力された画像内の人物が身につける服装の特徴変化量が、特徴変化モデル蓄積部102内に蓄積された複数の特徴変化量と類似する場合、部位総称情報を出力することができる。例えば、Tシャツを着た人物と、ニットやトレーナーを着た人物とを部位分離位置で区別するのは難しいため、部位分離位置抽出処理部101は、上半身に対してこれらを総称する“短い上半身の服装(上半身1)”といった部位総称情報を出力することができる。
 続いて第1の実施の形態の動作について、図2を用いて詳細に説明する。
 部位分離位置抽出処理部101は、部位分離可否判定情報を基に、画像上の人物領域から部位分離が可能であるか否かを判定する(ステップ1001)。部位分離位置抽出処理部101は、部位分離可否判定情報を基に分離の可否を判定する。
 部位分離可否判定が可である場合(ステップ1001、yes)、部位分離位置抽出処理部101は、人物領域の正規化を行う(ステップ1002)。正規化を行う理由は、人物領域サイズ変化による特徴変化量の変動を抑えるためである。具体的な正規化手法としては、部位分離位置抽出処理部101は、人物領域の鉛直方向の長さが基準値に揃うように、縦横の縦横比率を維持した画像の拡大または縮小処理を行う。拡大または縮小処理の方法は、一般的な画像の拡大・縮小アルゴリズムでよく、例えば、ニアレストネイバー法、バイリニア法、バイキュービック法などである。
 正規化後、部位分離位置抽出処理部101は、人物領域の水平方向の画像特徴量を算出する。部位分離位置抽出処理部101は、例えば、人物領域情報が示す範囲の画像の画素値を水平方向(x軸方向)にスキャンし、鉛直方向(y軸方向)に射影した画素値などの関数を求め、それを画像特徴量とする(ステップ1003)。部位分離位置抽出処理部101は、具体的には、例えば(式1)のように、人物領域とx軸とが交わる線分上で画像特徴量を求める。
Figure JPOXMLDOC01-appb-I000001
 ここで、I(x、y)は入力画像上における座標(x、y)における画素値(例えば、R(Red)、G(Green)、B(Blue)の3次元ベクトルから算出された値)、M(x、y)は、人物領域情報から生成したマスク情報、X0は人物領域におけるx座標の最小値、X1は人物領域におけるx座標の最大値である。
 次に部位分離位置抽出処理部101は、画像特徴量からその変化量(特徴変化量)を算出する。すなわち、部位分離位置抽出処理部101は、鉛直方向へ射影した画像特徴量の特徴変化量を算出する(ステップ1004)。例えば、部位分離位置抽出処理部101は、水平方向軸に平行な複数の軸と、人物領域とが交わった各々の線分について算出される画像特徴量の値の、鉛直方向軸上の位置の変化に対する特徴変化量を算出する。特徴変化量は、例えば(式2)のように求められる。Bは、特徴変化量を求めるために用いる画像特徴量の範囲の上限である。
Figure JPOXMLDOC01-appb-I000002
 部位分離位置抽出処理部101は、算出した特徴変化量と、あらかじめ特徴変化モデル蓄積部102に蓄積しておいた特徴変化モデルに含まれる各々の特徴変化量との類似度を算出し、類似する特徴変化量をもつ特徴変化モデルを抽出し、当該モデルの特徴変化量のピーク位置(部位分離位置)を出力する(ステップ1005)。例えば、部位分離位置抽出処理部101は、類似度が一番高い特徴変化量の部位分離位置を出力する。
 部位分離位置抽出処理部101は、類似度の算出方法として、例えば、(式2)のD(y)が閾値以上のピークとなっている値を抽出し、そのピークの位置を部位分離位置の候補(部位分離候補位置)とする。そして、部位分離位置抽出処理部101は、特徴変化モデル蓄積部102に蓄積した部位分離位置と、抽出した部位分離候補位置とを基に、各部位分離位置の数や部位分離位置間の差分距離を算出し、あるいは、DPマッチングのような手法を用いる等して、特徴変化量の類似度を算出することができる。
 また、部位分離位置抽出処理部101は、(式2)の特徴変化量と、特徴変化モデル蓄積部102に蓄積してある特徴変化モデルの特徴変化量との積を求め、その積の分布に含まれる閾値以上のピークの数とピークの高さを検出し、そのピークの数とピークの高さを基に特徴変化量の類似度を判定しても良い。部位分離位置抽出処理部101は、その他の方法を用いて特徴変化量の類似度を判定しても良い。
 本実施の形態によれば、入力された画像上の人物が身につけている服装に対して画像特徴量の変化量が類似する特徴変化モデルのピーク位置を出力することによって、服装の部位を分離して抽出することができる。入力された画像から抽出された特徴変化量やそのピーク位置は、画像のノイズなどが原因となって、部位の分離位置を正確に示しているとは限らない場合がある。そのような場合であっても、本実施の形態によれば、部位分離位置抽出処理部101が、既に部位分離位置が既知である特徴変化モデルを利用して、入力された画像から得られた特徴変化量と類似する特徴変化量を特定し、その特徴変化量の部位分離位置を出力するため、正確な部位分離位置を出力することができる。
 また、このようにして出力された部位分離位置は、例えば、図13が示すように、部位分離位置出力処理部2から出力され、服装領域特徴量を抽出する服装領域特徴抽出処理部3に入力されてもよい。本実施の形態によって部位分離位置が抽出されることで、服装領域特徴抽出処理部3は、服装領域特徴抽出処理部3に入力される画像情報、人物の領域を示す人物領域情報、部位分離位置を用いて各部位ごとの服装領域を求め、それらの服装領域に含まれる画素情報を用いて、服装領域の特徴量を抽出することができる。服装領域の特徴量としては、例えば、服装領域から色・エッジ(柄)を示すヒストグラムでもよいし、その他の色・柄を示す特徴量を用いてもよい。そしてこの特徴量によって、特定人物の捜索や、服装領域を用いた人物の検索などが可能となる。
 また、本実施の形態は人物の服装における部位分離位置の抽出について説明したが、本実施の形態によれば、人物や服装に限られず、動物や物体などについても上述した方法と同様に部位分離位置を抽出することが可能である。
 (第2の実施の形態)
 次に、本発明の第2の実施の形態について説明する。
 図3は、図1に示した部位分離位置出力処理部2の別の構成の一例を示すブロック図である。
 本実施の形態における部位分離位置出力処理部2は、部位分離位置抽出処理部107と、特徴変化モデル蓄積部102と、人物特定情報抽出・照合処理部301と、人物属性蓄積部201と、を含む。
 特徴変化モデル蓄積部102は、特徴変化量あるいは部位分離位置だけでなく、上述した人物属性(部位総称情報、部位性別情報、属性情報)と部位分離区間も蓄積している。すなわち、特徴変化モデルは、特徴変化モデルと人物属性とを対応付けて蓄積している。
 人物属性蓄積部201には、人物を特定する情報、例えば顔認識により人物の特定に用いられる人物個体特徴量と、人物属性(部位性別情報、部位総称情報、属性情報)とが対応付けられて蓄積されている。上述のとおり、部位総称情報とは、具体的な服装の名称(Tシャツ、スカートなど)や、特徴変化モデルに含まれる部位分離位置の確率分布を算出する際に類似する服装をまとめた部位総称情報の名称(上半身の服装(上半身1)など)や、服装全体を考慮した部位総称情報(スーツ類)などである。このように、人物属性蓄積部201には、特徴変化モデル蓄積部102が蓄積する部位総称情報と同一の粒度で名称が付けられている。
 人物特定情報抽出・照合処理部301は、入力された画像情報と人物領域情報とに基づいて、人物の顔領域を抽出し、既存の方法によって、その顔領域から顔照合に用いる人物個体特徴量を算出する。人物特定情報抽出・照合処理部301は、この人物個体特徴量と、人物属性蓄積部201に蓄積された人物個体特徴量とを比較し、最も類似する人物個体特徴量に対応する人物属性を人物属性蓄積部201から抽出し、部位分離位置抽出処理部107に出力する。
 なお、顔領域の抽出技術、顔認識技術は、様々な技術が提案されているが、人物特定情報抽出・照合処理部301が用いる技術は特段制限されない。また、人物個体特徴量は、顔領域から抽出したものに限らず、人物の体型や歩容特徴、虹彩情報、指紋情報、人物ごとに付与されたID情報などの特徴量であってもよい。また、人物特定情報抽出・照合処理部301には、性別、年齢、身長、人種、髪型などの判定機能を持たせてもよい。これにより、人物特定情報抽出・照合処理部301は、人物領域から抽出した性別、年齢、身長、人種、髪型などの情報に基づいて、人物個体特徴量を求めても良い。
 部位分離位置抽出処理部107は、第1の実施の形態で説明した図1の部位分離位置抽出処理部101の機能と同様の部位分離位置を抽出する機能をもつ。
 部位分離位置処理部107は、人物属性蓄積部201に蓄積された人物属性を人物特定情報抽出・照合処理部301から取得し、該人物属性と類似する人物属性を特徴変化モデル蓄積部102から抽出する。次に、部位分離位置処理部107は、抽出された人物属性に類似する人物属性と対応する特徴変化量を特徴変化モデル蓄積部102から抽出する。これにより、部位分離位置抽出処理部107は、特徴変化モデル蓄積部102に蓄積された特徴変化量の絞込みを行なうことができる。部位分離位置抽出処理部107は、図1の部位分離位置情報抽出部101に比べ、照合精度や処理速度が向上する。
 続いて第2の実施の形態の動作について図4を用いて詳細に説明する。
 ステップ1001の部位分離可能性判定処理からステップ1004の人物鉛直方向での特徴変化量の射影による特徴変化量の算出までは、図2と同一の処理が部位分離位置抽出処理部107で行われる。
 人物特定情報抽出・照合処理部301は、人物領域から抽出した人物個体特徴量と、人物属性蓄積部201に蓄積された人物個体特徴量との照合を行い、人物属性蓄積部201から、類似する人物個体特徴量に対応付けられた人物属性を抽出する。人物特定情報抽出・照合処理部301は、抽出した人物属性を部位分離位置抽出処理部107に出力する。
 部位分離位置抽出処理部107は、受け取った人物属性と、特徴変化モデル蓄積部102が蓄積する特徴変化モデルに含まれる人物属性とを照合し、類似する人物属性に対応付けられた特徴変化量を抽出する。すなわち、部位分離位置抽出処理部107は、特徴変化モデルを絞り込む(ステップ1006)。
 次に、部位分離位置抽出処理部107は、図4のステップ1005で、図2のステップ1005と同一の処理を行い、ステップ1006で抽出されて絞り込まれた特徴変化モデルの特徴変化量とステップ1004で得た特徴変化量との照合を行い、最も類似する特徴変化量の部位分離位置を出力する。
 本実施の形態によれば、第1の実施の形態に比べ、照合精度や処理速度が向上する。その理由は、部位分離位置抽出処理部107が、照合に用いる特徴変化モデルを絞り込むからである。
 (第3の実施の形態)
 図5は、図1の部位分離位置出力処理部2の別の構成の一例を示すブロック図である。
 本実施の形態における部位分離位置出力処理部2は、部位分離位置抽出処理部105、特徴変化モデル蓄積部102、水平方向特徴変化モデル蓄積部103を含んでいる。
 部位分離位置抽出処理部105は、鉛直方向特徴変化量照合処理部210、水平方向特徴変化量照合処理部211、部位分離位置統合処理部212を含んでいる。
 特徴変化モデル蓄積部102は、鉛直方向特徴変化量照合処理部210に接続されている。また、水平方向特徴変化モデル蓄積部103は水平方向特徴変化量照合処理部211に接続されている。部位分離位置統合処理部212は鉛直方向特徴変化量照合処理部210と水平方向特徴変化量照合処理部211と接続されている。特徴変化モデル蓄積部102は図1と同様である。
 水平方向特徴変化モデル蓄積部103には、特徴変化モデル蓄積部102と同様に、部位分離位置が既知の特徴変化量が蓄積されているが、その特徴変化量は人物の鉛直方向ではなく、水平方向に対して抽出したものである。つまり、水平方向特徴変化モデル蓄積部103は、人物領域の画像の画素値から人物の鉛直方向(y軸方向)にスキャンし、水平方向(x軸方向)の軸に射影した画素値から算出した特徴変化量(水平方向特徴変化量)を格納している。
 水平方向特徴変化モデル蓄積部103は、水平方向特徴変化量に加え、部位分離区間、人物属性(部位性別情報、部位総称情報、属性情報)などを、鉛直方向での特徴変化量算出時と同様に保存しても良い。その際、水平方向特徴変化モデル蓄積部103は、図14のようなリスト形式でそれらの情報を保存しても良い。
 水平方向特徴変化モデル蓄積部103に格納される部位分離区間の一例として、例えば、Tシャツにカーディガンを着て、ジーンズを履いている人物の画像(カーディガンの前は開いているとする)を例に説明する(図15参照)。この場合、Tシャツの生地を左右からカーディガンの生地が挟みこみ、またTシャツの下端にジーンズが位置する。この場合、一例として部位分離区間にはカーディガン、Tシャツ、ジーンズ、Tシャツ、カーディガンという部位総称情報が与えられている。ただし、水平方向の服装の長さが同程度の服装が多く存在するため、それらの水平方向特徴変化量は類似する可能性が高い。そのため、鉛直方向の特徴変化量について説明したのと同様に、部位分離位置出力処理部2は、それら類似する水平方向特徴変化量を持つ服装をまとめた部位総称情報を特徴変化モデルに付与してもよい。さらに、部位分離位置出力処理部2は、部位分離区間だけでなく、服装全体を考慮した部位総称情報を特徴変化モデルに付与してもよい。これらのような部位総称情報の付与の処理については、第1の実施の形態の鉛直方向における特徴変化モデルで説明した方法と同様である。
 また、第1の実施の形態と同様に、服装によっては、その服を身につけられる性別が定まっている場合もあるため、部位分離位置出力処理部2は、服装ごとに定まっている男性、女性、男女兼用などの部位性別情報を水平方向特徴変化モデルにも付与することができる。水平方向特徴変化モデル蓄積部103は、この部位性別情報を保存しても良い。部位分離位置出力処理部2は、例えば、ジーンズ、Tシャツ、カーディガンという服装情報から、男女兼用という部位性別情報を水平方向特徴変化モデル蓄積部103の特徴変化モデルに付与することができる。
 さらに、第1の実施の形態と同様に、部位分離位置出力処理部2は、年齢、身長、人種、髪型(スタイル、長さなど)などの属性情報も特徴変化モデルに付与しても良い。また、部位分離位置出力処理部2は、上述した情報以外の属性情報があれば、特徴変化モデルに付与しても良い。また、上述のように、水平方向特徴変化モデル蓄積部103は、部位性別情報、部位総称情報、属性情報、部位分離区間長、特徴変化量の対応関係を示すためのID、部位分離位置と服装の名称とを関連付けた情報などを保存しても良い。
 なお、上述の処理は、部位分離位置出力処理部2が行う処理として説明したが、予め外部装置等によって処理された結果が、水平方向特徴変化モデル蓄積部103に格納されても良い。
 鉛直方向特徴変化量照合処理部210は、図1を基に説明した、部位分離位置抽出処理部101と同様に、外部から、画像情報、人物領域情報、部位分離可否判定情報を取得する。そして、鉛直方向特徴変化量照合処理部210は、部位分離可否判定情報から部位分離が可能であると判定した場合に、入力された画像情報と人物領域情報とを基に特徴変化量を抽出する。そして、鉛直方向特徴変化量照合処理部210は、該特徴変化量と、特徴変化モデル蓄積部102に蓄積された特徴変化モデルの特徴変化量と、の類似度を算出し、最も類似度の高い鉛直方向の軸上の部位分離位置を出力する。
 水平方向特徴変化量照合処理部211は、鉛直方向特徴変化量照合処理部210と同様に、外部から、画像情報と人物領域情報、部位分離可否判定情報を取得する。水平方向特徴変化量照合処理部211は、部位分離可否判定情報から部位分離が可能であると判定した場合、部位分離位置を抽出する。ただし、水平方向特徴変化量照合処理部211は、人物の鉛直方向ではなく、人物の水平方向に対して部位分離位置を抽出する。具体的には、水平方向特徴変化量照合処理部211は、水平方向特徴変化量を算出し、算出した水平方向特徴変化量と、水平方向特徴変化モデル蓄積部103から取り出した水平方向特徴変化量との類似度の算出を行い、最も類似度の高い水平方向特徴変化量をもつ水平方向特徴変化モデルの部位分離位置を出力する。
 部位分離位置統合処理部212は、鉛直方向特徴変化量照合処理部210で抽出した鉛直方向の部位分離位置と、水平方向特徴変化量照合処理部211で抽出した水平方向の部位分離位置とを取得し、それらの情報を統合して、人物全体の部位分離位置を出力する。
 部位分離位置の統合処理として、部位分離位置統合処理部212は、各方向での部位分離位置を組み合わせ、人物領域を分割するための情報を生成する。例えば、部位分離位置統合処理部212は、鉛直方向の部位分離位置として保存された部位分離区間と部位総称情報と人物領域とから、頭部や上半身や下半身の領域の位置を抽出し、水平方向の部位分離位置を用いて、上半身の一部である腕領域(重ね着した際の服領域)を、人物の上半身領域から抽出することができる。
 本実施の形態によれば、部位分離位置統合処理部212は、水平・鉛直方向で各部位の領域の位置を決定するため、より詳細な部位分離位置を抽出することができる。また、部位分離位置統合処理部212は、水平・鉛直方向での部位分離位置を用いることで、人物領域の服装名称や部位性別情報の確度をあげることができる。例えば、各方向での部位分離位置に含まれる共通する部位総称は、その人物領域が着ている服装である可能性が高い。また、部位分離位置統合処理部212は、各方向での部位性別情報に含まれる性別が共通している場合は該性別の確度は高く、異なる場合は確度が低いといえる。このように部位分離位置統合処理部212は、方向ごとに服装名称や部位性別情報等を統合した部位分離情報を出力することが可能である。
 (第4の実施の形態)
 次に、本発明の第4の実施の形態について説明する。
 図6は、図5に示した部位分離位置抽出処理部105の別の構成の一例を示すブロック図である。
 人物領域から部位分離位置を抽出するために、第3の実施の形態(図5参照)においては、鉛直方向特徴変化量照合処理部210と水平方向特徴変化量照合処理部211とが、並列で特徴変化量照合処理を行っていた。一方、図6では、鉛直方向特徴変化量照合処理部210と水平方向特徴変化量照合処理213とが直列につながれている。直列でつながれる目的は、水平方向特徴変化量照合処理213が、鉛直方向特徴変化量照合処理部210で抽出された鉛直方向で分けられた部位ごとに、水平方向の部位分離位置を抽出するためである。人物の鉛直方向に対する特徴変化の発生回数は、人物の水平方向に対する特徴変化の発生回数に比べて多いため、水平方向の部位分離位置は、鉛直方向の部位分離位置よりも抽出が困難であるケースが発生しうる。例えば、頭、脚、前開きの服装などにおいて水平方向の部位分離位置の算出を行なう場合、水平方向の部位分離位置は同じ位置に集中しやすい。そのため、水平方向において頭、脚、前開きの服装などの部位分離位置を区別して抽出することは困難である。そこで、部位分離位置抽出処理部105は、図6の構成のようにまず鉛直方向の部位分離位置を特定し、鉛直方向の部位ごとに水平方向特徴変化量を抽出することで、より精度が高い部位分離位置が抽出できる。
 なお、この際、鉛直方向の部位について、全て水平方向特徴変化量を抽出する必要は無い。例えば、部位分離位置抽出処理部105は、人物の上半身が重ね着、下半身がズボンである場合等には水平方向特徴変化量抽出を行なっても良い。一方、部位分離位置抽出処理部105は、頭部分、靴部分など、それ以上分割する必要が無い領域については、水平方向特徴変化量を求めないようにしてもよい。
 本実施の形態における部位分離位置抽出処理部105は、鉛直方向特徴変化量照合処理部210、水平方向特徴変化量照合処理部213、部位分離位置統合処理部212を含んでいる。
 部位分離位置統合処理部212は、鉛直方向特徴変化量照合処理部210と、水平方向特徴変化量照合処理部213と接続されている。部位分離位置統合処理部212は、鉛直方向特徴変化量照合処理部210の鉛直方向の部位分離位置などの情報と、水平方向特徴変化量照合処理部213の水平方向の部位分離位置などの情報とを統合し、人物領域全体における部位分離位置などの情報を生成、出力する。
 水平方向特徴変化量照合処理部213は、鉛直方向特徴変化量照合処理部210と部位分離位置統合処理部212接続されている。鉛直方向特徴変化量照合処理部210には、外部から、画像情報、人物領域情報、鉛直方向の特徴変化モデルが入力される。また、水平方向特徴変化量照合処理部213には、鉛直方向特徴変化量照合処理部210で生成された水平方向の部位分離位置が入力される。
 水平方向特徴変化量照合処理部213は、鉛直方向の部位分離位置を基に、鉛直方向の部位分離区間ごとに水平方向特徴変化量を抽出し、この水平方向特徴変化量と、入力された水平方向の特徴変化モデルに含まれる水平方向特徴変化量と、を照合して、類似する特徴変化量を含む水平方向の特徴変化モデルを、水平方向特徴変化モデル蓄積部103から検索する。
 水平方向特徴変化量照合処理部213は、類似した特徴変化量を含む特徴変化モデルの水平方向の部位分離位置を部位分離位置統合処理部212に出力する。
 部位分離位置統合処理部212は、鉛直方向特徴変化量照合処理部210の鉛直方向の部位分離位置、水平方向特徴変化量照合処理部213の水平方向の部位分離位置を統合し、人物領域全体の部位分離情報を生成、出力する。
 本実施の形態によれば、まず人物の鉛直方向の部位分離位置を基に、人物の鉛直方向の部位を特定し、次に鉛直方向の部位ごとに水平方向特徴変化量を抽出することで、より精度が高い部位分離位置が抽出できる。
 (第5の実施の形態)
 図7は、第1の実施の形態の図1に示した部位分離位置抽出処理部101の別の構成の一例を示すブロック図である。
 本実施形態の部位分離位置抽出処理部101は、人物鉛直方向抽出処理部207と特徴変化量照合処理部206とを含み、特徴変化量照合処理部206は図示しない特徴変化モデル蓄積部102に接続されている。特徴変化モデル蓄積部102は図1と同じものである。
 人物鉛直方向抽出処理部207には、外部から、画像情報、人物領域情報が入力される。人物鉛直方向抽出処理部207は、入力された画像情報、人物領域情報から、その人物領域の鉛直方向を推定する。鉛直方向の推定方法には様々な方法がある。例えば、予め学習により生成した頭部検知器や足下検知器を利用して頭と足元を検知し、頭と足元の2点を結ぶベクトルから鉛直方向を推定する方法や、頭部または足元を表す画像上の二次元座標をカメラのキャリブレーション情報を用いて3次元空間座標に変換し、その点を通る鉛直方向に平行な直線上の点の3次元空間座標を求め、2次元に逆変換する方法などがある。具体的には、人物鉛直方向抽出処理部207は、人物領域の下端が足元である場合、撮影する空間の位置情報とカメラパラメタなどの情報を基に、画像座標を2次元から3次元またはその逆に射影する変換行列によって、2次元座標の足元位置を3次元座標の足元位置に変換し、その足元位置から適当な高さを設定した3次元座標を設定し、2次元に逆変換することで、2次元空間、つまり画像上での人物の鉛直方向を抽出することができる。これにより、レンズの周囲の領域などの、レンズのゆがみや画角によってy軸と鉛直方向が一致しない領域であっても人物の鉛直方向を自動で設定することができ、より精度の高い、部位分離位置を取得することができる。
 特徴変化量照合処理部206は、図1の部位分離位置抽出処理部101と同様に部位分離位置を抽出する機能を持つ。また、第1の実施の形態と同様に、特徴変化量照合処理部206は、人物鉛直方向抽出処理部207から取得した人物領域ごとに推定される鉛直方向の情報を基に第1の実施の形態と同様に人物領域の画像特徴量を算出する。
 本実施の形態によれば、人物鉛直方向抽出処理部207によって推定された人物領域の鉛直方向の情報を基に、該鉛直方向の画像特徴量を算出することができるため、手動でカメラ位置を設定せずとも鉛直方向の画像特徴量を算出できる。
 (第6の実施の形態)
 次に、本発明の第6の実施の形態について説明する。
 図8は、第1の実施の形態の図1に示した部位分離位置抽出処理部101の別の構成の一例を示すブロック図である。
 本実施の形態における部位分離位置抽出処理部101は、人物向き判定部209と特徴変化量照合処理部208とを含み、特徴変化量照合処理部208は図示しない特徴変化モデル蓄積部102に接続されている。また、人物向き判定手段209は特徴変化量照合処理部208に接続されている。
 図示しない特徴変化モデル蓄積部102には、人物の向きごとに、特徴変化モデルが蓄積されている。人物の向き(正面、背後、横向きなど)は、特定方向を基準に角度やベクトルなどで表現される。例えば人物がカメラに向かって正面を向いている方向を基準に座標系を定義し、角度やベクトルの方向で人物の向きが表現される。これらの角度やベクトル情報を、人物向き情報と定義する。特徴変化モデル蓄積部102には、人物向き情報で分類された特徴変化モデルが蓄積されている。
 人物向き判定部209は、画像情報と人物領域情報を基に人物が向いている方向を判定し、人物向き情報を出力する。人物の向き判定の方法は様々あるが、顔認識によって人物が向いている顔の方向を人物が向いている方向とする方法や、時系列で入力される画像を用いて、オプティカルフローによって進行方向から人物の向きを推定する方法や、抽出した人物領域の位置変化を算出して進行方向から人物の向きを推定する方法などがある。2番目と3番目の手法では、人物向き判定部209がカメラ座標系を知る必要があり、人物向き判定部209は、例えば、カメラの向きに対して人物が直線的にから遠ざかる場合は背面、逆に直線的に近づく場合は正面、横切る場合は横向きとして人物の向きを推定できる。
 特徴変化量照合処理部208は、図1の部位分離位置抽出処理部101と同様に部位分離位置を抽出する機能を持つ。また、特徴変化量照合処理部208は、入力された画像に対する特徴変化量と、特徴変化モデル蓄積部102に蓄積された特徴変化モデルの特徴変化量とを照合する前に、人物向き判定部209から得た人物向き情報と、特徴変化モデルを分類している人物向き情報とを照合する。照合方法として、特徴変化量照合処理部208は、例えば人物向き情報の角度やベクトル同士の類似度を比較する。特徴変化量照合処理部208は、例えば、人物向き情報が角度の場合には角度同士の差分によって類似度を比較しても良いし、ベクトルの場合には内積により類似度を比較しても良いし、その他の類似度の比較手法を用いても良い。
 その後、特徴変化量照合処理部208は、最も人物向き情報が類似する人物向き情報を選択し、当該人物向き情報に対応する特徴変化モデルを選択して特徴変化量の比較を行なう。これにより、特徴変化量照合処理部208は、同一ないし類似の人物向きの特徴変化量同士で照合を行うことができる。
 以上の構成によれば、まず人物の向きが類似している特徴変化モデルを抽出して絞り込みを行い、当該特徴変化モデルに対して特徴変化量同士の比較を行なうため、第1の実施形態の図1の構成よりも、服装の見え方の変動に対して高い精度で部位分離位置を取得することができる。
 (第7の実施の形態)
 図9は、第1の実施の形態の図1に示した部位分離位置抽出処理部101の別の構成の一例を示すブロック図である。
 本実施の形態における部位分離位置抽出処理部101は、部位分離位置蓄積部203と人物領域情報蓄積部204と特徴変化量照合処理部205とを含み、特徴変化量照合処理部205は、図示しない特徴変化モデル蓄積部102に接続されている。
 部位分離位置蓄積部203は、特徴変化量照合処理部205が算出した特徴変化量(あるいは部位分離位置)を、画像情報取得時のタイムスタンプ情報と共に蓄積する。また、部位分離位置蓄積部203は、既に蓄積された過去の特徴変化量(あるいは部位分離位置)とタイムスタンプ情報とを特徴変化量照合処理部205に出力する。
 人物領域情報蓄積部204は、画像情報取得時のタイムスタンプ情報と、人物領域情報とを蓄積する。
 また、人物領域情報蓄積部204は、既に蓄積された過去の人物領域情報をタイムスタンプ情報と共に特徴変化量照合処理部205に出力する。
 特徴変化量照合処理部205には、タイムスタンプ情報が付与された画像情報と、人物領域情報と、部位分離可否判定情報とが、外部から入力される。また、特徴変化量照合処理部205には、図示しない特徴変化モデル蓄積部102から特徴変化モデルが入力される。
 また、特徴変化量照合処理部205は、入力される画像情報のタイムスタンプ情報に近い時刻(タイムスタンプ情報が示す時刻と所定時間内にある時刻)、かつ、人物領域の位置やサイズが近い(誤差が所定範囲内の)過去の人物領域を照合するために、人物領域情報蓄積部204に接続されている。更に、特徴変化量照合処理部205は、タイムスタンプ情報に近い時刻の部位分離位置を取り出すために部位分離位置蓄積部203に接続されている。
 特徴変化量照合処理部205は、人物領域情報蓄積部204から、人物領域の位置やサイズが近く、タイムスタンプ情報に近い時刻の過去の人物領域の抽出が可能である場合、それらの人物領域は同一人物の人物領域とみなしても良い。これは、時間があまり経過していない画像に写った人物は急激にその位置やサイズが変化することがないと仮定できるためである。また、特徴変化量照合処理部205は、2つの人物領域が同一人物の人物領域か否かを判定する方法として、画像の輝度情報の変化を算出し、画素の移動量を推定するオプティカルフローを用いても良い。この場合、特徴変化量照合処理部205は、移動量が少ない人物領域を同一人物の人物領域とみなすことができる。また、特徴変化量照合処理部205は、画像情報から人物領域範囲の画素値を抽出し、その色や輝度の同一性を見ることで、同一人物の人物領域か否かを判定することもできる。ただし、この場合、人物領域情報蓄積部204には、人物領域情報の蓄積だけでなく、画像情報もタイムスタンプ情報と共に蓄積されている。
 特徴変化量照合処理部205は、人物領域情報蓄積部204から、同一人物の人物領域とみなされた人物領域を抽出し、その人物領域情報に付与されたタイムスタンプ情報に近い時刻のタイムスタンプ情報が付与された特徴変化量(あるいは部位分離位置)を、部位分離位置蓄積部203から取り出す。この特徴変化量(あるいは部位分離位置)は、同一人物の特徴変化量(あるいは部位分離位置)を表していると考えられる。
 特徴変化量照合処理部205は、この取り出した特徴変化量(あるいは部位分離位置)と、特徴変化量照合処理部205に現在入力されている画像情報と人物領域情報とから抽出した特徴変化量(あるいは第1の実施の形態と同様に算出した部位分離候補位置)との類似度を算出し、最も類似する部位分離位置を抽出することができる。
 一方、特徴変化量照合処理部205は、人物領域情報蓄積部204から、同一人物の人物領域とみなせる人物領域が存在しない場合、すなわち、人物領域情報蓄積部204に人物領域情報の位置やサイズが大きく異なる人物領域しか存在しない場合、特徴変化量照合処理部205は、第1の実施の形態と同様の処理で、部位分離位置を出力する。
 なお、部位分離位置蓄積部203に同一人物の部位分離位置が複数蓄積される場合、部位分離位置抽出処理部101は、その複数の部位分離位置の平均位置を用いて、あるいは、タイムスタンプが新しい部位分離位置ほど大きく重みを付ける重み平均位置を用いて、部位分離位置を作成することができる。また、部位分離位置抽出処理部101は、複数の画像から抽出した部位分離位置の頻度を示す確率分布(例えばガウス分布)を求めて部位分離位置を作成することができる。これにより、画像情報のノイズによって起こる部位分離位置を除去することができ、部位分離位置のマッチング精度が向上する。この場合、人物領域情報蓄積部204は部位分離位置に加え、人物を特定するためのIDを部位分離位置に関連付けて蓄積する。
 続いて本実施の形態の各部の動作について詳細に説明する。図10は、部位分離位置抽出処理部101の動作の一例を示すフローチャート図である。
 ステップ1001の部位分離可能性判定処理からステップ1004の人物鉛直方向での特徴変化量の射影による特徴変化量の算出までは、図2と同一の処理を特徴変化量照合処理部205で行う。
 次に、特徴変化量照合処理部205は、入力された画像に付与されたタイムスタンプと時刻が近いタイムスタンプが付与された人物領域を人物領域情報蓄積部204から取り出し、人物領域のサイズ、位置から同一人物の人物領域であるか否かを判定する(ステップ1007)。ステップ1007において同一人物の人物領域と判定された場合(ステップ1007、Yes)、特徴変化量照合処理部205は、部位分離位置蓄積部203から、タイムスタンプ情報に近い時刻のタイムスタンプ情報が付与された特徴変化量(あるいは部位分離位置)を取り出す。特徴変化量照合処理部205は、特徴変化量照合処理部205に入力されている画像情報と人物領域とを基に特徴変化量を求める(あるいは、第1の実施の形態のステップ1005で説明した手法で特徴変化量から部位分離候補位置を抽出する)。特徴変化量照合処理部205は、当該特徴変化量(あるいは部位分離候補位置)と、部位分離位置蓄積部203から抽出した特徴変化量(あるいは部位分離位置)との類似度を第1の実施の形態で説明した手法で算出し、部位分離位置を抽出する(ステップ1008)。
 一方、同一人物の人物領域と判定されなかった場合(ステップ1007、No)、特徴変化量照合処理部205は、第1の実施の形態と同様の処理で、部位分離位置を出力する。
 以上の構成によれば、過去の同一人物の人物領域を用いて部位分離位置を取得するため、高い精度で部位分離位置を取得することができる。
 (第8の実施の形態)
 本実施の形態における部位分離位置抽出装置1は、図16に示すように、特徴変化モデル蓄積部102と部位分離位置抽出処理部101とからなる。これらの構成、動作については上述した通りであるから詳細説明を省略する。
 特徴変化モデル蓄積部102は、複数の画像の人物領域ごとに、画像上で定義された所定軸(例えば水平軸、x軸)に平行な複数の軸と、各々の画像に含まれる人物領域と、が交わった各々の線分上で定義される画像特徴量の値の、所定軸に対する垂直軸(例えば鉛直軸、y軸)上の位置の変化に対する特徴変化量をモデルとして格納する。
 部位分離位置抽出処理部101は、外部から入力された画像上の人物領域に対して特徴変化量を算出し、当該特徴変化量と、特徴変化モデル蓄積部102に格納された特徴変化量との比較を行ない、類似する変化量を特定し、特定された変化量の値が所定値以上の箇所であるピーク位置を出力する。
 本実施の形態によれば、入力された画像上の人物が身につけている服装に対して画像特徴量の変化量が類似する特徴変化モデルのピーク位置を出力することによって、服装の部位を分離して抽出することができる。
 以上、実施形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されるものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解しうる様々な変更をすることができる。
 この出願は、2011年5月20日に出願された日本出願特願2011−113663を基礎とする優先権を主張し、その開示の全てをここに取り込む。
Hereinafter, embodiments of the present invention will be described with reference to the drawings. In all the drawings, the same components are denoted by the same reference numerals, and description thereof will be omitted as appropriate.
Each unit constituting the device of each embodiment includes a control unit, a memory, a program loaded in the memory, a storage unit such as a hard disk for storing the program, a network connection interface, and the like. Realized by a combination of And unless there is particular notice, the realization method and apparatus are not limited.
The control unit includes a CPU (Central Processing Unit) and the like, and controls the entire apparatus by operating an operating system, and reads programs and data from a recording medium mounted on a drive device to a memory, for example. Various processes are executed according to the above.
The recording medium is, for example, an optical disk, a flexible disk, a magnetic optical disk, an external hard disk, a semiconductor memory, or the like, and records a computer program so that the computer can read it. The computer program may be downloaded from an external computer (not shown) connected to the communication network. Here, unless otherwise specified, the communication network may be the Internet, a LAN (Local Area Network), a public line network, a wireless communication network, or a network configured by a combination thereof.
Each unit constituting the device of each embodiment is configured by a combination of hardware such as a logic circuit, software, and the like. The block diagram used in the description of each embodiment shows a functional unit configuration, not a hardware unit configuration. These functional blocks are realized by any combination of hardware and software. Further, in these drawings, the components of each embodiment may be described as being realized by one physically coupled device, but the means for realizing the same is not limited thereto. That is, the components, devices, systems, and the like described in each embodiment may be realized by a plurality of devices by connecting two or more physically separated devices in a wired or wireless manner. In addition, although there are cases where each component is described as two or more devices that are physically separated, the means for realizing it is not limited to this. That is, each component, device, system, and the like of each embodiment may be realized by arbitrarily combining hardware and software so as to be realized by one physically coupled device.
(First embodiment)
First, a first embodiment will be described with reference to FIG.
FIG. 1 is a block diagram illustrating an example of a site separation position extraction apparatus 1 according to the first embodiment of the present invention. The part separation position extraction apparatus 1 includes a part separation position output processing unit 2. The part separation position output processing unit 2 includes a part separation position extraction processing unit 101 and a feature change model storage unit 102, and the feature change model storage unit 102 is connected to the part separation position extraction processing unit 101.
As shown in FIG. 12, the feature change model accumulating unit 102 has an image feature in a predetermined axial direction defined on an image showing a person, for example, a vertical direction of the person (an axial direction connecting a head and a foot). Change amount (characteristic change amount) of the amount (described later) or the change amount and clothing and body part separation positions (the top of the head, the top of the T-shirt, the bottom of the T-shirt, the top of the skirt, the top of the skirt, etc.) And a feature change model in which a part separation position indicating a part of the body part) is associated. The image feature amount and feature change amount will be described later.
Next, a method in which the part separation position output processing unit 2 stores the feature change model in the feature change model storage unit 102 will be described. The part separation position output processing unit 2 extracts a person area indicating a person area from image information (also simply referred to as an image) input from the outside using an existing method, and determines the size of the person area in the vertical direction. The size of the person area is normalized by enlarging or reducing to a predetermined size. Further, the part separation position output processing unit 2 scans the human area in the horizontal direction, that is, pixel values on a line segment where the axis parallel to the horizontal axis intersects the human area (for example, the type of pixel color) Image feature quantity indicating the feature of the clothing at each position on the axis in the vertical direction of the person is calculated on the basis of the numerical value representing the intensity, brightness, brightness, etc. (see Equation 1 described later). Note that the vertical direction is an axial direction connecting the head and feet of a person in the image, and the part separation position output processing unit 2 may calculate from the image using a known method, or the vertical direction of the camera in advance. The camera is set so that the direction coincides with the vertical direction, and the part separation position output unit 2 may set the vertical direction of the image as the vertical direction.
Then, the part separation position output processing unit 2 calculates a feature change amount indicating a change in the image feature amount that occurs in the vertical direction of the person region, and associates the feature change amount with the type of clothes worn by the person. Information on clothes worn by persons on the image is given in advance.
The part separation position output processing unit 2 obtains each feature change amount for each clothing (see FIG. 12). The part separation position output processing unit 2 uses the peak position of each feature change amount as the part separation position, and uses the information associating the part separation position and the obtained change amount as the above feature change model. Store in the model storage unit 102.
By the way, when the feature change amounts are classified for each type of clothes, the feature change amounts of each classification may be similar. For example, there is a high possibility that the feature change amount obtained from clothes such as sweaters and jeans is similar to the feature change amount obtained from clothes such as Y-shirts and slacks. Therefore, the part separation position output processing unit 2 may combine similar feature change amounts into one. For example, the part separation position output processing unit 2 may combine feature change amounts using an average of each feature change amount when combining similar feature change amounts into one feature change amount. Are substantially the same, one representative feature change amount may be selected from similar feature change amounts based on a predetermined criterion. The part separation position output processing unit 2 may use another method for calculating a feature change amount that is representative of a plurality of feature change amounts.
The part separation position output processing unit 2 uses the degree of similarity calculated using the number of part separation positions and the difference distance between the part separation positions when collecting clothes having similar part separation positions, or DP (Dynamic Programming) matching. The degree of similarity calculated using a method such as the above can be used, but other methods for calculating the degree of similarity may be used.
In addition, the part separation position output processing unit 2 uses the pixel value obtained by scanning the image included in the person region in the horizontal direction as a method for calculating the image feature amount indicating the feature of the person's clothes. The method may be used. That is, the part separation position output processing unit 2 extracts a pixel value of a certain pixel included in the person area, and calculates an image feature amount based on the pixel value and the pixel values of pixels around the pixel. Then, a new image feature amount may be calculated by scanning the image feature amount in the horizontal direction. For example, the part separation position output processing unit 2 considers a plurality of pixels as one pixel instead of one pixel, and scans an average value of pixel values of the plurality of pixels in the horizontal direction to obtain an image feature amount. It may be calculated.
The part separation position output processing unit 2 assigns a clothing name and a body part name (part generic name information) to the section (part separation section) divided by the part separation position, and is characterized as a feature change model. The change model storage unit 102 may store the change model. For example, as shown in FIG. 12, a woman wearing a skirt, stockings, and shoes on a T-shirt has a site separation position at least between the top of the head, the top of the T-shirt, the bottom of the T-shirt, the top of the skirt, the bottom of the skirt, and the shoes. To do. Therefore, the part separation position output processing unit 2 refers to, for example, a storage unit (not shown) in which the section length and the part generic information are stored in association with each other, and each part separation section includes a face, a T-shirt, a skirt, You may give the part generic name information called a leg.
By the way, there are many clothes having the same length of clothes in the vertical direction even if they are different kinds of clothes. For this reason, when similar feature change amounts are combined into one, it is highly likely that feature change amounts of clothes of different types are combined and the same feature change model is calculated. Therefore, the part separation position output processing unit 2 may further add the name of part generic information that summarizes clothes having similar lengths when calculating the feature change amount included in the feature change model. For example, the part separation position output processing unit 2 may use T-shirts, jerseys, blousons, knits and the like as “short upper body clothes (upper body 1)”, and long coats, rain feathers, etc. as “long upper body”. Clothing (upper body 2) ”, mini skirts, shorts and half pants“ short lower body clothing (lower body 1) ”, lower body clothing such as seven-part trousers and long skirts that can be seen near the ankle, etc. “Lower body clothes (lower body 2)”, long pants that cannot see the ankles, long skirts, and the like may be classified as “long lower body clothes (lower body 3)”. Further, the part separation position output processing unit 2 may add not only the part separation section but also part generic name information about the entire clothing to the feature change model. For example, a combination of a jacket and slacks may be given part generic information “suits”.
By the way, depending on clothes, the gender that can wear the clothes may be determined. Therefore, the part separation position output processing unit 2 may add information (partial sex information) such as male, female, and unisex that is determined for each clothing to the feature change model. The part separation position output processing unit 2 may also store this part sex information in the feature change model storage unit 102. For example, since the person wearing the skirt is generally a woman, the part separation position output processing unit 2 refers to, for example, a storage unit (not shown) in which part generic information and part sex information are stored in association with each other. It is possible to give a woman as part sex information to the skirt.
The part separation position output processing unit 2 may further add attribute information such as a person's sex, age, height, race, hairstyle (such as length) to the feature change model together with the part sex information. For example, the attribute information includes “male / female” for gender, “length” and “style” for hairstyle, “age group” for age, and “skin color” for race. The part separation position output processing unit 2 can extract the attribute information from a person shown in the image using a known technique.
The part separation position output processing unit 2 may also store these attribute information in the feature change model storage unit 102. In addition to the information described above, the part separation position output processing unit 2 can add to the feature change model if there is other attribute information and can store it in the feature change model storage unit 102.
The accumulated part sex information, part generic name information, and attribute information can be used for narrowing down feature change models extracted from the feature change model storage unit 102 when matching feature change models to be described later. The part sex information, the part generic name information, and the attribute information are collectively referred to as a person attribute.
The part sex information, part generic information, attribute information, part separation section length (part separation section length), ID (IDentification) for indicating a correspondence relationship between feature change amounts, and the like are, for example, an association list as shown in FIG. Information may be stored in the feature change model storage unit 102 as information. The saved format may be another data format.
The feature change model storage unit 102 may store the feature change model for each direction of the person. The part separation position output processing unit 2 detects, for example, that a person is facing in the horizontal direction or the rear direction by a known technique, calculates a feature change amount and separation position information for each person's direction, and models each direction. May be stored in the feature change model storage unit 102. For example, the feature change model accumulating unit 102 stores feature change models constructed for the front direction, the horizontal direction, and the rear direction for suits that look different from the front direction, the horizontal direction, and the rear direction. You may accumulate according to direction.
Note that the above-described process has been described as the process in which the part separation position output processing unit 2 stores the feature change model in the feature change model storage unit 102, but is not limited thereto. For example, a feature change model generated in advance by an external device (not shown) or the like may be stored in the feature change model storage unit 102.
As described above, the feature change model storage unit 102 has the feature change amount of the image feature amount or the feature change amount and the part separation position at each position on the vertical axis of the person as shown in FIG. The associated feature change model is accumulated. In addition, as described above, the feature change model storage unit 102 may further store person attributes (partial sex information, part generic name information, attribute information) and part separation section length. Further, the feature change model storage unit 102 may store a probability distribution indicating the frequency of occurrence of a location where the feature change amount of the image feature amount on the vertical axis of the person is equal to or greater than a threshold value.
The part separation position extraction processing unit 101 determines whether the image information, the person area information (information indicating the position of the area in which the person is shown in the image information), and the clothes of the person area can be separated. Using the part separation possibility determination information to be determined, the feature change amount of the person region is calculated as shown in the right graph of FIG. 12, and the feature change amount and the feature change accumulated in the feature change model storage unit 102 are calculated. The similarity with the feature change amount of the model is calculated. The part separation position extraction processing unit 101 can calculate the degree of similarity using a method such as DP (Dynamic Programming) matching, using the number of part separation positions and the difference distance between the part separation positions. Other similarity calculation methods may be used.
The part separation possibility determination information is information indicating whether or not a person can be separated according to how the person in the image is captured. In the part separation possibility determination information, a value that determines whether separation is possible is described. For example, the part separation possibility determination information is a binary value indicating whether or not it is possible, or a numerical value for performing threshold processing. The part separation availability information is calculated by an external device (not shown). For example, the external device can determine how to capture a person showing that the person area is shown in the whole body in the image or the lower body is partially hidden, and can determine whether or not the part can be separated based on the way the image is taken. A specific example of the part separation availability information will be described with reference to FIG. The left figure of FIG. 11 is an input image, and shows a situation where only three upper persons can be seen because there are three persons and one person stands at the back of the desk. In this case, the external device can calculate a region as shown in the right diagram of FIG. 11 by a known method based on, for example, the position of the person's feet. In the right figure of FIG. 11, the upper end of the desk is an area where only the upper body is shown (black area), the floor is an area where the whole body is shown (shaded area), and the wall is an area where a person's feet cannot exist (white area) ). As described above, the part separation possibility determination information is “Yes” if there is a region in the image where there is a possibility of part separation, and “No” if there is no region.
The part separation position extraction processing unit 101 outputs the part separation position of the input person region based on the part separation position of the feature change model having the highest feature change amount. At this time, the part separation position extraction processing unit 101 can also receive and output information such as part generic name information, part sex information, attribute information, and part separation section length from the feature change model storage unit 102. For example, the part separation position extraction processing unit 101, when the feature change amount of clothes worn by a person in the input image is similar to a plurality of feature change amounts stored in the feature change model storage unit 102, Site generic information can be output. For example, since it is difficult to distinguish between a person wearing a T-shirt and a person wearing a knit or a trainer by a part separation position, the part separation position extraction processing unit 101 collectively refers to the upper body as “short upper body” The part generic information such as “clothing (upper body 1)” can be output.
Next, the operation of the first embodiment will be described in detail with reference to FIG.
The part separation position extraction processing unit 101 determines whether or not part separation is possible from the person region on the image based on the part separation possibility determination information (step 1001). The part separation position extraction processing unit 101 determines whether or not separation is possible based on the part separation possibility determination information.
When the part separation possibility determination is possible (step 1001, yes), the part separation position extraction processing unit 101 normalizes the person region (step 1002). The reason for performing the normalization is to suppress the variation of the feature change amount due to the change in the person area size. As a specific normalization method, the part separation position extraction processing unit 101 performs an enlargement or reduction process of an image while maintaining a vertical / horizontal aspect ratio so that the lengths of the human regions in the vertical direction are aligned with the reference value. The enlargement / reduction processing method may be a general image enlargement / reduction algorithm, such as the nearest neighbor method, the bilinear method, the bicubic method, or the like.
After normalization, the part separation position extraction processing unit 101 calculates an image feature amount in the horizontal direction of the person region. The part separation position extraction processing unit 101 scans pixel values of an image in the range indicated by the human region information in the horizontal direction (x-axis direction) and uses a function such as a pixel value projected in the vertical direction (y-axis direction). It is obtained and used as an image feature amount (step 1003). Specifically, the part separation position extraction processing unit 101 obtains an image feature amount on a line segment where the person region and the x axis intersect, for example, as in (Equation 1).
Figure JPOXMLDOC01-appb-I000001
Here, I (x, y) is a pixel value at coordinates (x, y) on the input image (for example, a value calculated from a three-dimensional vector of R (Red), G (Green), and B (Blue)). , M (x, y) is mask information generated from the person area information, X0 is the minimum value of the x coordinate in the person area, and X1 is the maximum value of the x coordinate in the person area.
Next, the part separation position extraction processing unit 101 calculates the change amount (feature change amount) from the image feature amount. That is, the part separation position extraction processing unit 101 calculates the feature change amount of the image feature amount projected in the vertical direction (step 1004). For example, the part separation position extraction processing unit 101 determines the position of the image feature value calculated for each line segment where a plurality of axes parallel to the horizontal axis intersect with the person area on the vertical axis. A feature change amount with respect to the change is calculated. The feature change amount is obtained as in (Equation 2), for example. B is an upper limit of the range of the image feature amount used for obtaining the feature change amount.
Figure JPOXMLDOC01-appb-I000002
The part separation position extraction processing unit 101 calculates the similarity between the calculated feature change amount and each feature change amount included in the feature change model stored in the feature change model storage unit 102 in advance. A feature change model having a feature change amount is extracted, and a peak position (part separation position) of the feature change amount of the model is output (step 1005). For example, the part separation position extraction processing unit 101 outputs the part separation position of the feature change amount having the highest similarity.
The part separation position extraction processing unit 101 extracts, for example, a value in which D (y) in (Equation 2) is a peak equal to or greater than a threshold as a method for calculating the similarity, and the position of the peak is determined as the part separation position. Let it be a candidate (part separation candidate position). The part separation position extraction processing unit 101 then determines the number of part separation positions and the difference distance between the part separation positions based on the part separation positions accumulated in the feature change model accumulation unit 102 and the extracted part separation candidate positions. Or the similarity of the feature change amount can be calculated by using a technique such as DP matching.
Further, the part separation position extraction processing unit 101 obtains a product of the feature change amount of (Equation 2) and the feature change amount of the feature change model stored in the feature change model storage unit 102, and uses the product distribution. The number of peaks that are greater than or equal to the threshold value and the height of the peak may be detected, and the similarity of the feature change amount may be determined based on the number of peaks and the peak height. The part separation position extraction processing unit 101 may determine the similarity of the feature change amount using other methods.
According to the present embodiment, by outputting the peak position of a feature change model in which the change amount of the image feature quantity is similar to the clothes worn by the person on the input image, the parts of the clothes are separated. And can be extracted. The feature change amount and its peak position extracted from the input image may not always accurately indicate the separation position of the part due to image noise or the like. Even in such a case, according to the present embodiment, the part separation position extraction processing unit 101 is obtained from the input image using the feature change model whose part separation position is already known. Since the feature change amount similar to the feature change amount is specified and the part separation position of the feature change amount is output, an accurate part separation position can be output.
Further, the part separation position output in this way is output from the part separation position output processing unit 2 and input to the clothing region feature extraction processing unit 3 that extracts the clothing region feature amount, for example, as shown in FIG. May be. By extracting the part separation position according to the present embodiment, the clothing region feature extraction processing unit 3 causes the image information input to the clothing region feature extraction processing unit 3, person region information indicating a person region, and the part separation position. It is possible to obtain a clothing region for each part using, and to extract a feature amount of the clothing region using pixel information included in the clothing region. As the feature amount of the clothing region, for example, a histogram indicating colors / edges (patterns) from the clothing region may be used, or feature amounts indicating other colors / patterns may be used. The feature amount enables searching for a specific person, searching for a person using a clothing area, and the like.
Further, although the present embodiment has been described with respect to the extraction of the part separation position in the clothes of the person, according to the present embodiment, the part separation is not limited to the person or the clothes, but also to animals and objects in the same manner as described above. It is possible to extract the position.
(Second Embodiment)
Next, a second embodiment of the present invention will be described.
FIG. 3 is a block diagram showing an example of another configuration of the part separation position output processing unit 2 shown in FIG.
The part separation position output processing unit 2 in the present embodiment includes a part separation position extraction processing unit 107, a feature change model storage unit 102, a person specifying information extraction / collation processing unit 301, and a person attribute storage unit 201. Including.
The feature change model accumulation unit 102 accumulates not only the feature change amount or the part separation position but also the above-described personal attributes (part generic name information, part sex information, attribute information) and the part separation section. That is, the feature change model stores the feature change model and the person attribute in association with each other.
In the person attribute storage unit 201, information for specifying a person, for example, a person individual feature amount used for specifying a person by face recognition, and a person attribute (part sex information, part generic information, attribute information) are associated with each other. Accumulated. As described above, the part generic information is a part generic name that summarizes similar clothes when calculating a probability distribution of a part separation position included in a feature change model (T-shirt, skirt, etc.) and a feature change model. Information name (upper body clothes (upper body 1), etc.), part generic information (suits) considering the entire clothes. As described above, the person attribute storage unit 201 is given a name with the same granularity as the part generic name information stored in the feature change model storage unit 102.
The person identification information extraction / collation processing unit 301 extracts a person's face area based on the input image information and person area information, and uses the existing method to identify the individual individual feature amount used for face collation from the face area. Is calculated. The person specifying information extraction / collation processing unit 301 compares this individual individual feature amount with the individual individual feature amount stored in the individual attribute storage unit 201, and determines the person attribute corresponding to the most similar individual individual feature amount as the person. Extracted from the attribute storage unit 201 and output to the part separation position extraction processing unit 107.
Various techniques have been proposed for the face area extraction technique and the face recognition technique, but the technique used by the person specifying information extraction / collation processing unit 301 is not particularly limited. The individual individual feature amount is not limited to the feature amount extracted from the face region, and may be a feature amount such as a person's body shape, gait feature, iris information, fingerprint information, ID information assigned to each person. Further, the person identification information extraction / collation processing unit 301 may have a determination function such as gender, age, height, race, hairstyle, and the like. Accordingly, the person specifying information extraction / collation processing unit 301 may obtain the individual individual feature amount based on information such as sex, age, height, race, and hairstyle extracted from the person region.
The part separation position extraction processing unit 107 has a function of extracting a part separation position similar to the function of the part separation position extraction processing unit 101 of FIG. 1 described in the first embodiment.
The part separation position processing unit 107 acquires the person attribute stored in the person attribute storage unit 201 from the person specifying information extraction / collation processing unit 301 and extracts a person attribute similar to the person attribute from the feature change model storage unit 102. To do. Next, the part separation position processing unit 107 extracts a feature change amount corresponding to a person attribute similar to the extracted person attribute from the feature change model storage unit 102. Accordingly, the part separation position extraction processing unit 107 can narrow down the feature change amount stored in the feature change model storage unit 102. The part separation position extraction processing unit 107 improves collation accuracy and processing speed compared to the part separation position information extraction unit 101 of FIG.
Next, the operation of the second embodiment will be described in detail with reference to FIG.
From the part separation possibility determination process in step 1001 to the calculation of the feature change amount by projection of the feature change amount in the person vertical direction in step 1004, the part separation position extraction processing unit 107 performs the same process as in FIG.
The person identification information extraction / collation processing unit 301 collates the individual individual feature amount extracted from the person region with the individual individual feature amount accumulated in the person attribute accumulation unit 201, and is similar to the person attribute accumulation unit 201. A person attribute associated with a person individual feature amount is extracted. The person specifying information extraction / collation processing unit 301 outputs the extracted person attribute to the part separation position extraction processing unit 107.
The part separation position extraction processing unit 107 collates the received person attribute with the person attribute included in the feature change model stored in the feature change model storage unit 102, and calculates the feature change amount associated with the similar person attribute. Extract. That is, the part separation position extraction processing unit 107 narrows down the feature change model (step 1006).
Next, the part separation position extraction processing unit 107 performs the same processing as step 1005 in FIG. 2 at step 1005 in FIG. 4 and the feature change amount of the feature change model extracted and narrowed down at step 1006 and step 1004. Is compared with the feature change amount obtained in step 1, and the part separation position of the most similar feature change amount is output.
According to the present embodiment, the collation accuracy and the processing speed are improved as compared with the first embodiment. This is because the part separation position extraction processing unit 107 narrows down the feature change model used for collation.
(Third embodiment)
FIG. 5 is a block diagram illustrating an example of another configuration of the part separation position output processing unit 2 of FIG.
The part separation position output processing unit 2 in the present embodiment includes a part separation position extraction processing unit 105, a feature change model storage unit 102, and a horizontal direction feature change model storage unit 103.
The site separation position extraction processing unit 105 includes a vertical direction feature change amount matching processing unit 210, a horizontal direction feature change amount matching processing unit 211, and a site separation position integration processing unit 212.
The feature change model storage unit 102 is connected to the vertical feature change amount matching processing unit 210. The horizontal feature change model storage unit 103 is connected to the horizontal feature change amount matching processing unit 211. The part separation position integration processing unit 212 is connected to the vertical direction feature change amount matching processing unit 210 and the horizontal direction feature change amount matching processing unit 211. The feature change model storage unit 102 is the same as that in FIG.
Like the feature change model storage unit 102, the horizontal feature change model storage unit 103 stores feature change amounts with known part separation positions. However, the feature change amounts are not the vertical direction of the person but the horizontal direction. Extracted with respect to direction. In other words, the horizontal direction feature change model accumulation unit 103 scans from the pixel value of the image of the person area in the vertical direction (y-axis direction) of the person, and calculates from the pixel value projected onto the axis in the horizontal direction (x-axis direction). The feature change amount (horizontal feature change amount) is stored.
The horizontal direction feature change model accumulating unit 103 includes, in addition to the horizontal direction feature change amount, a part separation section, a person attribute (part sex information, part generic information, attribute information), etc., in the same manner as the feature change amount calculation in the vertical direction. You may save it. At that time, the horizontal direction feature change model accumulating unit 103 may store such information in a list format as shown in FIG.
As an example of the part separation section stored in the horizontal direction feature change model storage unit 103, for example, an image of a person wearing a cardigan on a T-shirt and wearing jeans (assuming that the cardigan is open before the cardigan) (See FIG. 15). In this case, the cardigan fabric is sandwiched between the T-shirt fabric from the left and right, and the jeans are located at the lower end of the T-shirt. In this case, as an example, part generic information such as cardigan, T-shirt, jeans, T-shirt, and cardigan is given to the part separation section. However, since there are many clothes having the same length in the horizontal direction, there is a high possibility that the horizontal feature change amounts are similar. Therefore, similarly to the description of the feature change amount in the vertical direction, the part separation position output processing unit 2 assigns part generic information including clothes having similar horizontal feature change amounts to the feature change model. Also good. Furthermore, the part separation position output processing unit 2 may add not only the part separation section but also part generic name information considering the entire clothing to the feature change model. The process of giving the part generic information as described above is the same as the method described in the feature change model in the vertical direction of the first embodiment.
Similarly to the first embodiment, depending on the clothes, the gender on which the clothes can be worn may be determined. Therefore, the site separation position output processing unit 2 is determined by men and women determined for each clothes. In addition, it is also possible to assign part sex information such as unisex to the horizontal feature change model. The horizontal direction feature change model accumulating unit 103 may store this part sex information. The part separation position output processing unit 2 can add part sex information of unisex to the feature change model of the horizontal direction feature change model storage unit 103 from clothes information such as jeans, T-shirts, and cardigans.
Furthermore, as in the first embodiment, the part separation position output processing unit 2 may also add attribute information such as age, height, race, hairstyle (style, length, etc.) to the feature change model. . In addition, the part separation position output processing unit 2 may add to the feature change model if there is attribute information other than the information described above. In addition, as described above, the horizontal direction feature change model storage unit 103 includes the part sex information, the part generic information, the attribute information, the part separation section length, the ID for indicating the correspondence between the feature change amounts, the part separation position, and the clothes. You may save the information etc. which linked | related with the name of.
Although the above-described processing has been described as processing performed by the part separation position output processing unit 2, a result processed in advance by an external device or the like may be stored in the horizontal direction feature change model accumulation unit 103.
Similar to the part separation position extraction processing unit 101 described based on FIG. 1, the vertical feature change amount matching processing unit 210 obtains image information, person area information, and part separation possibility determination information from the outside. Then, when it is determined that the part separation is possible from the part separation possibility determination information, the vertical direction feature change amount matching processing unit 210 extracts the feature change amount based on the input image information and person area information. . Then, the vertical feature change amount matching processing unit 210 calculates the similarity between the feature change amount and the feature change amount of the feature change model stored in the feature change model storage unit 102, and has the highest similarity. The part separation position on the vertical axis is output.
Similar to the vertical direction feature change amount matching processing unit 210, the horizontal direction feature change amount matching processing unit 211 acquires image information, person area information, and part separation possibility determination information from the outside. When it is determined that the part separation is possible from the part separation possibility determination information, the horizontal direction feature change amount matching processing unit 211 extracts the part separation position. However, the horizontal feature change amount matching processing unit 211 extracts a part separation position in the horizontal direction of the person, not in the vertical direction of the person. Specifically, the horizontal direction feature change amount matching processing unit 211 calculates the horizontal direction feature change amount, and the calculated horizontal direction feature change amount and the horizontal direction feature change amount extracted from the horizontal direction feature change model storage unit 103. And the part separation position of the horizontal feature change model having the highest horizontal feature change amount is output.
The part separation position integration processing unit 212 acquires the vertical part separation position extracted by the vertical feature change amount matching processing unit 210 and the horizontal part separation position extracted by the horizontal direction feature change amount matching processing unit 211. Then, those pieces of information are integrated to output the part separation position of the entire person.
As the part separation position integration processing, the part separation position integration processing unit 212 generates information for dividing the person region by combining the part separation positions in each direction. For example, the part separation position integration processing unit 212 extracts the positions of the head, upper body, and lower body areas from the part separation section, the part generic information, and the person area that are stored as the part separation positions in the vertical direction, and the horizontal direction Using this part separation position, an arm region (a clothing region when layered) that is a part of the upper body can be extracted from the upper body region of the person.
According to the present embodiment, the part separation position integration processing unit 212 determines the position of each part region in the horizontal and vertical directions, and thus can extract more detailed part separation positions. Further, the part separation position integration processing unit 212 can increase the accuracy of the clothes name and part sex information of the person region by using the part separation positions in the horizontal and vertical directions. For example, the common part generic name included in the part separation positions in each direction is highly likely to be clothes worn by the person region. In addition, the part separation position integration processing unit 212 can be said to have high accuracy when the gender included in the part gender information in each direction is common, and low accuracy when the gender is different. As described above, the part separation position integration processing unit 212 can output part separation information in which clothing names, part sex information, and the like are integrated for each direction.
(Fourth embodiment)
Next, a fourth embodiment of the present invention will be described.
FIG. 6 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 105 illustrated in FIG.
In order to extract the part separation position from the person region, in the third embodiment (see FIG. 5), the vertical direction feature change amount matching processing unit 210 and the horizontal direction feature change amount matching processing unit 211 are arranged in parallel. A feature change amount matching process was performed. On the other hand, in FIG. 6, the vertical direction feature change matching processing unit 210 and the horizontal direction feature change matching processing 213 are connected in series. The purpose of connecting in series is that the horizontal feature change amount matching process 213 extracts a horizontal part separation position for each part divided in the vertical direction extracted by the vertical direction feature change amount matching processing unit 210. It is. Cases where the number of occurrences of feature changes in the vertical direction of a person is greater than the number of occurrences of feature changes in the horizontal direction of the person, so that the part separation position in the horizontal direction is more difficult to extract than the part separation position in the vertical direction Can occur. For example, when calculating the horizontal part separation position in the head, legs, front-opening clothes, etc., the horizontal part separation position tends to concentrate on the same position. Therefore, it is difficult to distinguish and extract part separation positions such as the head, legs, and front-opening clothes in the horizontal direction. Therefore, the part separation position extraction processing unit 105 first specifies the part separation position in the vertical direction as in the configuration of FIG. 6, and extracts the horizontal feature change amount for each part in the vertical direction, thereby achieving higher accuracy. The site separation position can be extracted.
At this time, it is not necessary to extract the horizontal direction feature change amount for all the parts in the vertical direction. For example, the part separation position extraction processing unit 105 may perform the horizontal feature change amount extraction when the upper body of the person is layered and the lower body is pants. On the other hand, the part separation position extraction processing unit 105 may not calculate the horizontal feature change amount for an area that does not need to be further divided, such as a head part or a shoe part.
The site separation position extraction processing unit 105 in the present embodiment includes a vertical direction feature change amount matching processing unit 210, a horizontal direction feature change amount matching processing unit 213, and a site separation position integration processing unit 212.
The part separation position integration processing unit 212 is connected to the vertical direction feature variation matching processing unit 210 and the horizontal direction feature variation matching processing unit 213. The part separation position integration processing unit 212 includes information such as the vertical part separation position of the vertical feature change amount matching processing unit 210 and information such as the horizontal part separation position of the horizontal feature change amount matching processing unit 213. To generate and output information such as part separation positions in the entire person area.
The horizontal direction feature amount matching processing unit 213 is connected to the vertical direction feature amount matching processing unit 210 and the part separation position integration processing unit 212. Image information, person region information, and a vertical feature change model are input to the vertical feature change amount matching processing unit 210 from the outside. Also, the horizontal part separation position generated by the vertical direction feature variation matching processor 210 is input to the horizontal feature variation matching processor 213.
The horizontal feature change amount matching processing unit 213 extracts a horizontal feature change amount for each vertical part separation section based on the vertical part separation position, and the horizontal feature change amount and the input horizontal feature change amount are extracted. The horizontal feature change model included in the direction feature change model is collated, and a horizontal feature change model including a similar feature change amount is searched from the horizontal feature change model storage unit 103.
The horizontal feature change amount matching processing unit 213 outputs the horizontal part separation position of the feature change model including the similar feature change amount to the part separation position integration processing unit 212.
The part separation position integration processing unit 212 integrates the vertical part separation position of the vertical direction feature change amount matching processing unit 210 and the horizontal part separation position of the horizontal direction feature change amount matching processing unit 213, and Generate and output site separation information.
According to the present embodiment, first, based on the vertical part separation position of the person, specify the vertical part of the person, and then extract the horizontal feature change amount for each vertical part, A site separation position with higher accuracy can be extracted.
(Fifth embodiment)
FIG. 7 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
The part separation position extraction processing unit 101 of this embodiment includes a person vertical direction extraction processing unit 207 and a feature change amount matching processing unit 206, and the feature change amount matching processing unit 206 is connected to a feature change model storage unit 102 (not shown). Has been. The feature change model storage unit 102 is the same as that shown in FIG.
The person vertical direction extraction processing unit 207 receives image information and person area information from the outside. The person vertical direction extraction processing unit 207 estimates the vertical direction of the person area from the input image information and person area information. There are various methods for estimating the vertical direction. For example, a head detector or a foot detector generated by learning in advance is used to detect the head and feet, and a method for estimating the vertical direction from a vector connecting the two points of the head and feet, or represents the head or feet 2D coordinates on the image are converted into 3D space coordinates using the camera calibration information, and 3D space coordinates of a point on a straight line passing through the point and parallel to the vertical direction are obtained and converted back to 2D. There are methods. Specifically, when the lower end of the person area is a foot, the person vertical direction extraction processing unit 207 converts the image coordinates from two-dimensional to three-dimensional based on the position information of the space to be photographed and information such as camera parameters. Convert the foot position of the two-dimensional coordinates to the foot position of the three-dimensional coordinates by using a transformation matrix that projects in reverse, and set the three-dimensional coordinates set with an appropriate height from the foot positions and reversely transform to two dimensions. Thus, the vertical direction of the person on the two-dimensional space, that is, the image can be extracted. As a result, the vertical direction of the person can be automatically set even in a region where the y-axis and the vertical direction do not coincide with each other, such as a region around the lens, due to lens distortion or the angle of view. The separation position can be acquired.
The feature change amount matching processing unit 206 has a function of extracting a part separation position in the same manner as the part separation position extraction processing unit 101 in FIG. Similarly to the first embodiment, the feature change amount matching processing unit 206 performs the first embodiment based on the information in the vertical direction estimated for each person area acquired from the person vertical direction extraction processing unit 207. Similar to the form, the image feature amount of the person region is calculated.
According to the present embodiment, the image feature amount in the vertical direction can be calculated based on the information on the vertical direction of the person area estimated by the person vertical direction extraction processing unit 207, so that the camera position is manually set. The image feature amount in the vertical direction can be calculated without setting.
(Sixth embodiment)
Next, a sixth embodiment of the present invention will be described.
FIG. 8 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
The part separation position extraction processing unit 101 in this embodiment includes a person orientation determination unit 209 and a feature change amount matching processing unit 208, and the feature change amount matching processing unit 208 is connected to a feature change model storage unit 102 (not shown). ing. The person orientation determination unit 209 is connected to the feature change amount matching processing unit 208.
A feature change model storage unit 102 (not shown) stores feature change models for each direction of a person. The direction of a person (front, back, sideways, etc.) is expressed by an angle, a vector, or the like based on a specific direction. For example, a coordinate system is defined on the basis of the direction in which the person is facing the front toward the camera, and the direction of the person is expressed by an angle or a vector direction. These angles and vector information are defined as person orientation information. The feature change model storage unit 102 stores feature change models classified according to person orientation information.
The person orientation determination unit 209 determines the direction in which the person is facing based on the image information and the person area information, and outputs the person orientation information. There are various methods for determining the direction of a person, but the direction of the face that the person faces by face recognition is set to the direction that the person is facing, or the direction of travel by optical flow using images input in time series There are a method for estimating the direction of a person from the above, a method for calculating the position change of the extracted person region, and a method for estimating the direction of the person from the traveling direction. In the second and third methods, the person orientation determination unit 209 needs to know the camera coordinate system. For example, the person orientation determination unit 209 is configured to display the rear side when the person moves away from a straight line with respect to the camera orientation. On the other hand, the direction of the person can be estimated as a front direction when approaching linearly and as a landscape direction when crossing.
The feature change amount matching processing unit 208 has a function of extracting a part separation position in the same manner as the part separation position extraction processing unit 101 in FIG. In addition, the feature change amount matching processing unit 208 compares the feature change amount for the input image with the feature change amount of the feature change model stored in the feature change model storage unit 102 before the person orientation determination unit 209. The person orientation information obtained from the above and the person orientation information that classifies the feature change model are collated. As a matching method, the feature change amount matching processing unit 208 compares, for example, the angle of person orientation information and the similarity between vectors. For example, when the person orientation information is an angle, the feature change amount matching processing unit 208 may compare the similarity based on the difference between the angles, and may compare the similarity based on an inner product if the information is a vector. However, other methods for comparing similarities may be used.
Thereafter, the feature change amount matching processing unit 208 selects the person orientation information with the most similar person orientation information, selects the feature change model corresponding to the person orientation information, and compares the feature change amounts. Accordingly, the feature change amount matching processing unit 208 can perform matching between feature change amounts of the same or similar person orientation.
According to the above configuration, first, a feature change model with a similar person orientation is extracted and narrowed down, and feature change amounts are compared with each other in the feature change model. Compared to the configuration of FIG. 1, the site separation position can be acquired with higher accuracy with respect to fluctuations in the appearance of clothes.
(Seventh embodiment)
FIG. 9 is a block diagram illustrating an example of another configuration of the part separation position extraction processing unit 101 illustrated in FIG. 1 according to the first embodiment.
The part separation position extraction processing unit 101 in the present embodiment includes a part separation position accumulation unit 203, a person region information accumulation unit 204, and a feature change amount matching processing unit 205, and the feature change amount matching processing unit 205 is not illustrated. The feature change model storage unit 102 is connected.
The part separation position accumulating unit 203 accumulates the feature change amount (or part separation position) calculated by the feature change amount matching processing unit 205 together with time stamp information at the time of image information acquisition. The part separation position accumulating unit 203 outputs the past feature change amount (or part separation position) and time stamp information that have already been accumulated to the feature change amount collation processing unit 205.
The person area information accumulation unit 204 accumulates time stamp information at the time of image information acquisition and person area information.
In addition, the person area information storage unit 204 outputs the past person area information that has already been stored to the feature change amount matching processing unit 205 together with the time stamp information.
The feature change amount matching processing unit 205 receives externally input image information to which time stamp information is added, person region information, and part separation possibility determination information. Further, the feature change amount matching processing unit 205 receives a feature change model from a feature change model storage unit 102 (not shown).
In addition, the feature change amount matching processing unit 205 is close to the time stamp information of the input image information (the time indicated by the time stamp information and a time within a predetermined time), and the position and size of the person area are close ( In order to collate a past person area (with an error within a predetermined range), it is connected to the person area information storage unit 204. Further, the feature change amount matching processing unit 205 is connected to the part separation position accumulating part 203 in order to extract a part separation position at a time close to the time stamp information.
When the person region information storage unit 204 can extract past person regions at a time close to the time stamp information, the feature change amount matching processing unit 205 can extract the past person regions at a time close to the time stamp information. It may be regarded as a person area of the same person. This is because it can be assumed that a person who appears in an image that has not passed much time will not change its position or size abruptly. In addition, as a method for determining whether or not the two person regions are the same person, the feature change amount matching processing unit 205 calculates an optical flow that calculates the change in the luminance information of the image and estimates the movement amount of the pixel. It may be used. In this case, the feature change amount matching processing unit 205 can regard a person area with a small amount of movement as a person area of the same person. Further, the feature change amount matching processing unit 205 can also determine whether or not the person area is the same person by extracting the pixel value of the person area range from the image information and looking at the same color and luminance. . However, in this case, the person area information storage unit 204 stores not only the person area information but also the image information together with the time stamp information.
The feature change amount matching processing unit 205 extracts a person area regarded as the person area of the same person from the person area information storage unit 204, and time stamp information at a time close to the time stamp information given to the person area information. The amount of feature change (or part separation position) to which is added is extracted from the part separation position accumulation unit 203. This feature change amount (or part separation position) is considered to represent the feature change amount (or part separation position) of the same person.
The feature change amount matching processing unit 205 extracts the feature change amount extracted from the extracted feature change amount (or part separation position) and the image information and person area information currently input to the feature change amount matching processing unit 205. Alternatively, the similarity with the part separation candidate position calculated in the same manner as in the first embodiment can be calculated, and the most similar part separation position can be extracted.
On the other hand, when there is no person area that can be regarded as a person area of the same person from the person area information storage section 204, that is, the position and size of the person area information is stored in the person area information storage section 204. When only a significantly different person region exists, the feature change amount matching processing unit 205 outputs the part separation position by the same processing as in the first embodiment.
When a plurality of part separation positions of the same person are stored in the part separation position storage unit 203, the part separation position extraction processing unit 101 uses the average position of the plurality of part separation positions or the time stamp is new. A part separation position can be created using a weighted average position that gives a greater weight to the part separation position. In addition, the part separation position extraction processing unit 101 can create a part separation position by obtaining a probability distribution (for example, a Gaussian distribution) indicating the frequency of the part separation positions extracted from a plurality of images. Thereby, the part separation position caused by the noise of the image information can be removed, and the matching accuracy of the part separation position is improved. In this case, in addition to the part separation position, the person region information storage unit 204 stores an ID for specifying a person in association with the part separation position.
Next, the operation of each part of the present embodiment will be described in detail. FIG. 10 is a flowchart showing an example of the operation of the part separation position extraction processing unit 101.
From the part separability determination process in step 1001 to the calculation of the feature change amount by the projection of the feature change amount in the human vertical direction in step 1004, the same process as that in FIG.
Next, the feature change amount collation processing unit 205 extracts a person area to which a time stamp that is similar to the time stamp given to the input image is given from the person area information storage unit 204, and determines the size and position of the person area. From this, it is determined whether or not the same person area (step 1007). When it is determined in step 1007 that the person area is the same person (step 1007, Yes), the feature change amount matching processing unit 205 is given time stamp information at a time close to the time stamp information from the part separation position accumulating unit 203. The amount of feature change (or part separation position) is extracted. The feature change amount matching processing unit 205 obtains a feature change amount based on the image information input to the feature change amount matching processing unit 205 and the person area (or described in step 1005 of the first embodiment). The part separation candidate position is extracted from the feature change amount by the technique). The feature change amount matching processing unit 205 determines the similarity between the feature change amount (or part separation candidate position) and the feature change amount (or part separation position) extracted from the part separation position storage unit 203 in the first embodiment. Calculation is performed by the method described in the embodiment, and a part separation position is extracted (step 1008).
On the other hand, when it is not determined that the person area is the same person (No in Step 1007), the feature change amount matching processing unit 205 outputs the part separation position by the same process as in the first embodiment.
According to the above configuration, since the part separation position is acquired using the person region of the same person in the past, the part separation position can be obtained with high accuracy.
(Eighth embodiment)
The part separation position extraction apparatus 1 according to the present embodiment includes a feature change model accumulation unit 102 and a part separation position extraction processing unit 101, as shown in FIG. Since these configurations and operations are as described above, detailed description thereof will be omitted.
For each person area of the plurality of images, the feature change model storage unit 102 includes a plurality of axes parallel to a predetermined axis (for example, a horizontal axis and an x axis) defined on the image, and a person area included in each image. , And a feature change amount corresponding to a change in a position on a vertical axis (for example, a vertical axis or a y-axis) of a value of an image feature amount defined on each line segment intersecting with a predetermined axis is stored as a model.
The part separation position extraction processing unit 101 calculates a feature change amount for a human region on an image input from the outside, and calculates the feature change amount and the feature change amount stored in the feature change model storage unit 102. The comparison is performed, the similar change amount is specified, and the peak position where the specified change value is a predetermined value or more is output.
According to the present embodiment, by outputting the peak position of a feature change model in which the change amount of the image feature quantity is similar to the clothes worn by the person on the input image, the parts of the clothes are separated. And can be extracted.
While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
This application claims the priority on the basis of Japanese application Japanese Patent Application No. 2011-113663 for which it applied on May 20, 2011, and takes in those the indications of all here.

Claims (10)

  1.  画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納する特徴変化モデル蓄積手段と、
     外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記モデルの変化量との比較を行ない、類似する変化量をもつモデルを前記特徴変化モデル蓄積手段から抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出処理手段と、
    を備える部位分離位置抽出装置。
    For each person area on the image, the predetermined value of the feature value calculated for each line segment where a plurality of axes parallel to the predetermined axis on the image and the person area included in the image intersect A feature change model accumulating means for storing a change amount with respect to a change in position on the vertical axis with respect to the axis as a model;
    The amount of change is calculated for a human region of an external input image, the amount of change is compared with the amount of change of the model, and a model having a similar amount of change is extracted from the feature change model storage unit And a part separation position extraction processing means for outputting a peak position where the value of the amount of change of the extracted model is a position that is a predetermined value or more
    A site separation position extraction device comprising:
  2.  前記モデルは、前記変化量と、前記ピーク位置によって区切られた前記垂直軸上の区間に対して付与された、服装の種類、名称、性別の少なくともいずれかの情報である人物属性とを対応付けて包含し、
     前記部位分離位置抽出処理手段は、抽出された前記モデルの前記人物属性をさらに出力する請求項1の部位分離位置抽出装置。
    The model associates the amount of change with a person attribute, which is information on at least one of clothing type, name, and gender, assigned to the section on the vertical axis divided by the peak position. Including
    The part separation position extraction device according to claim 1, wherein the part separation position extraction processing unit further outputs the person attribute of the extracted model.
  3.  人物ごとに算出される個体特徴量と、前記人物属性と、を対応付けて格納する人物属性蓄積手段と、
     前記入力画像の前記人物領域から前記個体特徴量を抽出し、
    抽出した個体特徴量に類似する個体特徴量に対応する前記人物属性を前記人物属性蓄積手段から取得する人物特定情報抽出照合処理手段と、
    を備え、
     前記部位分離位置抽出処理手段は、前記人物特定情報抽出照合処理手段が取得した前記人物属性と類似する人物属性に対応する前記モデルの変化量を前記特徴変化モデル蓄積手段から抽出し、抽出した変化量と、前記入力画像から算出した変化量との前記比較を行なう請求項2に記載の部位分離位置抽出処理装置。
    A person attribute storage means for storing the individual feature amount calculated for each person and the person attribute in association with each other;
    Extracting the individual feature amount from the person region of the input image;
    A person specifying information extraction collation processing means for acquiring the person attribute corresponding to the individual feature quantity similar to the extracted individual feature quantity from the person attribute storage means;
    With
    The part separation position extraction processing unit extracts a change amount of the model corresponding to a person attribute similar to the person attribute acquired by the person specifying information extraction collation processing unit from the feature change model storage unit, and extracts the change The part separation position extraction processing device according to claim 2, wherein the comparison is performed between an amount and a change amount calculated from the input image.
  4.  画像上の人物領域ごとに、前記垂直軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸上の位置の変化に対する変化量(第2の変化量)を第2のモデルとして格納する第2の特徴変化モデル蓄積手段を備え、
     前記部位分離抽出手段は、前記入力画像の前記人物領域に対して前記第2の変化量を算出し、当該第2の変化量と、前記第2のモデルの前記第2の変化量との比較を行ない、類似する第2の変化量をもつ第2のモデルを前記第2の特徴変化モデル蓄積手段から抽出し、抽出された前記第2のモデルの第2の変化量の値が所定値以上の第2ピーク位置をさらに出力する請求項1乃至3のいずれかに記載の部位分離位置抽出装置。
    For each person area on the image, a feature value calculated on each line segment where a plurality of axes parallel to the vertical axis intersects the person area included in the image is on the predetermined axis. A second feature change model storage unit that stores a change amount (second change amount) with respect to a change in position as a second model;
    The part separation / extraction means calculates the second change amount for the person region of the input image, and compares the second change amount with the second change amount of the second model. And extracting a second model having a similar second change amount from the second feature change model accumulating means, and the value of the second change amount of the extracted second model is not less than a predetermined value The part separation position extraction apparatus according to any one of claims 1 to 3, further outputting the second peak position.
  5.  前記部位分離位置抽出処理手段は、前記ピーク位置によって前記人物領域を前記垂直軸方向に分割し、当該分割された人物領域の部位について前記第2の変化量を算出する請求項4に記載の部位分離位置抽出装置。 The part according to claim 4, wherein the part separation position extraction processing unit divides the person area in the vertical axis direction based on the peak position, and calculates the second change amount for the part of the divided person area. Separation position extraction device.
  6.  前記モデルは、さらに前記人物領域の人物の向きと前記変化量とを対応付けて包含し、
     前記部位分離位置抽出処理手段は、前記入力画像の前記人物領域から人物の向きを抽出し、当該人物の向きと、前記特徴変化モデル蓄積手段に格納された前記人物の向きとを比較し、類似する人物の向きと対応する前記モデルの変化量と、前記入力画像から算出した変化量との前記比較を行なう請求項1乃至5のいずれかに記載の部位分離位置抽出装置。
    The model further includes the direction of the person in the person area and the amount of change in association with each other,
    The part separation position extraction processing means extracts the orientation of the person from the person area of the input image, compares the orientation of the person with the orientation of the person stored in the feature change model accumulation means, and is similar The part separation position extraction apparatus according to claim 1, wherein the comparison is performed between a change amount of the model corresponding to a direction of a person to be performed and a change amount calculated from the input image.
  7.  前記部位分離位置抽出装置が外部から前記入力画像を取得した入力画像取得時刻と、当該画像に対して前記部位分離位置抽出処理手段が算出した前記変化量と、を対応付けて格納する部位分離位置蓄積手段と、
     前記入力画像取得時刻と、前記入力画像の人物領域と、を対応づけて格納する人物領域情報蓄積手段と、
     をさらに備え、
     前記部位分離位置抽出処理手段は、さらに外部から新規入力画像が入力された時刻の所定時間内で、かつ、前記新規入力画像の人物領域と類似する人物領域に対応する前記入力画像取得時刻である特定時刻を前記人物領域情報蓄積手段から抽出し、
     前記特定時刻の所定時間内である前記入力画像取得時刻に対応する前記変化量を前記部位分離位置蓄積手段から抽出し、抽出した変化量と、前記部位分離位置抽出処理手段が算出した前記変化量との前記比較を行なう請求項1乃至請求項6のいずれかに記載の部位分離位置抽出装置。
    The part separation position for storing the input image acquisition time when the part separation position extraction apparatus has acquired the input image from the outside and the change amount calculated by the part separation position extraction processing unit with respect to the image in association with each other Storage means;
    Person area information storage means for storing the input image acquisition time and the person area of the input image in association with each other;
    Further comprising
    The part separation position extraction processing means is the input image acquisition time corresponding to a person area similar to the person area of the new input image within a predetermined time when the new input image is input from the outside. Extract a specific time from the person area information storage means,
    The change amount corresponding to the input image acquisition time that is within a predetermined time of the specific time is extracted from the part separation position accumulation unit, and the extracted change amount and the change amount calculated by the part separation position extraction processing unit The part separation position extraction apparatus according to any one of claims 1 to 6, wherein the comparison is performed.
  8.  前記所定軸は、前記人物領域を基に定まる人物の鉛直方向の軸である請求項1乃至7のいずれかに記載の部位分離位置抽出装置。 8. The site separation position extracting apparatus according to claim 1, wherein the predetermined axis is a vertical axis of a person determined based on the person area.
  9.  画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納する特徴変化モデル蓄積手段を備えるコンピュータに、
     外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記特徴変化モデル蓄積手段に格納された前記変化量との比較を行ない、類似する変化量をもつモデルを抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出処理ステップを実行させる部位分離位置抽出プログラムを格納するプログラム記憶媒体。
    For each person area on the image, the predetermined value of the feature value calculated for each line segment where a plurality of axes parallel to the predetermined axis on the image and the person area included in the image intersect A computer comprising a feature change model storage means for storing, as a model, a change amount with respect to a change in position on a vertical axis with respect to an axis
    The amount of change is calculated for a human region of an input image from the outside, the amount of change is compared with the amount of change stored in the feature change model storage means, and a model having a similar amount of change is obtained. A program storage medium for storing a part separation position extraction program for executing a part separation position extraction processing step for outputting a peak position that is extracted and outputs a peak position where a value of a change amount of the extracted model is a predetermined value or more.
  10.  画像上の人物領域ごとに、前記画像上の所定軸に平行な複数の軸と、前記画像に含まれる人物領域と、が交わった各々の線分について算出される特徴量の値の、前記所定軸に対する垂直軸上の位置の変化に対する変化量をモデルとして格納し、
     外部からの入力画像の人物領域に対して前記変化量を算出し、当該変化量と、前記特徴変化モデル蓄積手段に格納された前記変化量との比較を行ない、類似する変化量をもつモデルを抽出し、抽出されたモデルの変化量の値が所定値以上の箇所であるピーク位置を出力する部位分離位置抽出方法。
    For each person area on the image, the predetermined value of the feature value calculated for each line segment where a plurality of axes parallel to the predetermined axis on the image and the person area included in the image intersect The amount of change with respect to the change in position on the vertical axis relative to the axis is stored as a model,
    The amount of change is calculated for a human region of an input image from the outside, the amount of change is compared with the amount of change stored in the feature change model storage means, and a model having a similar amount of change is obtained. A part separation position extraction method for extracting and outputting a peak position where a value of a change amount of the extracted model is a predetermined value or more.
PCT/JP2012/063403 2011-05-20 2012-05-18 Site separation location extraction device, program, and method WO2012161291A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2013516448A JP6090162B2 (en) 2011-05-20 2012-05-18 Site separation position extraction apparatus, program, and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011113663 2011-05-20
JP2011-113663 2011-05-20

Publications (1)

Publication Number Publication Date
WO2012161291A1 true WO2012161291A1 (en) 2012-11-29

Family

ID=47217362

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/063403 WO2012161291A1 (en) 2011-05-20 2012-05-18 Site separation location extraction device, program, and method

Country Status (2)

Country Link
JP (1) JP6090162B2 (en)
WO (1) WO2012161291A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2017141454A1 (en) * 2016-05-13 2018-02-22 株式会社日立製作所 Congestion status visualization device, congestion status visualization system, congestion status visualization method, and congestion status visualization program
US10127310B2 (en) 2015-03-24 2018-11-13 Fujitsu Limited Search method and system
JP2019158433A (en) * 2018-03-09 2019-09-19 パナソニック株式会社 Device and method for estimating weather
JP2020086636A (en) * 2018-11-19 2020-06-04 株式会社ニコン Detector, processor, detection method, and processing program

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08153187A (en) * 1994-09-27 1996-06-11 Oki Electric Ind Co Ltd Image recognizing method
JP2002344946A (en) * 2001-05-16 2002-11-29 Oki Electric Ind Co Ltd Monitoring system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5664553B2 (en) * 2009-10-16 2015-02-04 日本電気株式会社 Person clothes feature extraction device, person search device, person clothes feature extraction method, person search method, and program
US8379920B2 (en) * 2010-05-05 2013-02-19 Nec Laboratories America, Inc. Real-time clothing recognition in surveillance videos

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08153187A (en) * 1994-09-27 1996-06-11 Oki Electric Ind Co Ltd Image recognizing method
JP2002344946A (en) * 2001-05-16 2002-11-29 Oki Electric Ind Co Ltd Monitoring system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
TAKAFUMI KOSHINAKA ET AL.: "Development for Collaborative Integration of Speech and Image Recognition", JOHO SHORI, vol. 52, no. 1, 15 January 2011 (2011-01-15), pages 74 - 76 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10127310B2 (en) 2015-03-24 2018-11-13 Fujitsu Limited Search method and system
JPWO2017141454A1 (en) * 2016-05-13 2018-02-22 株式会社日立製作所 Congestion status visualization device, congestion status visualization system, congestion status visualization method, and congestion status visualization program
US10607356B2 (en) 2016-05-13 2020-03-31 Hitachi, Ltd. Congestion analysis device, congestion analysis method, and congestion analysis program
JP2019158433A (en) * 2018-03-09 2019-09-19 パナソニック株式会社 Device and method for estimating weather
JP7094120B2 (en) 2018-03-09 2022-07-01 パナソニックホールディングス株式会社 Weather estimation device and weather estimation method
JP2020086636A (en) * 2018-11-19 2020-06-04 株式会社ニコン Detector, processor, detection method, and processing program
JP7234595B2 (en) 2018-11-19 2023-03-08 株式会社ニコン Detection device, processing device, detection method, and processing program

Also Published As

Publication number Publication date
JP6090162B2 (en) 2017-03-08
JPWO2012161291A1 (en) 2014-07-31

Similar Documents

Publication Publication Date Title
US10846554B2 (en) Hash-based appearance search
TWI559242B (en) Visual clothing retrieval
JP7132387B2 (en) Image processing device, image processing method and program
JP6631179B2 (en) Foreground image division method and apparatus, program, and recording medium
JP5010937B2 (en) Image processing apparatus, program, and image processing method
US9495754B2 (en) Person clothing feature extraction device, person search device, and processing method thereof
US10002460B2 (en) Method and apparatus for estimating body shape
Pala et al. Multimodal person reidentification using RGB-D cameras
US9928411B2 (en) Image processing apparatus, image processing system, image processing method, and computer program product
JP5613741B2 (en) Image processing apparatus, method, and program
US10489640B2 (en) Determination device and determination method of persons included in imaging data
JP6090162B2 (en) Site separation position extraction apparatus, program, and method
JP2010262425A (en) Computer execution method for recognizing and classifying clothes
JP2014016968A (en) Person retrieval device and data collection device
JP6739200B2 (en) Video processing device, video processing system and control method
CA3011713A1 (en) Hash-based appearance search
JP6430102B2 (en) Person attribute estimation device, person attribute estimation method and program
JP6444331B2 (en) Object identification device
JP6544970B2 (en) IMAGE PROCESSING APPARATUS, IMAGE PROCESSING METHOD, AND PROGRAM
JP5876121B2 (en) Image processing apparatus, method, and program
Ng et al. Multi-view gait based human identification system with covariate analysis.
JP7396364B2 (en) Image processing device, image processing method, and image processing program
WO2022009279A1 (en) Image selection device, image selection method, and program
Wu et al. Clothing extraction by coarse region localization and fine foreground/background estimation
Xia Predict 3D body measurements with 2D images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12789587

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2013516448

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12789587

Country of ref document: EP

Kind code of ref document: A1