WO2016127883A1 - Image area detection method and device - Google Patents

Image area detection method and device Download PDF

Info

Publication number
WO2016127883A1
WO2016127883A1 PCT/CN2016/073274 CN2016073274W WO2016127883A1 WO 2016127883 A1 WO2016127883 A1 WO 2016127883A1 CN 2016073274 W CN2016073274 W CN 2016073274W WO 2016127883 A1 WO2016127883 A1 WO 2016127883A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
image
cluster
probability
processed
Prior art date
Application number
PCT/CN2016/073274
Other languages
French (fr)
Chinese (zh)
Inventor
石克阳
Original Assignee
阿里巴巴集团控股有限公司
石克阳
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司, 石克阳 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2016127883A1 publication Critical patent/WO2016127883A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/40Analysis of texture

Definitions

  • the present application belongs to a computer information processing neighborhood, and particularly relates to an image area detecting method and apparatus.
  • the product image can better reflect the intuitive characteristics of the product.
  • the main body area (or foreground area, such as windbreaker, casual pants, leather shoes, mobile phone, sofa stool) in the product is usually in the product image.
  • the largest and most important part of the information For example, when displaying and placing an advertisement, it is usually necessary to consider whether the main body of the product is centered in an image, whether it occupies a prescribed ratio in the image displayed by the image, and whether the main body region is prominent with respect to the background.
  • most of the product images are uploaded by the seller's merchants and uploaded on the website display window. The seller's merchants often do not have the professional shooting and image editing ability, and can not highlight the product features.
  • the business platform service party usually needs to analyze the image provided by the seller merchant, obtain the main body of the product, adjust the display angle of the product, the background matching, the placement position, the size of the main product, etc., so as to have the best display effect. Images so that consumers can more accurately get the items they are interested in, or be attracted by the merchant's items. Therefore, users of business platform servants or terminal applications often need to accurately and efficiently separate the product body area from the background area from the product image.
  • the commonly used separation technology between the main body area and the background area mainly includes the image saliency area detection technology based on the color quantization feature in the academic world.
  • image saliency area detection technology based on the color quantization feature in the academic world.
  • Such techniques are typically processed only by relying on color features and can only process simple product images.
  • the image of the product in the platform e-commerce website such as Taobao and Tmall can be uploaded by the seller, and the quality of the image is uneven and the complexity is very high. For example, when the color of the subject and the background are similar, it is easy to mix the two when using color modeling, which is difficult to distinguish and cannot effectively extract the main body area.
  • the use of color feature-based methods tends to model the background and foreground as too many blocks, resulting in the inability to accurately separate the foreground and background.
  • the commodity image subject recognition technology in the prior art cannot accurately and effectively detect and separate the subject region when facing complex images with similar colors of the subject and the background region or high complexity of the background region.
  • a more efficient and accurate detection method is needed.
  • the purpose of the present application is to provide an image region detecting method and device, which can effectively cope with various complicated situations in an actual image scene, realize accurate and effective separation of the main body region in the complex image, and improve extraction precision.
  • An image area detecting method and apparatus provided by the present application is implemented as follows:
  • An image area detecting method comprising:
  • Clustering the mixed feature vectors to obtain clusters after clustering
  • the image to be processed is detected based on the pixel probability to acquire a target area.
  • An image area detecting device comprising:
  • a feature calculation module configured to calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
  • a clustering module configured to cluster the mixed feature vectors to obtain clusters after clustering
  • a clustering probability module configured to calculate a clustering probability of the cluster according to a predetermined rule
  • a pixel probability module configured to calculate a pixel probability of a pixel point in the cluster based on the clustering probability
  • a detecting module configured to detect the image to be processed based on the pixel probability, and acquire a target area.
  • An image area detecting device configured to include:
  • a first processing unit configured to acquire a to-be-processed image of the user/client, calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
  • a second processing unit configured to cluster the mixed feature vector to obtain clustered clusters; and further configured to calculate a clustering probability of the cluster according to a predetermined rule, and calculate the clustering probability based on the clustering probability The pixel probability of a pixel in the cluster;
  • an output unit configured to acquire a target area of the to-be-processed image based on the pixel probability, and store or display the acquired target area in a specified location.
  • the image area detecting method and apparatus provided by the present application adopts a unique mixed feature vector for each pixel in the image.
  • the mixed feature vector includes a gradient feature in addition to the color feature of the pixel point.
  • the hybrid feature vector described in the present application can combine the color feature and the gradient feature to describe the pixel of the foreground and the pixel of the background into two different clusters, in the Euclidean distance. It is easy to separate the two when calculating.
  • the hybrid features are clustered, and the clustering probability that the clusters belong to the body region after clustering is calculated, and the pixel probability of each pixel in the cluster belongs to the body region is calculated based on the clustering probability, as described in the present application.
  • the calculated saliency as a probability of belonging to the subject area can effectively and accurately detect the subject area in the image to be processed.
  • the clustering and other clustering distances and the sum of the sums are used as the clustering saliency to express the probability that the cluster belongs to the body region, which is more in line with the actual user's perception of the product subject in the image, so that the processing result More precise and effective.
  • FIG. 1 is a schematic flow chart of an embodiment of an image area detecting method according to the present application.
  • FIG. 2 is a schematic diagram of a neighborhood window extraction of an image boundary point to be processed according to the present application
  • FIG. 3 is a schematic diagram of performing body region extraction using an image region detecting method according to the present application.
  • FIG. 4 is a schematic diagram of performing body region extraction using an image region detecting method according to the present application.
  • FIG. 5 is a schematic structural diagram of a module of an image area detecting apparatus according to the present application.
  • FIG. 6 is a schematic structural diagram of a module of an embodiment of a feature calculation module according to the present application.
  • FIG. 7 is a block diagram showing the structure of an embodiment of a color feature module according to the present application.
  • FIG. 8 is a schematic structural diagram of a module of an embodiment of a pixel probability calculation module according to the present application.
  • the product image uploaded by the seller merchant may include one or more subjects.
  • the seller merchant may merge the multiple images into one image and upload the image as an item.
  • An image area detecting method described in the present application may be applied to an image including one or more product bodies, where the image includes a plurality of In the main body, the image to be processed may be divided into a plurality of sub-pictures, each of which may include a single body, and then processed for each of the sub-pictures using the subject area extraction method described in the present application.
  • the method for dividing a to-be-processed image including a plurality of subjects may adopt an image segmentation method described in Patent No. CN102567952A, entitled "An Image Segmentation Method and System".
  • the product image including the plurality of subjects may be divided into a plurality of sub-pictures including a single body.
  • FIG. 1 is a flowchart of a method for detecting an image region according to an embodiment of the present invention. As shown in FIG. 1, the method may include:
  • S1 Calculating a color feature and a gradient feature of the pixel of the image to be processed, and constructing a mixed feature vector of the image to be processed.
  • the image to be processed described in this embodiment may be a single product image including one main body, or may be a sub-picture including a single main body after being divided by the image.
  • the mixed features of the pixel points of the image to be processed may be constructed based on the feature values including the color and the gradient to form a mixed feature vector.
  • feature extraction of each pixel point can usually be performed by using local features. For example, for a certain pixel point P, a neighborhood window W(p) can be selected, and the neighborhood window is selected. W(p) can be a square region of N*N centered at point P.
  • the value of the N may be reasonably selected according to the accuracy or speed of the image information processing, for example, an odd number such as 3, 5, 7, or 9 may be used according to the image size or the number of pixels included.
  • the value of N may be 5, and a 5*5 square neighborhood window region centered on the P point may be taken each time the color feature or the gradient feature in the mixed feature of the pixel is calculated.
  • the hybrid feature of the image to be processed constructed in this embodiment may include a color feature and a gradient feature of the pixel, and the color feature and the gradient feature may be combined in a predetermined format to form a high-dimensional mixed feature vector.
  • the process of calculating the color feature of the pixel of the image to be processed may include:
  • S102 extract a pixel point of the neighborhood window W(p) in the image to be processed centering on the pixel to be processed, and divide the three channels L, a, and b of the pixel in the neighborhood window W(p) For K groups, form a color feature vector of 3*K dimensions;
  • the image color feature extraction of the image to be processed may generally include uniformly quantizing the three channels L, a, and b into K packets, and ensuring that each packet of each channel has the same length as possible.
  • the image to be processed may be image information of an RGB channel color model
  • the color model of the Lab channel generally refers to a color model based on a person's perception of color and independent of light and equipment, and is more in line with human visual perception. Therefore, in this embodiment, the body region in the image detected from the Lab space is used, which is more in line with the human perception result, so that the processing result of the body region extraction is more accurate.
  • the image to be processed can be converted from an RGB channel to a Lab channel.
  • the RGB channel consists of three variable color vector vectors (R, G, B) as follows:
  • R red, an integer from 0 to 255, with a change value of 256;
  • G green, an integer from 0 to 255, with a change value of 256;
  • B Blue, an integer from 0 to 255, with a change value of 256.
  • the Lab channel can include three variables as shown below:
  • L brightness, an integer from 0 to 100, a change value of 100
  • A from green to red, an integer from -128 to 127, with a value of 256;
  • B an integer from -28 to 127 from blue to yellow, with a value of 256.
  • the conversion may be performed by using a given algorithm, or may be converted by using a software tool such as Photoshop, and will not be discussed in detail herein.
  • the pixel points in the image to be processed of the Lab channel may be extracted by using a preset neighborhood window W(p), and the three channels of L, a, and b of the pixel in the neighborhood window W(p) are respectively
  • the uniform quantization is K bins.
  • the values of the pixel points quantized by the three channels L, a, and b can be spliced together to form a 3*K-dimensional color feature vector.
  • the formed 3*K-dimensional color feature vector can be expressed as ⁇ L1, L2, ...
  • K described in this embodiment can be customized to indicate a description of the color space of the image to be processed.
  • the value of the K is too large, the image to be processed is divided into finer colors in the color space, the color feature is more accurately expressed, and the corresponding calculation time is increased; Small, then the overall division degree of the image to be processed in the color space is low, and the dimension of the color feature vector is small, which can improve the data processing speed.
  • the present application provides a range of values of K.
  • the specific value of the K may be: 6 ⁇ K ⁇ 16, which can ensure that the color feature vector can be accurate, effective, and suitable within the above range of values.
  • the value of K may be a value of 6, that is, an 18-dimensional mixed feature vector of a pixel to be processed in the neighborhood window may be constructed.
  • the color values of the three channels L, a, b of each pixel in the neighborhood window W(p) it is added to the corresponding dimension of the color feature vector. For example, in a neighborhood window of 5*5 and 25 pixels, the Lab color values of the 25 pixels jointly construct an 18-dimensional color feature vector.
  • each of the 25 pixels has a set of Lab color values, and the L channel is taken as an example. If the value of the L channel of the first pixel is 10, it can be mapped to the L channel. One of the corresponding six bins (grouping), for example, divided into L1. The L pixel value of the second pixel is 98, which can be divided into L6.
  • one pixel After calculating the color feature vector of the pixel to be processed in the current neighborhood window, one pixel can be shifted once in a certain direction, and then the pixel of the neighborhood window is extracted again according to the above manner, and the new neighborhood window is calculated to be processed.
  • the color feature vector of the pixel After calculating the color feature vector of the pixel to be processed in the current neighborhood window, one pixel can be shifted once in a certain direction, and then the pixel of the neighborhood window is extracted again according to the above manner, and the new neighborhood window is calculated to be processed.
  • the color feature vector of the pixel The color feature vectors of all the pixels in the image to be processed are sequentially calculated, and the color features of the pixels of the image to be processed are acquired.
  • the pixel to be processed in the neighborhood window described in the present application is usually the center point of the set square neighborhood window.
  • a square neighborhood window may be extracted at one time.
  • the extraction specification preset according to the neighborhood window may be used, that is, the pixel point of the boundary point or the pixel point close to the boundary may be centered. The calculation is performed on the pixel points that the neighborhood window actually covers in the image to be processed.
  • FIG. 2 is a schematic diagram of a neighborhood window extraction of an image boundary point to be processed according to the present application. As shown in FIG.
  • the set neighborhood window extraction rule is a 5*5 square region, and for a non-corner point of a certain boundary point, the pixel point P1 of the boundary point is extracted as a neighborhood window center.
  • the pixel size specification is 5*3.
  • the extracted pixel point has a specification of 3*3.
  • the blending features described in this application can include gradient features of the image to be processed.
  • the HoG feature may be used to perform gradient feature extraction to form a gradient feature of M-dimensionality of each pixel in the image to be processed.
  • the meaning of the gradient may include the difference between each pixel in the image and the adjacent pixel, which can be used to detect an area where the color is not obvious after being constructed as a gradient feature.
  • the image to be processed can be converted from an RGB color channel to a grayscale image, thus simplifying the complexity of the gradient feature.
  • the HoG feature may be used to perform gradient feature extraction, and obtain a gradient direction and a gradient value of the pixel in the preset neighborhood window W(p), and then the neighbor window W(p) may be included.
  • the total gradient direction value of all pixel point gradient directions is divided into M bins, for example, the total gradient direction of 180 degrees is divided into 12 bins, and each bin represents a range of 15 degrees.
  • a linear interpolation method is used to accumulate the corresponding bin (packet) to form an M-dimensional of the pixel to be processed in the neighborhood window.
  • the gradient feature vector such as the 12-dimensional gradient feature vector in this embodiment, can be expressed as ⁇ g1, g2, ... g12 ⁇ .
  • the gradient direction of a point in the neighborhood window W(p) of the pixel to be processed is 44 degrees and the gradient value is 10
  • the color feature calculation is performed.
  • a gradient value of 44 degrees can be accumulated in the group g3 to which the value belongs.
  • the gradient direction and the gradient value of all the pixel points of the neighborhood window are traversed, and the gradient feature of the pixel to be processed of the neighborhood window is calculated.
  • one pixel can be shifted, and the gradient feature of the next pixel to be processed is continuously calculated.
  • the gradient features of all the pixels of the image to be processed are calculated in sequence, and the calculation manner of the above-mentioned color features may be referred to, and details are not described herein.
  • the mixed feature vector of the image to be processed may be constructed.
  • the constructing the mixed feature vector of the image to be processed may include splicing and combining the K-dimensional color feature and the M-dimensional gradient feature of each pixel of the image to be processed to form a (K+M) dimension for the pixel.
  • Mixed feature vector For example, in this embodiment, the 18-dimensional color feature and the 12-dimensional gradient feature value can be sequentially combined and stitched.
  • the first 18-dimensional data is a color feature
  • the latter 12-dimensional data is a gradient feature, which can be expressed as ⁇ L1, L2,... L6, a1, a2, ... a6, b1, b2, ...
  • the method can be constructed by the above method.
  • the feature value of the pixel can be more accurately established, so that the foreground and the background region are similar when the feature space is mixed.
  • the distance between the mixed feature vectors of the two points is greatly increased compared with the distance using only the color features, and the foreground and background regions can be effectively distinguished, and the accuracy of the detection of the main body region is improved.
  • the product image of the image size [W, H] to be processed in the foregoing may generate a mixed feature vector of W*H (K+M) dimensions.
  • these feature vectors can be clustered.
  • the clustering algorithm used in this embodiment may be a Kmeans clustering algorithm.
  • the specific operation process of the Kmeans clustering algorithm may mainly include:
  • S201 randomly select L mixed feature vectors from the mixed feature vectors of the W*H (K+M) dimensions as an initial cluster center.
  • the value range of the L may be tested to select a suitable value. Generally, the L value is too large, which may result in a long calculation time. If the L is too small, the feature space may not be divided finely.
  • S202 Traverse all W*K mixed feature vectors, and calculate the distance between each mixed feature vector and the current cluster center respectively.
  • the distances described in this embodiment are Euclidean distances.
  • two mixed feature vectors are p and q, respectively, where q is a randomly selected current cluster center, then the mixed feature vector and the current cluster center.
  • the Euclidean distance D(p,q) between q can be:
  • S203 Calculate, for each mixed feature vector, a distance from the selected L initial cluster centers, where the mixed feature vector belongs to a cluster with the smallest distance from the L initial cluster centers. After a round of calculation and classification, the mixed feature vector can be reasonably divided into the clusters of the L initial cluster centers closest to each other.
  • S204 Update the cluster center of each cluster. After each pixel in the image to be processed is divided into corresponding clusters, To update the cluster center of each cluster.
  • the specific update calculation method in this embodiment may include calculating an average value of each mixed feature vector in each cluster in each dimension, and then using the calculated average value of each dimension as the cluster new Cluster center.
  • the above-mentioned S201-S204 is a process of one-time clustering.
  • the steps of dividing the clustering and updating the clustering center for each pixel point may be repeatedly performed by clustering until the clustering center of the clustering is no longer A larger amount of movement is performed (the amplitude threshold of the movement can be set according to requirements) or the number of times the cluster calculation reaches the preset calculation requirement.
  • the number of clustering of the mixed feature vector may be set to 1000 times, or the distance between the cluster center of the new cluster and the cluster center of the cluster is less than 0.5. If the cluster center is the old_C and the new cluster center is New_C, the stop condition of the cluster calculation can be set to D (Old_C, New_C) ⁇ 0.5.
  • the mixed feature vectors are clustered to form L clusters, and the calculation of the mixed features of the plurality of pixel points in the image to be processed can be reduced to the calculation of L clusters, thereby improving subsequent image area detection.
  • the further calculation rate improves the overall image information processing efficiency.
  • S3 Calculate a clustering probability of the cluster according to a predetermined rule, and calculate a pixel probability of the pixel point in the cluster based on the clustering probability.
  • the to-be-processed image is clustered into L clusters in the (K+M)-dimensional mixed feature vector space described in the present application, wherein each cluster of the L clusters The pixels within are similar in the feature space.
  • a clustering probability that each cluster belongs to a body region may be calculated in units of each of the clusters, and then all pixel points in the cluster are calculated based on the clustering probability of the clustering.
  • the pixel probability of the body area may be used to describe the probability that each cluster belongs to the body region.
  • calculating the clustering probability of the cluster according to a predetermined rule may include:
  • clustering centers of the L clusters obtained after clustering are respectively C1, C2, ..., CL, and the degree of saliency of the cluster in this embodiment may be adopted by distance from all other clusters. And the ratio of the sum to the sum.
  • the embodiment provides a method for calculating the distance sum of each cluster and other clusters in the cluster, specifically the Ci of the cluster and other
  • the cluster distance and D(Ci) can be calculated by the following formula (1):
  • L is the number of clusters, as set in the embodiment 120
  • is the mixed feature vector of the cluster center of the current cluster Ci and the clustering of other clusters
  • the larger the difference of the mixed feature vectors between the two clusters the larger the Euclidean distance between the two cluster centers. If the distance between a cluster and other clusters is larger overall, it can be said that the higher the difference between the cluster and other clusters, the more likely it is to approach the body region of the image to be processed, and the corresponding calculation is obtained. The sum of the distances from other clusters is also larger.
  • a factor Wj is added, and the Wj may be a weight set according to a pixel point included in the current cluster Ci.
  • the Wj can be set according to the pixel points included in the cluster. For example, it may be set to the number of pixels included in the cluster, or the ratio of the number of pixels included in the current cluster to the total number of pixels of the image to be processed, etc., and may be specifically set according to requirements.
  • the weight Wj of the cluster is added, and the number of pixels included in the cluster is counted, which is more in line with the actual image body area in some application scenarios.
  • the calculation result of extracting the main map area can be more accurate.
  • the clustering probability that each cluster belongs to the body region may be further calculated according to the saliency.
  • the ratio of the cluster and D(Ci) to the sum of the distance sums of all the clusters may be used as the clustering probability that the cluster Ci belongs to the body region, and the specific formula may be (2) Calculated:
  • the above middle ⁇ 1 ⁇ j ⁇ L D(c j ) is the sum of the calculated cluster sums of all the clusters, and the distance of the current cluster and the ratio of the sum may be used as the current cluster
  • the clustering probability of the subject area Since the mixed feature vector values in the clustered clusters are relatively close, in one embodiment of the present application, it can be considered that the pixel probability of the pixel points belonging to the body region in the cluster is equivalent to the cluster of the cluster belonging to the body region. Probability, such that a probability value for each pixel can be derived from the probability of the cluster. Therefore, in an embodiment of the present application, the calculating a pixel probability of a pixel point in the cluster based on the clustering probability may include:
  • the pixel probability of the pixel in the cluster may be a clustering probability of the cluster to which the pixel belongs.
  • the pixels in the cluster may be distributed in other regions of the image to be processed.
  • the extracted body region has a compact feature, and the extracted body region is more accurate. Calculate again The pixel probability of each pixel in each cluster belonging to the body area.
  • the present application may set a second neighborhood window W(p)', and may extract the pixel points of the second neighborhood window W(p)' centering on the pixel point P by referring to the manner of calculating the color feature.
  • the probability of a certain pixel point q in the second neighborhood window W(p)' is the clustering probability of the cluster to which the pixel point q belongs, which is represented by P(q), and in another embodiment Calculating the probability that the pixel points in the cluster belong to the body region based on the probability of the clustering may include:
  • S302 extract a pixel point of the first neighborhood window W(p)' centering on the pixel to be obtained p, and calculate a pixel probability Sal(p) of the pixel to be obtained p belonging to the body region by using the following formula:
  • P(q) is the clustering probability of the cluster to which the pixel point q in the first neighborhood window W(p)' belongs belongs to the body region
  • t is the cluster to which the pixel point p to be sought belongs.
  • the number of pixels, ⁇ is a set of smoothing parameters, which can indicate the size of the currently calculated pixel point p is affected by the surrounding pixels. If the value of ⁇ is large, it can be said that the calculation result of the pixel point p is easily affected by the surrounding pixel points, and vice versa.
  • the ⁇ value can be set according to the experience or the estimation of the result. Generally, for the image sold by the website product, the ⁇ value may be small, for example, the specific value in the embodiment may be 0.17. If the image is in a natural scene (usually a non-commodity image), the value of ⁇ may be too large, for example, may be 0.25.
  • the setting of the first neighborhood window W(p)' described above may be the same as the neighborhood window set in the foregoing color feature extraction, for example, a square neighborhood window of 5*5 may be set.
  • the pixel of the first neighborhood window W(p)′ such as 5*5
  • the probability of all pixel points in a neighborhood window W(p)' can be calculated as the pixel probability that the pixel point p to be sought belongs to the body region.
  • the pixel point according to the above S302 belongs to the body region probability calculation method, and the pixel probability that each pixel in the image to be processed belongs to the body region can be calculated, and the probability value adopts the first neighborhood window W ( The probability value of the pixel point in p)' is smoothed and calculated, which can improve the accuracy of the final extraction result.
  • the image to be processed is detected based on the pixel probability, and the target area is acquired.
  • the body region and the background region may be separated, and the target region in the image to be processed is extracted and acquired.
  • the target area described in the present application may be a body area (foreground area) in the image to be processed, and in other embodiments, the target area may also be For the background area, that is, the background area of the image to be processed can be detected.
  • the detecting the image to be processed based on the pixel probability to obtain the target area may include:
  • a pixel point that meets a pixel probability value of a pixel in the image to be processed according to a determination threshold PV is used as a target area of the image to be processed.
  • a determination threshold PV of the pixel point probability such as 0.85
  • the pixel point of the pixel probability of the pixel to be processed may be greater than 0.85.
  • the value range of the threshold, the specific predetermined threshold value PV may be: 0.8 ⁇ PV ⁇ 0.95.
  • the pixel probability of the pixel point in the above S401 is preferably a probability value obtained by smoothing the probability value of the pixel point in the first neighborhood window W(p)'.
  • a value that satisfies the determination threshold PV determined as the background area may be set, and the specific determination may be performed according to the actual scenario application.
  • the detecting, by the pixel probability of the pixel, the target to be processed may be:
  • S4021 The pixel point in the image to be processed that belongs to the body region and whose probability value is greater than the first threshold PF is used as the seed pixel point;
  • S4022 Calculate a Euclidean distance from a pixel in a surrounding second neighborhood window centering on the seed pixel point;
  • S2044 traverse the Euclidean distance of all the seed pixel points and the pixel points in the surrounding second neighborhood window and make a determination, and use the calculated seed pixel point as the target area of the to-be-processed image.
  • the pixel probability that the pixel belongs to the body region is preferably a clustering probability of the cluster to which the pixel belongs.
  • the first threshold PF and the second threshold and the third neighborhood window may be set according to actual data processing requirements, for example, the first valve PF value may also be set to 0.85 or selected as a clustering probability.
  • the value of the higher value, the second threshold can be set to 0.5. If the threshold value of the first threshold PF is too small, the pixel value of the non-subject area is extracted too much, and if the value is too large, the integrity of the extracted image of the body area is reduced.
  • the value range of the first threshold PF is set, and the value of the first threshold PF may be: 0.8 ⁇ PF ⁇ 0.95.
  • the third neighborhood window described in this embodiment is generally a 3*3 eight-contiguous window centered on the seed pixel point, and then the Euclidean distance calculation can be performed according to the 30-dimensional feature mixed feature vector described in the present application. . If the distance satisfies the second threshold requirement, a pixel point satisfying the second threshold requirement around the seed may be used as a new seed pixel point, and a new seed pixel point that meets the second threshold requirement may be considered to belong to the body area. . when However, during processing, a pixel point that does not satisfy the third neighborhood window may be set as a background area. It should be noted that the body area described in the present application is generally connected.
  • a pixel point that has not been judged by the second threshold may be set as a background area.
  • a pixel point having a larger probability value may be used as a seed pixel point, and then the surrounding neighboring points are continuously traversed and a judgment is made to finally obtain a body region.
  • the manner of acquiring the target area may include, but is not limited to, the embodiments described in the present application, and other processing methods that do not require creative labor based on the method described in this application.
  • the body region obtained by separating the body region from the background region is performed, for example, using a geodesic distance algorithm.
  • the image region detecting method provided by the present application constructs a mixed feature vector including pixel color features and gradient features, which can more accurately establish feature values of pixel points, can effectively distinguish foreground and background regions, and improve the subject.
  • the accuracy of regional extraction Similarly, in a complex background image, the hybrid feature vector described in the present application can combine the color feature and the gradient feature to describe the pixel of the foreground and the pixel of the background into two different clusters, in the Euclidean distance. It is easy to separate the two when calculating.
  • the mixed features are clustered, and the clustering and other clustering distances and the sum of the sums are used as the clustering saliency to express the probability that the cluster belongs to the main body, which is more realistic.
  • the user perceives the situation of the product body in the image, making the processing result more accurate and effective.
  • the accuracy of extracting the main body region of the image to be processed by using the main body region extraction method of the present application reaches 89.62%, and the recall rate reaches 88.83%, which solves the problem in the prior art when facing a complex image.
  • the problem of low regional extraction accuracy 89.62%, and the recall rate reaches 88.83%, which solves the problem in the prior art when facing a complex image.
  • FIG. 3 and FIG. 4 are schematic diagrams of extracting a main body region by using an image region detecting method according to the present application, and FIG. 3 and FIG. 4 are respectively to-be-processed images, existing algorithm extraction results, and extraction by the present invention from left to right. result.
  • FIG. 3 an image with a very similar color between the foreground and the background area is selected.
  • the existing algorithm cannot detect the highlighted portion of the garment when processing such an image because The color here is very close to the white of the background.
  • the mixed feature vector of (K+M) dimension of the present application can effectively distinguish similar foreground and background regions.
  • FIG. 4 is a case where the background is complicated. It can be seen from FIG.
  • the method of the present application uses the cluster to acquire the cluster to calculate the pixel points.
  • the pixel probability of the main body region can effectively solve the problem of image subject extraction on the background not only in color but also in structure, which greatly improves the detection accuracy.
  • FIG. 5 is a schematic structural diagram of a module of an image area detecting apparatus according to the present application. As shown in FIG. 5, the apparatus may include:
  • the feature calculation module 101 is configured to calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
  • the clustering module 102 can be configured to cluster the mixed feature vectors to obtain clusters after clustering;
  • the clustering probability module 103 may be configured to calculate a clustering probability of the cluster according to a predetermined rule
  • a pixel probability module 104 configured to calculate a pixel probability of a pixel point in the cluster based on a probability of the clustering
  • the detecting module 105 is configured to detect the image to be processed based on the pixel probability to acquire a target area.
  • FIG. 6 is a schematic diagram of a module structure of an embodiment of a feature calculation module 101 according to the present application. As shown in FIG. 6, the feature calculation module 101 may be configured to include:
  • the color feature module 1011 is configured to calculate a color feature of the pixel of the image to be processed
  • the gradient feature module 1012 can be configured to calculate a gradient feature of the pixel of the image to be processed
  • the blending feature module 1013 can be configured to combine the color features and the gradient features to form a mixed feature vector of the image to be processed.
  • FIG. 7 is a schematic diagram of a module structure of an embodiment of a feature calculation module 1011 according to the present application. As shown in FIG. 7, the color feature module 1011 may include:
  • the Lab conversion module 111 can be configured to convert the image to be processed into data in a Lab format
  • the color feature vector module 112 may be configured to extract pixel points of the neighborhood window in the image to be processed centering on the pixel to be processed, and divide the three channels L, a, and b of the pixel in the neighborhood window into two K groups, forming a color feature vector of 3*K dimensions;
  • the feature calculation module 113 may be configured to add color values of each of the L, a, and b channels of each pixel in the neighborhood window to a dimension corresponding to the color feature vector to form the neighbor. The color characteristics of the pixels to be processed in the domain window.
  • the color characteristics of the image to be processed can be obtained.
  • the present application provides a value range of K for the device, and the specific value of the K may be: 6 ⁇ K ⁇ 16, and the color feature vector extracted by the device of the present application can be ensured within the above range. Effectively and appropriately express the color characteristics of the image to be processed.
  • the clustering probability module 103 in the foregoing device calculates the probability that the cluster belongs to the body region, and may specifically include:
  • a distance and calculation module which can be used to calculate a distance between each cluster and other clusters in the cluster
  • the clustering probability calculation module may be configured to calculate a clustering probability of the cluster according to the cluster and the sum of the distance sums of all clusters.
  • the distance calculating module calculates the distance and specificity of each cluster in the cluster from other clusters, and may include:
  • L is the number of clusters
  • is the Euclidean distance of the mixed feature vector of the cluster center of the current cluster Ci and the clustered feature vector of the cluster cluster of other clusters
  • Wj is a weight set according to the pixel points included in the current cluster Ci.
  • the pixel probability module 104 may include at least one of the following:
  • the first probability module 1041 may be configured to use the clustering probability of the cluster to which the pixel belongs to be the pixel probability of the pixel;
  • the second probability module 1042 may be configured to extract a pixel point of the first neighborhood window W(p)′ centering on the pixel point p to be obtained, and calculate a pixel probability Sal(p) of the pixel point p to be obtained by using the following formula: :
  • P(q) is the probability that the cluster to which the pixel point q in the first neighborhood window W(p)' belongs belongs to the body region
  • t is the pixel point in the cluster to which the pixel point p to be sought belongs
  • the number of ⁇ is a smoothing parameter set.
  • the extraction module 105 may extract a body region of the image to be processed by using different extraction methods set in advance. Specifically, at least one of the following modules may be included:
  • a first extraction module configured to use, as a target area of the to-be-processed image, a pixel point in which a pixel probability value of a pixel in the image to be processed meets a determination threshold PV requirement;
  • the second extraction module may be configured to use, as a seed pixel point, a PF pixel point whose probability value of the pixel in the image to be processed belongs to the body region is greater than the first threshold; and may also be used to calculate and center the pixel pixel a Euclidean distance of a pixel in the surrounding second neighborhood window; and may also be used to use the pixel point whose Euclidean distance is less than the second threshold as a new seed pixel point; and may also be used to traverse all of the seed pixel points and surrounding areas Determining the Euclidean distance of the pixel in the second neighborhood window and making a judgment, and using the calculated seed pixel as the target area of the image to be processed.
  • the value of the determination threshold PV may be: 0.8 ⁇ PV ⁇ 0.95;
  • the value of the first threshold PF may be: 0.8 ⁇ PF ⁇ 0.95.
  • the determination threshold PV or the value range of the first threshold PF provided in this embodiment can effectively ensure the correctness and validity of the extraction of the main body region, and improve the accuracy of the image region detection with high image complexity.
  • An image area detecting device can be used in a platform type e-commerce website to separate a body area and a background area in a complex and varied product image, and can effectively cope with various complicated situations in an actual image scene. It can accurately and effectively separate the main body area in complex images and improve the accuracy of image detection.
  • An image area detecting apparatus described in the present application can be used in a variety of terminal devices, such as a mapping application of a user mobile client, or a client or server dedicated to image body or background area extraction. Generally, after performing image detection and acquiring a target area, the image detecting apparatus may save or display the image of the acquired target area to the user for further processing.
  • the present application provides an image area detecting apparatus, which can be applied to process an image of a user or a client, perform image detection, and acquire a target area.
  • the device may be configured to include:
  • the first processing unit may be configured to acquire a to-be-processed image of the user/client, calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
  • a second processing unit configured to perform clustering on the hybrid feature vector to obtain clustered clusters; and may be further configured to calculate a clustering probability of the cluster according to a predetermined rule, and based on the clustering probability Calculating a pixel probability of a pixel point in the cluster;
  • the output unit may be configured to acquire a target area of the image to be processed based on the pixel probability, and store or display the acquired target area at a specified location.
  • the image de-detection device provided in this embodiment can effectively and accurately extract the target area of the to-be-processed picture in the client or the server, and can improve the accuracy of the client-side picture processing user experience or the client/server image information processing.
  • the present application is not limited to a format conversion, a clustering method, or the present application, which must be completely standard.
  • the case of the fixed formula The above description of the various embodiments in the present application is only an application in some embodiments of the present application.
  • the slightly modified processing method may also implement the foregoing embodiments of the present application. Program.
  • the same application can still be implemented without any inventive variation of the processing method steps described in the above embodiments of the present application, and details are not described herein again.
  • the unit or module illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product having a certain function.
  • the above devices are described as being separately divided into various modules by function.
  • the functions of the modules may be implemented in the same software or software and/or hardware when implementing the present application, or the modules implementing the same functions may be implemented by multiple sub-modules or a combination of sub-units.
  • the controller can be logically programmed by means of logic gates, switches, ASICs, programmable logic The form of the controller and embedded microcontroller, etc. to achieve the same function. Therefore, such a controller can be considered as a hardware component, and a device for internally implementing it for implementing various functions can also be regarded as a structure within a hardware component. Or even a device for implementing various functions can be considered as a software module that can be both a method of implementation and a structure within a hardware component.
  • the application can be described in the general context of computer-executable instructions executed by a computer, such as a program module.
  • program modules include routines, programs, objects, components, data structures, classes, and the like that perform particular tasks or implement particular abstract data types.
  • the present application can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are connected through a communication network.
  • program modules can be located in both local and remote computer storage media including storage devices.
  • the technical solution of the present application may be embodied in the form of a software product in essence or in the form of a software product, which may be stored in a storage medium such as a ROM/RAM or a disk. , an optical disk, etc., includes instructions for causing a computer device (which may be a personal computer, mobile terminal, server, or network device, etc.) to perform the methods described in various embodiments of the present application or portions of the embodiments.
  • a computer device which may be a personal computer, mobile terminal, server, or network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

An image area detection method and device. The method comprises: calculating color features and gradient features of pixels in an image to be processed, and constructing mixed feature vectors of the image to be processed (S1); clustering the mixed feature vectors to obtain a cluster after clustering (S2); calculating a clustering probability of the cluster according to predetermined rules, and calculating a pixel probability of pixels in the cluster according to the clustering probability (S3); and detecting the image to be processed according to the pixel probability to acquire a target area (S4). The method and the device can effectively cope with various complex situations in an actual image scene, realize accurate and effective separation of the main area in a commodity image, and improve extraction accuracy.

Description

一种图像区域检测方法及装置Image area detecting method and device
技术邻域Technical neighborhood
本申请属于计算机信息处理邻域,尤其涉及一种图像区域检测方法及装置。The present application belongs to a computer information processing neighborhood, and particularly relates to an image area detecting method and apparatus.
背景技术Background technique
随着互联网消费时代的发展,例如一淘、淘宝以及天猫商城等提供在线商品搜索和在线购物的网站在商品信息展示时通常会提供大量有关商品的图像,以便于消费者进行直观的选择。商品图像作为在线搜索和购物的网站中承载较多,是非常重要信息,对于商品成交有着极大的影响。With the development of the era of Internet consumption, websites that provide online product search and online shopping, such as Amoy, Taobao, and Tmall Mall, often provide a large number of images of products when displaying product information, so that consumers can make intuitive choices. Commodity images are more important information as a website for online search and shopping, and have a great impact on the transaction of goods.
在网上商品信息展示中,通常商品图像可以较好的体现商品的直观特性,商品中的主体区域(或者称为前景区域,比如风衣、休闲裤、皮鞋、手机、沙发凳)通常为商品图像中信息量最大、最主要的部分。例如,在商品展示、投放广告时,通常需要考虑在一幅图像当中,商品主体是否居中、是否在图像所展示的画面中占据符合规定的比例、主体区域相对于背景是否突出等。而实际的应用中绝大部分商品图像由卖家商户自行拍摄上传在网站展示窗口,卖家商户往往不具备专业的拍摄和图像编辑能力,不能很好的突出展示商品特征。因此一些应用场景中商务平台服务方通常需要对卖家商户提供的图像进行分析,获取商品主体,调整商品的展示角度、背景搭配、摆放位置、主体商品大小等,使其具有最佳展示效果的图像,以便于消费者能够更准确获取其感兴趣的商品,或者被商户的商品吸引。因此,商务平台服务方或者终端应用的用户通常需要精准且高效的从商品图像中将商品主体区域和背景区域分离出来。In the online product information display, usually the product image can better reflect the intuitive characteristics of the product. The main body area (or foreground area, such as windbreaker, casual pants, leather shoes, mobile phone, sofa stool) in the product is usually in the product image. The largest and most important part of the information. For example, when displaying and placing an advertisement, it is usually necessary to consider whether the main body of the product is centered in an image, whether it occupies a prescribed ratio in the image displayed by the image, and whether the main body region is prominent with respect to the background. In the actual application, most of the product images are uploaded by the seller's merchants and uploaded on the website display window. The seller's merchants often do not have the professional shooting and image editing ability, and can not highlight the product features. Therefore, in some application scenarios, the business platform service party usually needs to analyze the image provided by the seller merchant, obtain the main body of the product, adjust the display angle of the product, the background matching, the placement position, the size of the main product, etc., so as to have the best display effect. Images so that consumers can more accurately get the items they are interested in, or be attracted by the merchant's items. Therefore, users of business platform servants or terminal applications often need to accurately and efficiently separate the product body area from the background area from the product image.
目前常用的商品主体区域与背景区域分离技术主要包括采用学术界中基于颜色量化特征的图像显著性区域检测技术。这类技术通常由于仅仅依赖于颜色特征进行处理,仅能对简单的商品图像进行处理。而淘宝、天猫等平台型电商网站中的商品图像可以由卖家上传,图像的质量参差不齐,复杂度也非常高。例如在主体和背景颜色相似的情况下,在使用颜色建模的时候很容易将两者混在一起,难以区分,无法有效提取主体区域。同样,在背景复杂度较高即非主体区域的颜色分布复杂时,使用基于颜色特征的方法往往会将背景和前景建模为过多的区块,导致也无法精确的分离前景和背景。At present, the commonly used separation technology between the main body area and the background area mainly includes the image saliency area detection technology based on the color quantization feature in the academic world. Such techniques are typically processed only by relying on color features and can only process simple product images. The image of the product in the platform e-commerce website such as Taobao and Tmall can be uploaded by the seller, and the quality of the image is uneven and the complexity is very high. For example, when the color of the subject and the background are similar, it is easy to mix the two when using color modeling, which is difficult to distinguish and cannot effectively extract the main body area. Similarly, when the background complexity is high, that is, the color distribution of the non-subject area is complex, the use of color feature-based methods tends to model the background and foreground as too many blocks, resulting in the inability to accurately separate the foreground and background.
目前现有技术中商品图像主体识别技术在面临主体和背景区域颜色相近或者背景区域复杂度高等复杂图像时不能精确、有效的进行主体区域的检测、分离。现有技术中尤其是复杂图像区域检测时亟需一种更加高效、精确的检测方法。 At present, the commodity image subject recognition technology in the prior art cannot accurately and effectively detect and separate the subject region when facing complex images with similar colors of the subject and the background region or high complexity of the background region. In the prior art, especially for complex image area detection, a more efficient and accurate detection method is needed.
发明内容Summary of the invention
本申请目的在于提供一种图像区域检测方法及装置,能有效应对实际图像场景中各种复杂的情况,实现对复杂图像中主体区域进行准确、有效的分离,提高提取精确度。The purpose of the present application is to provide an image region detecting method and device, which can effectively cope with various complicated situations in an actual image scene, realize accurate and effective separation of the main body region in the complex image, and improve extraction precision.
本申请提供的一种图像区域检测方法和装置是这样实现的:An image area detecting method and apparatus provided by the present application is implemented as follows:
一种图像区域检测方法,所述方法包括:An image area detecting method, the method comprising:
计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量;Calculating a color feature and a gradient feature of the pixel of the image to be processed, and constructing a mixed feature vector of the image to be processed;
对所述混合特征向量进行聚类,获取聚类后的聚簇;Clustering the mixed feature vectors to obtain clusters after clustering;
根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率;Calculating a clustering probability of the cluster according to a predetermined rule, and calculating a pixel probability of the pixel point in the cluster based on the clustering probability;
基于所述像素概率对所述待处理图像进行检测,获取目标区域。The image to be processed is detected based on the pixel probability to acquire a target area.
一种图像区域检测装置,所述装置包括:An image area detecting device, the device comprising:
特征计算模块,用于计算得出待处理图像像素点的颜色特征和梯度特征,并构建所述待处理图像的混合特征向量;a feature calculation module, configured to calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
聚类模块,用于对所述混合特征向量进行聚类,获取聚类后的聚簇;a clustering module, configured to cluster the mixed feature vectors to obtain clusters after clustering;
聚簇概率模块,用于根据预定规则计算所述聚簇的聚簇概率;a clustering probability module, configured to calculate a clustering probability of the cluster according to a predetermined rule;
像素概率模块,用于基于所述聚簇概率计算所述聚簇中像素点的像素概率;a pixel probability module, configured to calculate a pixel probability of a pixel point in the cluster based on the clustering probability;
检测模块,用于基于所述像素概率对所述待处理图像进行检测,获取目标区域。And a detecting module, configured to detect the image to be processed based on the pixel probability, and acquire a target area.
一种图像区域检测装置,所述装置被设置成,包括:An image area detecting device, the device being configured to include:
第一处理单元,用于获取用户/客户端的待处理图像,计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量;a first processing unit, configured to acquire a to-be-processed image of the user/client, calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
第二处理单元,用于对所述混合特征向量进行聚类,获取聚类后的聚簇;还用于根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率;a second processing unit, configured to cluster the mixed feature vector to obtain clustered clusters; and further configured to calculate a clustering probability of the cluster according to a predetermined rule, and calculate the clustering probability based on the clustering probability The pixel probability of a pixel in the cluster;
输出单元,用于基于所述像素概率对所述待处理图像进行获取目标区域,并将所述获取的目标区域存储或者展示于指定位置。And an output unit, configured to acquire a target area of the to-be-processed image based on the pixel probability, and store or display the acquired target area in a specified location.
本申请提供的一种图像区域检测方法及装置,采用为图像中的每个像素点构建其特有的混合特征向量。所述混合特征向量中除了可以包括像素点的颜色特征外还包括梯度特征,在计算像素点时同时考虑了像素点周围的信息,可以更加准确的建立像素点的特征值,使得混合特征空间时前景和背景区域相近的两个点的混合特征向量的距离比仅仅使用颜色特征的 距离大大增加,可以有效的区分前景和背景相近的区域,提高目前区域检测的精准度。同样的,在复杂背景图像中,本申请所述的混合特征向量可以很好的结合颜色特征和梯度特征将前景的像素点和背景的像素点描述到两个不同的聚簇中,在欧式距离计算时可以很容易将两者分离。本申请中对混合特征进行聚类,计算聚类后聚簇属于主体区域的聚簇概率,基于所述聚簇概率计算聚簇中每个像素属于主体区域的像素概率,以本申请所述的计算出来的显著度作为属于主体区域的概率,可以有效、精确的检测待处理图像中的主体区域。本申请以所述聚簇与其他聚簇距离和与总和的比值作为聚簇的显著度,用于表述聚簇属于主体区域的概率,更加符合实际用户感知图像中商品主体的情况,使得处理结果更加精确、有效。The image area detecting method and apparatus provided by the present application adopts a unique mixed feature vector for each pixel in the image. The mixed feature vector includes a gradient feature in addition to the color feature of the pixel point. When calculating the pixel point, the information around the pixel point is simultaneously considered, and the feature value of the pixel point can be more accurately established, so that when the feature space is mixed The distance between the foreground and the background regions is similar to that of the mixed feature vectors of the two points. The distance is greatly increased, and the foreground and background areas can be effectively distinguished, and the accuracy of the current area detection is improved. Similarly, in a complex background image, the hybrid feature vector described in the present application can combine the color feature and the gradient feature to describe the pixel of the foreground and the pixel of the background into two different clusters, in the Euclidean distance. It is easy to separate the two when calculating. In the present application, the hybrid features are clustered, and the clustering probability that the clusters belong to the body region after clustering is calculated, and the pixel probability of each pixel in the cluster belongs to the body region is calculated based on the clustering probability, as described in the present application. The calculated saliency as a probability of belonging to the subject area can effectively and accurately detect the subject area in the image to be processed. In the present application, the clustering and other clustering distances and the sum of the sums are used as the clustering saliency to express the probability that the cluster belongs to the body region, which is more in line with the actual user's perception of the product subject in the image, so that the processing result More precise and effective.
附图说明DRAWINGS
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请中记载的一些实施例,对于本邻域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings to be used in the embodiments or the prior art description will be briefly described below. Obviously, the drawings in the following description are only It is a part of the embodiments described in the present application. For those of ordinary skill in the art, other drawings can be obtained according to the drawings without any inventive labor.
图1是本申请一种图像区域检测方法一种实施例的流程示意图;1 is a schematic flow chart of an embodiment of an image area detecting method according to the present application;
图2是本申请所述待处理图像边界点邻域窗口提取的示意图;2 is a schematic diagram of a neighborhood window extraction of an image boundary point to be processed according to the present application;
图3是利用本申请所述的一种图像区域检测方法进行主体区域提取的示意图;3 is a schematic diagram of performing body region extraction using an image region detecting method according to the present application;
图4是利用本申请所述的一种图像区域检测方法进行主体区域提取的示意图;4 is a schematic diagram of performing body region extraction using an image region detecting method according to the present application;
图5是本申请所述一种图像区域检测装置的模块结构示意图;FIG. 5 is a schematic structural diagram of a module of an image area detecting apparatus according to the present application; FIG.
图6是本申请所述一种特征计算模块一种实施例的模块结构示意图;6 is a schematic structural diagram of a module of an embodiment of a feature calculation module according to the present application;
图7是本申请所述一种颜色特征模块一种实施例的模块结构示意图;7 is a block diagram showing the structure of an embodiment of a color feature module according to the present application;
图8是本申请所述一种像素概率计算模块一种实施例的模块结构示意图。FIG. 8 is a schematic structural diagram of a module of an embodiment of a pixel probability calculation module according to the present application.
具体实施方式detailed description
为了使本技术邻域的人员更好地理解本申请中的技术方案,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本邻域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都应当属于本申请保护的范围。The technical solutions in the embodiments of the present application are clearly and completely described in the following, in order to better understand the technical solutions in the present application. The embodiments are only a part of the embodiments of the present application, and not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by a person of ordinary skill in the art without creative efforts shall fall within the scope of protection of the present application.
卖家商户上传的商品图像中可以包括一个或者多个主体,例如为了节约商品展示的窗口资源,卖家商户可以将多个图像合并在一张图像中后上传作为某商品的图像。本申请所述的一种图像区域检测方法可以适用于包括一个或者多个商品主体的图像,在所述图像包括多个 主体时,可以将待处理图像划分为多个子图,每个子图可以包括单个主体,然后对每个所述子图采用本申请所述的主体区域提取方法进行处理。具体的所述将包括多个主体的待处理图像进行划分的方法可以采用专利号为CN102567952A,名称为《一种图像分割方法及系统》中所述的图像分割方法。经过上述方法处理后,可以将包括多个主体的商品图像分割成多个包括单个主体的子图。The product image uploaded by the seller merchant may include one or more subjects. For example, in order to save the window resource of the product display, the seller merchant may merge the multiple images into one image and upload the image as an item. An image area detecting method described in the present application may be applied to an image including one or more product bodies, where the image includes a plurality of In the main body, the image to be processed may be divided into a plurality of sub-pictures, each of which may include a single body, and then processed for each of the sub-pictures using the subject area extraction method described in the present application. Specifically, the method for dividing a to-be-processed image including a plurality of subjects may adopt an image segmentation method described in Patent No. CN102567952A, entitled "An Image Segmentation Method and System". After processing by the above method, the product image including the plurality of subjects may be divided into a plurality of sub-pictures including a single body.
下面以包括单个主体的商品图像或者以经过上述图像分割后的子图为例对本申请所述的图像处理方法进行详细的描述。图1是本申请所述一种图像区域检测方法一个实施例的方法流程图,如图1所述,所述方法可以包括:The image processing method described in the present application will be described in detail below by taking an article image including a single subject or a sub-picture divided by the above image as an example. 1 is a flowchart of a method for detecting an image region according to an embodiment of the present invention. As shown in FIG. 1, the method may include:
S1:计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量。S1: Calculating a color feature and a gradient feature of the pixel of the image to be processed, and constructing a mixed feature vector of the image to be processed.
如前所述,本实施例中所述的待处理图像可以为包括一个主体的单独一张商品图像,也可以为经过图像分割出来后的包括单个主体的一个子图。在获取所述待处理图像后,可以基于包括颜色和梯度的特征值构建待处理图像像素点的混合特征,形成混合特征向量。在实际图像信息处理时,通常可以使用局部特征的方式进行每个像素点的特征提取,例如对于某个像素点P来说,可以选取一个邻域窗口W(p),所述的邻域窗口W(p)可以为一个以P点为中心的N*N的正方形区域。所述的N的取值可以根据图像信息处理的精度或者速度等要求进行合理选择,例如可以根据图像尺寸或包括像素点的多少取值为3、5、7、9等奇数。本实施例中所述N可以取值为5,可以在每次计算像素点的混合特征中的颜色特征或者梯度特征时取以P点为中心的5*5的正方形邻域窗口区域。As described above, the image to be processed described in this embodiment may be a single product image including one main body, or may be a sub-picture including a single main body after being divided by the image. After acquiring the image to be processed, the mixed features of the pixel points of the image to be processed may be constructed based on the feature values including the color and the gradient to form a mixed feature vector. In the actual image information processing, feature extraction of each pixel point can usually be performed by using local features. For example, for a certain pixel point P, a neighborhood window W(p) can be selected, and the neighborhood window is selected. W(p) can be a square region of N*N centered at point P. The value of the N may be reasonably selected according to the accuracy or speed of the image information processing, for example, an odd number such as 3, 5, 7, or 9 may be used according to the image size or the number of pixels included. In this embodiment, the value of N may be 5, and a 5*5 square neighborhood window region centered on the P point may be taken each time the color feature or the gradient feature in the mixed feature of the pixel is calculated.
本实施例中所述构建的待处理图像的混合特征可以包括像素点的颜色特征和梯度特征,可以对所述颜色特征和梯度特征等进行预定格式的组合,形成高维度的混合特征向量。具体的实现过程中,所述计算得出待处理图像像素点的颜色特征的处理过程可以包括:The hybrid feature of the image to be processed constructed in this embodiment may include a color feature and a gradient feature of the pixel, and the color feature and the gradient feature may be combined in a predetermined format to form a high-dimensional mixed feature vector. In a specific implementation process, the process of calculating the color feature of the pixel of the image to be processed may include:
S101:如果所述待处理图像不为Lab格式的数据,将所述待处理图像的数据格式转化为Lab格式;S101: If the image to be processed is not data in the Lab format, convert the data format of the image to be processed into a Lab format;
S102:以待处理像素为中心提取所述待处理图像中邻域窗口W(p)的像素点,将所述邻域窗口W(p)中像素点的L、a、b三个通道分别分为K个分组,形成3*K维的颜色特征向量;S102: extract a pixel point of the neighborhood window W(p) in the image to be processed centering on the pixel to be processed, and divide the three channels L, a, and b of the pixel in the neighborhood window W(p) For K groups, form a color feature vector of 3*K dimensions;
S103:将所述邻域窗口W(p)中每个像素点在所述L、a、b三个通道的颜色值累加到所述颜色特征向量所对应的维中,形成所述邻域窗口中待处理像素点的颜色特征。S103: accumulating color values of each of the L, a, and b channels in each of the neighboring window W(p) in a dimension corresponding to the color feature vector to form the neighborhood window. The color characteristics of the pixel to be processed.
在待处理图像颜色特征提取时通常可以包括将L、a、b三个通道分别均匀的量化为K个分组,尽可能的保证每个通道的每个分组的长度相等。 The image color feature extraction of the image to be processed may generally include uniformly quantizing the three channels L, a, and b into K packets, and ensuring that each packet of each channel has the same length as possible.
通常所述待处理图像可以为RGB通道颜色模型的图像信息,所述Lab通道的颜色模型通常指基于人对颜色的感觉建立的并且与光线及设备无关的颜色模型,更加符合人的视觉感知。因而本实施例中采用从Lab空间检测出的图像中的主体区域,更加符合人的感知结果,使得主体区域提取的处理结果更加准确。Generally, the image to be processed may be image information of an RGB channel color model, and the color model of the Lab channel generally refers to a color model based on a person's perception of color and independent of light and equipment, and is more in line with human visual perception. Therefore, in this embodiment, the body region in the image detected from the Lab space is used, which is more in line with the human perception result, so that the processing result of the body region extraction is more accurate.
本实施例中可以将所述待处理图像从RGB通道转化为Lab通道。通常所述的RGB通道包括三个变量的三维颜色向量(R,G,B),如下所示:In this embodiment, the image to be processed can be converted from an RGB channel to a Lab channel. Typically the RGB channel consists of three variable color vector vectors (R, G, B) as follows:
R:红色,0~255的整数,变化值为256;R: red, an integer from 0 to 255, with a change value of 256;
G:绿色,0~255的整数,变化值为256;G: green, an integer from 0 to 255, with a change value of 256;
B:蓝色,0~255的整数,变化值为256。B: Blue, an integer from 0 to 255, with a change value of 256.
所述的Lab通道可以包括如下所示的三个变量:The Lab channel can include three variables as shown below:
L:亮度,0~100的整数,变化值100;L: brightness, an integer from 0 to 100, a change value of 100;
A:从绿色至红色,-128~127的整数,变化值256;A: from green to red, an integer from -128 to 127, with a value of 256;
B:从蓝色至黄色,-128~127的整数,变化值256。B: an integer from -28 to 127 from blue to yellow, with a value of 256.
在所述将待处理图像从RGB通道转化为Lab通道时,可以采用给定的算法进行转化,也可以采用例如Photoshop等软件工具进行转化,在此不做详细论述。然后可以以预先设置的邻域窗口W(p)提取所述Lab通道的待处理图像中的像素点,将所述邻域窗口W(p)中像素点的L、a、b三个通道分别均匀的量化为K个bin(分组)。进一步的可以将L、a、b三个通道量化后的像素点的值拼接在一起可以形成一个3*K维的颜色特征向量,例如所述形成的3*K维的颜色特征向量可以表示为{L1,L2,…LK,a1,a2,…aK,b1,b2,…bK}。本实施例中所述的K的取值可以自定义设置,用来表示对待处理图像颜色空间的一个描述。本申请中如果所述K的取值偏大,那么所述待处理图像在颜色空间会被划分得较细,颜色特征表述的更为准确,相应的计算时间增加;相反若K值取值较小,那么对所述待处理图像在颜色空间的整体划分度较低,颜色特征向量维数较小,可以提高数据处理速度。经过多次实验,本申请提供一种K的取值范围,具体的所述K的取值可以为:6≤K≤16,在上述取值范围内可以保证颜色特征向量可以准确、有效、合适的表述待处理图像的颜色特征。在本实施例中所述K的取值可以取值为6,即可以构建所述邻域窗口中待处理像素点的18维的混合特征向量。最后可以根据所述邻域窗口W(p)中每个像素点在L、a、b三个通道的颜色值,将其累加到该颜色特征向量相对应的维中。例如在5*5一共25个像素点的邻域窗口中,所述25个像素点的Lab颜色值共同构建一个18维的颜色特征向量。具体的,所述25个像素点中每个像素点都会有一组Lab颜色值,以L通道为例,假设第一像素点的L通道的值为10,可以将其映射到所述L通道总共划分的6个bin(分组)中相对应的一个分组中,例如划分到 L1中。第二像素点的L通道值为98,则可以将其划分到L6中。依次类推,将所述邻域窗口W(p)中的25个像素点全部遍历一遍,将相应bin(分组)中的颜色值累加可以得到一个所述邻域窗口W(p)中待处理像素点总的L、a、b颜色特征的分布向量。When the image to be processed is converted from the RGB channel to the Lab channel, the conversion may be performed by using a given algorithm, or may be converted by using a software tool such as Photoshop, and will not be discussed in detail herein. Then, the pixel points in the image to be processed of the Lab channel may be extracted by using a preset neighborhood window W(p), and the three channels of L, a, and b of the pixel in the neighborhood window W(p) are respectively The uniform quantization is K bins. Further, the values of the pixel points quantized by the three channels L, a, and b can be spliced together to form a 3*K-dimensional color feature vector. For example, the formed 3*K-dimensional color feature vector can be expressed as {L1, L2, ... LK, a1, a2, ... aK, b1, b2, ... bK}. The value of K described in this embodiment can be customized to indicate a description of the color space of the image to be processed. In the present application, if the value of the K is too large, the image to be processed is divided into finer colors in the color space, the color feature is more accurately expressed, and the corresponding calculation time is increased; Small, then the overall division degree of the image to be processed in the color space is low, and the dimension of the color feature vector is small, which can improve the data processing speed. After many experiments, the present application provides a range of values of K. The specific value of the K may be: 6 ≤ K ≤ 16, which can ensure that the color feature vector can be accurate, effective, and suitable within the above range of values. Represents the color characteristics of the image to be processed. In this embodiment, the value of K may be a value of 6, that is, an 18-dimensional mixed feature vector of a pixel to be processed in the neighborhood window may be constructed. Finally, according to the color values of the three channels L, a, b of each pixel in the neighborhood window W(p), it is added to the corresponding dimension of the color feature vector. For example, in a neighborhood window of 5*5 and 25 pixels, the Lab color values of the 25 pixels jointly construct an 18-dimensional color feature vector. Specifically, each of the 25 pixels has a set of Lab color values, and the L channel is taken as an example. If the value of the L channel of the first pixel is 10, it can be mapped to the L channel. One of the corresponding six bins (grouping), for example, divided into L1. The L pixel value of the second pixel is 98, which can be divided into L6. By analogy, all 25 pixels in the neighborhood window W(p) are traversed once, and the color values in the corresponding bins are grouped to obtain a pixel to be processed in the neighborhood window W(p). The distribution vector of the total L, a, b color features.
计算完成当前邻域窗口中待处理像素点的颜色特征向量后,可以按照一定方向一次移位一个像素点,然后按照上述方式再次提取邻域窗口的像素点,计算新的邻域窗口中待处理像素点的颜色特征向量。依次计算得出所述待处理图像中所有像素点的颜色特征向量,获取所述待处理图像像素点的颜色特征。After calculating the color feature vector of the pixel to be processed in the current neighborhood window, one pixel can be shifted once in a certain direction, and then the pixel of the neighborhood window is extracted again according to the above manner, and the new neighborhood window is calculated to be processed. The color feature vector of the pixel. The color feature vectors of all the pixels in the image to be processed are sequentially calculated, and the color features of the pixels of the image to be processed are acquired.
需要说明的是,本申请中所述的邻域窗口中的待处理像素点通常为所述设置的正方形邻域窗口的中心点。对于所述待处理图像中的非边界点像素点,可以一次提取一个正方形的邻域窗口。对于边界点或者靠近边界点不能满足正方形邻域窗口提取的像素点,则仍然按照所述邻域窗口预先设置的提取规格,即可以以所述边界点像素点或者靠近边界的像素点为中心、以所述邻域窗口实际在所述待处理图像中覆盖的像素点进行计算。图2是本申请所述待处理图像边界点邻域窗口提取的示意图。如图2中所示,例如设置的邻域窗口提取规则为5*5的正方形区域,对于某边界点的非角点来说,以所述边界点的像素点P1为邻域窗口中心提取到的像素点规格为5*3,相应的,对于所述待处理图像的角点P2,则提取到的像素点的规格为3*3。It should be noted that the pixel to be processed in the neighborhood window described in the present application is usually the center point of the set square neighborhood window. For a non-boundary point pixel in the image to be processed, a square neighborhood window may be extracted at one time. For a boundary point or a pixel point that is not close to the boundary point and cannot be extracted by the square neighborhood window, the extraction specification preset according to the neighborhood window may be used, that is, the pixel point of the boundary point or the pixel point close to the boundary may be centered. The calculation is performed on the pixel points that the neighborhood window actually covers in the image to be processed. FIG. 2 is a schematic diagram of a neighborhood window extraction of an image boundary point to be processed according to the present application. As shown in FIG. 2, for example, the set neighborhood window extraction rule is a 5*5 square region, and for a non-corner point of a certain boundary point, the pixel point P1 of the boundary point is extracted as a neighborhood window center. The pixel size specification is 5*3. Correspondingly, for the corner point P2 of the image to be processed, the extracted pixel point has a specification of 3*3.
本申请中所述的混合特征可以包括待处理图像的梯度特征。本实施例中可以采用HoG特征进行梯度特征提取,形成待处理图像中每个像素点M维的梯度特征。通常所述的梯度的含义可以包括图像中每个像素点与邻近像素点的差异,在构建为梯度特征后可以用于检测颜色不明显的区域。本实施例中可以将待处理图像从RGB颜色通道转化为灰度图,这样简化梯度特征的复杂性。具体的在实现方式上可以采用HoG特征进行梯度特征提取,获取预先设置的邻域窗口W(p)中像素点的梯度方向和梯度值,然后可以将所述邻域窗口W(p)中包括所有像素点梯度方向的总梯度方向值分割为M个bin(分组),例如将180度的总梯度方向分割为12个bin(分组),那么每个bin代表的是一个15度的范围。最后可以根据所述邻域窗口W(p)中每个像素点的梯度值,使用线性插值的方法累加到对应的bin(分组)中,形成邻域窗口中待处理像素点的一个M维的梯度特征向量,例如本实施例中的12维的梯度特征向量,可以表示为{g1,g2,…g12}。例如若待处理像素点的邻域窗口W(p)中某一点的梯度方向为44度,梯度值为10,那么该梯度方向为44度所属的bin(分组)为g3,与前述颜色特征计算方式类似,可以将44度的梯度值10累加值所属的分组g3中。遍历所述邻域窗口所有像素点的梯度方向和梯度值,计算得到所述邻域窗口待处理像素点的梯度特征。同样计算一个邻域窗口后可以移位一个像素点,继续计算下一个待处理像素点的梯度特征。 依次计算完成所述待处理图像所有像素点的梯度特征,具体的可以参照上述颜色特征的计算方式,在此不做赘述。The blending features described in this application can include gradient features of the image to be processed. In this embodiment, the HoG feature may be used to perform gradient feature extraction to form a gradient feature of M-dimensionality of each pixel in the image to be processed. Generally, the meaning of the gradient may include the difference between each pixel in the image and the adjacent pixel, which can be used to detect an area where the color is not obvious after being constructed as a gradient feature. In this embodiment, the image to be processed can be converted from an RGB color channel to a grayscale image, thus simplifying the complexity of the gradient feature. Specifically, in the implementation manner, the HoG feature may be used to perform gradient feature extraction, and obtain a gradient direction and a gradient value of the pixel in the preset neighborhood window W(p), and then the neighbor window W(p) may be included. The total gradient direction value of all pixel point gradient directions is divided into M bins, for example, the total gradient direction of 180 degrees is divided into 12 bins, and each bin represents a range of 15 degrees. Finally, according to the gradient value of each pixel in the neighborhood window W(p), a linear interpolation method is used to accumulate the corresponding bin (packet) to form an M-dimensional of the pixel to be processed in the neighborhood window. The gradient feature vector, such as the 12-dimensional gradient feature vector in this embodiment, can be expressed as {g1, g2, ... g12}. For example, if the gradient direction of a point in the neighborhood window W(p) of the pixel to be processed is 44 degrees and the gradient value is 10, the bin direction to which the gradient direction is 44 degrees belongs to g3, and the color feature calculation is performed. In a similar manner, a gradient value of 44 degrees can be accumulated in the group g3 to which the value belongs. The gradient direction and the gradient value of all the pixel points of the neighborhood window are traversed, and the gradient feature of the pixel to be processed of the neighborhood window is calculated. Similarly, after calculating a neighborhood window, one pixel can be shifted, and the gradient feature of the next pixel to be processed is continuously calculated. The gradient features of all the pixels of the image to be processed are calculated in sequence, and the calculation manner of the above-mentioned color features may be referred to, and details are not described herein.
在计算得出所述待处理图像像素点的颜色特征和梯度特征后,可以构建所述待处理图像的混合特征向量。具体的所述构建待处理图像的混合特征向量可以包括将所述待处理图像每个像素点的K维颜色特征和M维梯度特征进行拼接组合,形成对于像素点的(K+M)维的混合特征向量。例如本实施例中可以将18维的颜色特征与12维的梯度特征的值按序拼接组合,前面18维数据为颜色特征,后面12维数据为梯度特征,可以表示为{L1,L2,…L6,a1,a2,…a6,b1,b2,…b6,g1,g2,…g12}。当然,如果所述待处理图像的大小为[W,H],其中W为所述待处理图像的宽度,H为所述待处理图像的高度,单位均为像素点,那么通过上述方法可以构建所述待处理图像W*H个(K+M)维的混合特征向量。After calculating the color feature and the gradient feature of the pixel of the image to be processed, the mixed feature vector of the image to be processed may be constructed. Specifically, the constructing the mixed feature vector of the image to be processed may include splicing and combining the K-dimensional color feature and the M-dimensional gradient feature of each pixel of the image to be processed to form a (K+M) dimension for the pixel. Mixed feature vector. For example, in this embodiment, the 18-dimensional color feature and the 12-dimensional gradient feature value can be sequentially combined and stitched. The first 18-dimensional data is a color feature, and the latter 12-dimensional data is a gradient feature, which can be expressed as {L1, L2,... L6, a1, a2, ... a6, b1, b2, ... b6, g1, g2, ... g12}. Of course, if the size of the image to be processed is [W, H], where W is the width of the image to be processed, and H is the height of the image to be processed, and the unit is a pixel point, then the method can be constructed by the above method. The mixed feature vector of the W*H (K+M) dimension of the image to be processed.
在本申请中计算像素点的颜色特征和梯度特征时考虑计算到了每个待处理像素点周围像素点的信息,可以更加准确的建立像素点的特征值,使得混合特征空间时前景和背景区域相近的两个点的混合特征向量的距离比仅仅使用颜色特征的距离大大增加,可以有效的区分前景和背景相近的区域,提高主体区域检测的精准度。When calculating the color feature and the gradient feature of the pixel in the present application, considering the calculation of the information of the pixel around each pixel to be processed, the feature value of the pixel can be more accurately established, so that the foreground and the background region are similar when the feature space is mixed. The distance between the mixed feature vectors of the two points is greatly increased compared with the distance using only the color features, and the foreground and background regions can be effectively distinguished, and the accuracy of the detection of the main body region is improved.
S2:对所述混合特征向量进行聚类,获取聚类后的聚簇。S2: Clustering the mixed feature vectors to obtain clusters after clustering.
前述中所述待处理图像尺寸为[W,H]的商品图像会产生W*H个(K+M)维的混合特征向量。本申请中为了提高计算效率,可以对这些特征向量进行聚类。本实施例中所采用的聚类算法采可以为Kmeans聚类算法。所述的Kmeans聚类算法具体的操作过程主要可以包括:The product image of the image size [W, H] to be processed in the foregoing may generate a mixed feature vector of W*H (K+M) dimensions. In order to improve computational efficiency in the present application, these feature vectors can be clustered. The clustering algorithm used in this embodiment may be a Kmeans clustering algorithm. The specific operation process of the Kmeans clustering algorithm may mainly include:
S201:从所述W*H个(K+M)维的混合特征向量中随机选取L个混合特征向量作为初始聚类中心。在具体的实施例中,所述L的取值范围可以经过试验选取合适的值,通常所述L取值太大会导致计算时间较长,L太小则无法将特征空间划分得比较精细。S201: randomly select L mixed feature vectors from the mixed feature vectors of the W*H (K+M) dimensions as an initial cluster center. In a specific embodiment, the value range of the L may be tested to select a suitable value. Generally, the L value is too large, which may result in a long calculation time. If the L is too small, the feature space may not be divided finely.
S202:遍历所有W*K个混合特征向量,分别计算每一个混合特征向量与当前聚类中心之间的距离。本实施例中所述的距离采用的为欧式距离,例如两个混合特征向量分别为p和q,其中q为随机选取的当前聚类中心,那么所述混合特征向量与所述当前聚类中心q之间的欧式距离D(p,q)可以为:S202: Traverse all W*K mixed feature vectors, and calculate the distance between each mixed feature vector and the current cluster center respectively. The distances described in this embodiment are Euclidean distances. For example, two mixed feature vectors are p and q, respectively, where q is a randomly selected current cluster center, then the mixed feature vector and the current cluster center. The Euclidean distance D(p,q) between q can be:
D(p,q)=||(p1-q1)2+(p2-q2)2……+(p(K+M)-q(K+M))2||D(p,q)=||(p 1 -q 1 ) 2 +(p 2 -q 2 ) 2 ......+(p (K+M) -q (K+M) ) 2 ||
S203:对于每个混合特征向量,计算其与所述选取的L个初始聚类中心的距离,所述混合特征向量属于与所述L个初始聚类中心距离最小的聚簇。经过一轮计算分类后可以将混合特征向量合理的划分到距离最近的所述L个初始聚类中心的聚簇中。S203: Calculate, for each mixed feature vector, a distance from the selected L initial cluster centers, where the mixed feature vector belongs to a cluster with the smallest distance from the L initial cluster centers. After a round of calculation and classification, the mixed feature vector can be reasonably divided into the clusters of the L initial cluster centers closest to each other.
S204:更新每个聚簇的聚类中心。将待处理图像中每个像素点划分到对应的聚簇后,可 以更新每个聚簇的聚类中心。本实施例中具体的更新计算方法可以包括计算所述每个聚簇中所有混合特征向量在每一维上的平均值,然后将所述计算得到的每一维的平均值作为该聚簇新的聚类中心。S204: Update the cluster center of each cluster. After each pixel in the image to be processed is divided into corresponding clusters, To update the cluster center of each cluster. The specific update calculation method in this embodiment may include calculating an average value of each mixed feature vector in each cluster in each dimension, and then using the calculated average value of each dimension as the cluster new Cluster center.
上述所述的S201~S204为一次聚类的过程,本申请中可以反复聚类计算上述为每个像素点划分聚簇和更新聚簇中心的步骤,直到所述聚簇的聚类中心不再进行较大幅度的移动(该移动的幅度阈值可以根据需求进行设置)或者所述聚类计算的次数达到预置计算要求为止。具体例如在本实施例中,可以设置所述混合特征向量聚类次数为1000次,或者,新的聚簇的聚类中心与该聚簇上一次的聚类中心之间的欧式距离小于0.5,如表示为上一次的聚类中心为Old_C,新的聚类中心为New_C,那么聚类计算的停止条件可以设置为D(Old_C,New_C)<0.5。The above-mentioned S201-S204 is a process of one-time clustering. In the present application, the steps of dividing the clustering and updating the clustering center for each pixel point may be repeatedly performed by clustering until the clustering center of the clustering is no longer A larger amount of movement is performed (the amplitude threshold of the movement can be set according to requirements) or the number of times the cluster calculation reaches the preset calculation requirement. For example, in this embodiment, the number of clustering of the mixed feature vector may be set to 1000 times, or the distance between the cluster center of the new cluster and the cluster center of the cluster is less than 0.5. If the cluster center is the old_C and the new cluster center is New_C, the stop condition of the cluster calculation can be set to D (Old_C, New_C) < 0.5.
本实施例中将混合特征向量进行聚类,形成L个聚簇,可以将所述待处理图像中的大量像素点的混合特征计算量缩小至L个聚簇的计算,可以提高后续图像区域检测的进一步计算速率,提高整体图像信息处理效率。In this embodiment, the mixed feature vectors are clustered to form L clusters, and the calculation of the mixed features of the plurality of pixel points in the image to be processed can be reduced to the calculation of L clusters, thereby improving subsequent image area detection. The further calculation rate improves the overall image information processing efficiency.
S3:根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率。S3: Calculate a clustering probability of the cluster according to a predetermined rule, and calculate a pixel probability of the pixel point in the cluster based on the clustering probability.
经过上述步骤处理后,所述待处理图像在本申请所述的(K+M)维的混合特征向量空间中聚类成了L个聚簇,其中所述L个聚簇中每个聚簇内的像素点在所述特征空间上是相近的。本申请中可以以每个所述聚簇为单元计算每个聚簇属于主体区域的聚簇概率,然后进一步的基于所述聚簇的聚簇概率计算所述聚簇中所有像素点属于所述主体区域的像素概率。本实施例中可以采用每个聚簇在整个所述待处理图像中的显著度来描述所述每个聚簇属于主体区域的概率。具体的所述根据预定规则计算所述聚簇的聚簇概率可以包括:After processing in the foregoing steps, the to-be-processed image is clustered into L clusters in the (K+M)-dimensional mixed feature vector space described in the present application, wherein each cluster of the L clusters The pixels within are similar in the feature space. In the present application, a clustering probability that each cluster belongs to a body region may be calculated in units of each of the clusters, and then all pixel points in the cluster are calculated based on the clustering probability of the clustering. The pixel probability of the body area. In this embodiment, the degree of saliency of each cluster in the entire image to be processed may be used to describe the probability that each cluster belongs to the body region. Specifically, calculating the clustering probability of the cluster according to a predetermined rule may include:
计算所述L个聚簇中每个聚簇Ci与其他聚簇的距离和D(Ci),以所述聚簇和D(Ci)与所有聚簇的所述距离和的总和的比值作为所述聚簇Ci的聚簇概率。Calculating a distance and D(Ci) of each cluster Ci in the L clusters from other clusters, using a ratio of the clusters and D(Ci) to the sum of the distance sums of all clusters The clustering probability of the cluster Ci.
本实施例中假设经过聚类后得到的所述L个聚簇的聚簇中心分别为C1、C2、…、CL,本实施例聚簇的显著度可以采用由与其他所有聚簇的距离之和与总和的比值表示。那么对于任意聚簇Ci来说,1≤i≤L,本实施例提供一种计算所述聚簇中每个聚簇与其他聚簇的距离和的方法,具体的该聚簇的Ci与其他聚簇的距离和D(Ci)可以采用下式公式(1)计算得出: In this embodiment, it is assumed that clustering centers of the L clusters obtained after clustering are respectively C1, C2, ..., CL, and the degree of saliency of the cluster in this embodiment may be adopted by distance from all other clusters. And the ratio of the sum to the sum. Then, for any cluster Ci, 1≤i≤L, the embodiment provides a method for calculating the distance sum of each cluster and other clusters in the cluster, specifically the Ci of the cluster and other The cluster distance and D(Ci) can be calculated by the following formula (1):
Figure PCTCN2016073274-appb-000001
Figure PCTCN2016073274-appb-000001
上式中,L为聚簇的个数,如本实施例中设置的120,||ci,cj||为当前聚簇Ci的聚簇中心的混合特征向量与其他聚簇的聚簇中心混合特征向量的欧氏距离。一般的,所述两个聚簇之间混合特征向量差距越大,两个聚簇中心之间的欧氏距离越大。若某个聚簇与其他聚簇的欧式距离整体上都较大,可以表示该聚簇与其他聚簇的区别显著性越高,则越有可能接近待处理图像的主体区域,相应的计算得到的与其他聚簇的距离之和D(Ci)值也越大。在本实施中计算所述距离和的方法中加入了因子Wj,所述Wj可以为根据当前聚簇Ci所包括的像素点设置的权重。本实施例中一般的,所述聚簇中包括的像素点的个数越多,那么其对应显著度值的贡献也越大。因此,所述Wj可以根据聚簇中所包括的像素点进行设置。例如可以设置为聚簇所包括的像素点个数,或者当前聚簇所包括的像素点个数与所述待处理图像总像素点个数的比值等,具体的可以根据需求进行设置。这样,在计算所述聚簇的距离和时加入所述聚簇的权重Wj,将所述聚簇中所包括的像素点个数计算在内,在一些应用场景中更加符合实际图像主体区域的特性,在该类应用场景中可以使提取主图区域的计算结果更加准确。In the above formula, L is the number of clusters, as set in the embodiment 120, ||c i , c j || is the mixed feature vector of the cluster center of the current cluster Ci and the clustering of other clusters The Euclidean distance of the center blending feature vector. Generally, the larger the difference of the mixed feature vectors between the two clusters, the larger the Euclidean distance between the two cluster centers. If the distance between a cluster and other clusters is larger overall, it can be said that the higher the difference between the cluster and other clusters, the more likely it is to approach the body region of the image to be processed, and the corresponding calculation is obtained. The sum of the distances from other clusters is also larger. In the method for calculating the distance sum in the present embodiment, a factor Wj is added, and the Wj may be a weight set according to a pixel point included in the current cluster Ci. In the embodiment, generally, the more the number of pixels included in the cluster, the greater the contribution of the corresponding saliency value. Therefore, the Wj can be set according to the pixel points included in the cluster. For example, it may be set to the number of pixels included in the cluster, or the ratio of the number of pixels included in the current cluster to the total number of pixels of the image to be processed, etc., and may be specifically set according to requirements. In this way, when the distance of the cluster is calculated, the weight Wj of the cluster is added, and the number of pixels included in the cluster is counted, which is more in line with the actual image body area in some application scenarios. Features, in this type of application scenario, the calculation result of extracting the main map area can be more accurate.
在得到每个聚簇在所述待处理图像中的显著度后,可以进一步根据所述显著度计算每个聚簇属于所述主体区域的聚簇概率。本实施例中可以以所述聚簇和D(Ci)与所有聚簇的所述距离和的总和的比值作为所述聚簇Ci属于所述主体区域的聚簇概率,具体的可以采用下式(2)计算得出:After obtaining the saliency of each cluster in the image to be processed, the clustering probability that each cluster belongs to the body region may be further calculated according to the saliency. In this embodiment, the ratio of the cluster and D(Ci) to the sum of the distance sums of all the clusters may be used as the clustering probability that the cluster Ci belongs to the body region, and the specific formula may be (2) Calculated:
Figure PCTCN2016073274-appb-000002
Figure PCTCN2016073274-appb-000002
上述中∑1≤j≤LD(cj)为计算得出的所有聚簇的聚簇和的总和,可以采用当前聚簇的距离和与所述总和的比值作为所述当前聚簇属于所述主体区域的聚簇概率。由于聚类后的聚簇中混合特征向量值较为接近,在本申请的一种实施例中可以认为该聚簇中像素点属于主体区域的像素概率等价于该聚簇属于主体区域的聚簇概率,这样可以根据所述聚簇的概率得到每个像素点的一个概率值。因此,本申请的一种实施例中,所述基于所述聚簇概率计算所述聚簇中像素点的像素概率可以包括:The above middle ∑ 1 j ≤ L D(c j ) is the sum of the calculated cluster sums of all the clusters, and the distance of the current cluster and the ratio of the sum may be used as the current cluster The clustering probability of the subject area. Since the mixed feature vector values in the clustered clusters are relatively close, in one embodiment of the present application, it can be considered that the pixel probability of the pixel points belonging to the body region in the cluster is equivalent to the cluster of the cluster belonging to the body region. Probability, such that a probability value for each pixel can be derived from the probability of the cluster. Therefore, in an embodiment of the present application, the calculating a pixel probability of a pixel point in the cluster based on the clustering probability may include:
S301:所述聚簇中像素点的像素概率可以为该像素点所属聚簇的聚簇概率。S301: The pixel probability of the pixel in the cluster may be a clustering probability of the cluster to which the pixel belongs.
在本申请其他实施例中,聚簇中的像素点可能分布于所述待处理图像的分散的其他区域中,本申请为使提取的主体区域具有的紧凑特性,提取的主体区域更加准确,可以再次计算 每个聚簇中每个像素点属于主体区域的像素概率。在此,本申请可以设置第二邻域窗口W(p)’,可以参照上述计算颜色特征的方式以像素点P为中心提取所述第二邻域窗口W(p)’的像素点,所述第二邻域窗口W(p)’中的某个像素点q的概率为该像素点q所属聚簇的聚簇概率,在此以P(q)表示,则另一种实施例中所述基于所述聚簇的概率计算所述聚簇中像素点属于所述主体区域的概率可以包括:In other embodiments of the present application, the pixels in the cluster may be distributed in other regions of the image to be processed. In the present application, the extracted body region has a compact feature, and the extracted body region is more accurate. Calculate again The pixel probability of each pixel in each cluster belonging to the body area. Here, the present application may set a second neighborhood window W(p)', and may extract the pixel points of the second neighborhood window W(p)' centering on the pixel point P by referring to the manner of calculating the color feature. The probability of a certain pixel point q in the second neighborhood window W(p)' is the clustering probability of the cluster to which the pixel point q belongs, which is represented by P(q), and in another embodiment Calculating the probability that the pixel points in the cluster belong to the body region based on the probability of the clustering may include:
S302:以待求像素点p为中心提取第一邻域窗口W(p)’的像素点,采用下式计算所述待求像素点p属于所述主体区域的像素概率Sal(p):S302: extract a pixel point of the first neighborhood window W(p)' centering on the pixel to be obtained p, and calculate a pixel probability Sal(p) of the pixel to be obtained p belonging to the body region by using the following formula:
Figure PCTCN2016073274-appb-000003
Figure PCTCN2016073274-appb-000003
上式中,P(q)为所述第一邻域窗W(p)’内的像素点q所属的聚簇属于主体区域的聚簇概率,t为待求像素点p所属的聚簇中像素点的个数,σ为设置的一个平滑参数,可以表示当前计算的像素点p的结果受到周围像素点影响的大小。若σ取值较大,可以表示像素点p的计算结果容易受周围像素点的影响,反之不容易受到周围像素点的影响。该σ值可以根据经验或者结果的预估进行合理设置,一般来说,对于网站产品销售的图像来说,σ可以取值偏小,例如本实施例中具体的可以取值为0.17。若在自然场景下的图像(通常为非商品图像),所述σ的取值可以偏大,例如可以取值为0.25。In the above formula, P(q) is the clustering probability of the cluster to which the pixel point q in the first neighborhood window W(p)' belongs belongs to the body region, and t is the cluster to which the pixel point p to be sought belongs. The number of pixels, σ is a set of smoothing parameters, which can indicate the size of the currently calculated pixel point p is affected by the surrounding pixels. If the value of σ is large, it can be said that the calculation result of the pixel point p is easily affected by the surrounding pixel points, and vice versa. The σ value can be set according to the experience or the estimation of the result. Generally, for the image sold by the website product, the σ value may be small, for example, the specific value in the embodiment may be 0.17. If the image is in a natural scene (usually a non-commodity image), the value of σ may be too large, for example, may be 0.25.
上述中所述的第一邻域窗口W(p)’的设置可以与前述颜色特征提取时设置的邻域窗口相同,例如可以设置为5*5的正方形邻域窗口。这样,在计算所述待处理图像中像素点的像素概率时可以以待求像素点为中心所述第一邻域窗口W(p)’如5*5的像素点进行计算,遍历所述第一邻域窗口W(p)’中所有像素点的概率可以计算得到该待求像素点p属于所述主体区域的像素概率。The setting of the first neighborhood window W(p)' described above may be the same as the neighborhood window set in the foregoing color feature extraction, for example, a square neighborhood window of 5*5 may be set. In this way, when calculating the pixel probability of the pixel in the image to be processed, the pixel of the first neighborhood window W(p)′, such as 5*5, may be calculated centering on the pixel to be obtained, and the method is traversed. The probability of all pixel points in a neighborhood window W(p)' can be calculated as the pixel probability that the pixel point p to be sought belongs to the body region.
通过上述S302所述的像素点属于主体区域概率计算方法,可以计算得到所述待处理图像中每个像素点属于主体区域的像素概率,并且该概率值采用了所述第一邻域窗口W(p)’中像素点的概率值进行平滑计算得出,可以提高最终的提取结果的准确性。The pixel point according to the above S302 belongs to the body region probability calculation method, and the pixel probability that each pixel in the image to be processed belongs to the body region can be calculated, and the probability value adopts the first neighborhood window W ( The probability value of the pixel point in p)' is smoothed and calculated, which can improve the accuracy of the final extraction result.
S4:基于所述像素概率对所述待处理图像进行检测,获取目标区域。S4: The image to be processed is detected based on the pixel probability, and the target area is acquired.
在计算完成所述待处理图像每个像素属于所述主体区域的像素概率后,可以进行主体区域与背景区域的分离,提取获取所述待处理图像中的目标区域。本申请中所述的目标区域可以为所述待处理图像中的主体区域(前景区域),在其他的实施例中,所述目标区域也可以 为背景区域,即可以检测获取待处理图像的背景区域、本申请一种实施方式中,所述基于所述像素概率对所述待处理图像进行检测获取目标区域具体的可以包括:After calculating the pixel probability that each pixel of the image to be processed belongs to the body region, the body region and the background region may be separated, and the target region in the image to be processed is extracted and acquired. The target area described in the present application may be a body area (foreground area) in the image to be processed, and in other embodiments, the target area may also be For the background area, that is, the background area of the image to be processed can be detected. In an embodiment of the present application, the detecting the image to be processed based on the pixel probability to obtain the target area may include:
S401:将所述待处理图像中像素点的像素概率值符合判断阈值PV要求的像素点作为所述待处理图像的目标区域。S401: A pixel point that meets a pixel probability value of a pixel in the image to be processed according to a determination threshold PV is used as a target area of the image to be processed.
具体的例如在检测主体区域的实施过程中,例如可以预先设置像素点概率的判定阈值PV如0.85,然后可以将所述待处理图像中所述像素点的像素概率的值大于0.85的像素点提取出来,作为所述待处理图像的主体区域。本申请所述预定判断阈值取值过小会导致提取较多的非主体区域的像素点,取值过大则会降低提取到的主体区域图像的完整性,本实施例提供一种所述判断阈值的取值范围,具体的所述预定判断阈值PV的取值范围可以为:0.8≤PV≤0.95。上述S401中所述像素点的像素概率优选的方式为采用所述第一邻域窗口W(p)’中像素点的概率值进行平滑计算得出的概率值。Specifically, for example, in the implementation process of detecting the body region, for example, a determination threshold PV of the pixel point probability, such as 0.85, may be set in advance, and then the pixel point of the pixel probability of the pixel to be processed may be greater than 0.85. Coming out as the body area of the image to be processed. If the value of the predetermined judgment threshold is too small, the pixel points of the non-subject area are extracted. If the value is too large, the integrity of the extracted image of the body area is reduced. This embodiment provides a determination. The value range of the threshold, the specific predetermined threshold value PV may be: 0.8≤PV≤0.95. The pixel probability of the pixel point in the above S401 is preferably a probability value obtained by smoothing the probability value of the pixel point in the first neighborhood window W(p)'.
当然,在检测背景区域的实施例中,可以设置满足判断为背景区域的判断阈值PV的值,具体的可以根据实际场景应用进行确定。Certainly, in the embodiment for detecting the background area, a value that satisfies the determination threshold PV determined as the background area may be set, and the specific determination may be performed according to the actual scenario application.
本申请还提供另一种优选的实施例,在所述另一种实施例中,所述基于所述像素点的像素概率对所述待处理图像进行检测获取目标具体的可以包括:The present application further provides another preferred embodiment. In the another embodiment, the detecting, by the pixel probability of the pixel, the target to be processed may be:
S4021:将所述待处理图像中像素点属于主体区域的概率值大于第一阈值PF的像素点作为种子像素点;S4021: The pixel point in the image to be processed that belongs to the body region and whose probability value is greater than the first threshold PF is used as the seed pixel point;
S4022:以所述种子像素点为中心计算与周围第二邻域窗口中像素点的欧式距离;S4022: Calculate a Euclidean distance from a pixel in a surrounding second neighborhood window centering on the seed pixel point;
S4023:将所述欧式距离小于第二阈值的像素点作为新的种子像素点;S4023: The pixel point whose Euclidean distance is less than the second threshold is used as a new seed pixel point;
S2044:遍历所有所述种子像素点与周围所述第二邻域窗口中像素点的欧式距离并做出判断,将所述计算得出的种子像素点作为所述待处理图像的目标区域。S2044: traverse the Euclidean distance of all the seed pixel points and the pixel points in the surrounding second neighborhood window and make a determination, and use the calculated seed pixel point as the target area of the to-be-processed image.
本实施例中,所述像素点属于主体区域的像素概率优选的可以为所述像素点所属聚簇的聚簇概率。另外,所述的第一阈值PF和第二阈值以及所述第三邻域窗口可以根据实际数据处理需求进行设置,例如所述第一阀PF值同样可以设置为0.85或者选取为聚簇概率中值较高的值,所述第二阈值可以设置为0.5。如上述预定判断阈值,本申请所述第一阈值PF取值过小会导致提取较多的非主体区域的像素点,取值过大则会降低提取到的主体区域图像的完整性,本实施例提供一种所述第一阈值PF的取值范围,具体的所述第一阈值PF的取值范围可以为:0.8≤PF≤0.95。本实施例中所述的第三邻域窗口一般的为以种子像素点为中心的3*3的八邻近窗口,然后可以根据本申请所述的例如30维的特征混合特征向量进行欧式距离计算。如果所述距离满足第二阈值要求,可以将所述种子周围满足第二阈值要求的像素点作为新的种子像素点,可以认为符合所述第二阈值要求的新的种子像素点同样属于主体区域。当 然,在处理过程中,可以设置将不满足所述第三邻域窗口的像素点作为背景区域。需要说明的,本申请中所述的主体区域通常是连通的,在其他的应用场景中,可以将没有经过第二阈值判断过的像素点设置为背景区域。本实施例中可以根据概率值较大的像素点作为种子像素点,然后不断的遍历周围的邻近点并做出判断,最终得到主体区域。In this embodiment, the pixel probability that the pixel belongs to the body region is preferably a clustering probability of the cluster to which the pixel belongs. In addition, the first threshold PF and the second threshold and the third neighborhood window may be set according to actual data processing requirements, for example, the first valve PF value may also be set to 0.85 or selected as a clustering probability. The value of the higher value, the second threshold can be set to 0.5. If the threshold value of the first threshold PF is too small, the pixel value of the non-subject area is extracted too much, and if the value is too large, the integrity of the extracted image of the body area is reduced. For example, the value range of the first threshold PF is set, and the value of the first threshold PF may be: 0.8≤PF≤0.95. The third neighborhood window described in this embodiment is generally a 3*3 eight-contiguous window centered on the seed pixel point, and then the Euclidean distance calculation can be performed according to the 30-dimensional feature mixed feature vector described in the present application. . If the distance satisfies the second threshold requirement, a pixel point satisfying the second threshold requirement around the seed may be used as a new seed pixel point, and a new seed pixel point that meets the second threshold requirement may be considered to belong to the body area. . when However, during processing, a pixel point that does not satisfy the third neighborhood window may be set as a background area. It should be noted that the body area described in the present application is generally connected. In other application scenarios, a pixel point that has not been judged by the second threshold may be set as a background area. In this embodiment, a pixel point having a larger probability value may be used as a seed pixel point, and then the surrounding neighboring points are continuously traversed and a judgment is made to finally obtain a body region.
当然,本申请所述基于所述像素点的像素概率后,获取目标区域的方式可以包括但不限于本申请所述的实施例,其他的基于本申请所述的方法无需创造性劳动的其他处理方法仍在本申请所述的申请范围内,例如利用测地线距离算法进行主体区域与背景区域分离提取得到的主体区域。Certainly, after the pixel probability of the pixel is used in the present application, the manner of acquiring the target area may include, but is not limited to, the embodiments described in the present application, and other processing methods that do not require creative labor based on the method described in this application. Still within the scope of the application described herein, the body region obtained by separating the body region from the background region is performed, for example, using a geodesic distance algorithm.
本申请提供的一种图像区域检测方法,构建了包括像素点颜色特征和梯度特征的混合特征向量,可以更加准确的建立像素点的特征值,可以有效的区分前景和背景相近的区域,提高主体区域提取的精准度。同样的,在复杂背景图像中,本申请所述的混合特征向量可以很好的结合颜色特征和梯度特征将前景的像素点和背景的像素点描述到两个不同的聚簇中,在欧式距离计算时可以很容易将两者分离。本申请中对混合特征进行聚类,获得聚簇后以所述聚簇与其他聚簇距离和与总和的比值作为聚簇的显著度,用于表述聚簇属于主体区域的概率,更加符合实际用户感知图像中商品主体的情况,使得处理结果更加精确、有效。在实际的应用中,利用本申请所述主体区域提取方法提取待处理图像主体区域的准确率达到了89.62%,召回率达到了88.83%,解决了现有技术中面临复杂度高的图像时主体区域提取准确率低的问题。The image region detecting method provided by the present application constructs a mixed feature vector including pixel color features and gradient features, which can more accurately establish feature values of pixel points, can effectively distinguish foreground and background regions, and improve the subject. The accuracy of regional extraction. Similarly, in a complex background image, the hybrid feature vector described in the present application can combine the color feature and the gradient feature to describe the pixel of the foreground and the pixel of the background into two different clusters, in the Euclidean distance. It is easy to separate the two when calculating. In the present application, the mixed features are clustered, and the clustering and other clustering distances and the sum of the sums are used as the clustering saliency to express the probability that the cluster belongs to the main body, which is more realistic. The user perceives the situation of the product body in the image, making the processing result more accurate and effective. In practical applications, the accuracy of extracting the main body region of the image to be processed by using the main body region extraction method of the present application reaches 89.62%, and the recall rate reaches 88.83%, which solves the problem in the prior art when facing a complex image. The problem of low regional extraction accuracy.
图3、图4分别是利用本申请所述的一种图像区域检测方法进行主体区域提取的示意图,图3、图4从左到右分别是待处理图像、现有算法提取结果和本发明提取结果。如图3所示,选取的是一张前景和背景区域颜色非常相近的图像,从图3中可以看到现有算法在处理这样的图像时无法对该服装中间高亮的部分进行检测,因为此处的颜色非常靠近背景的白色。而本申请的(K+M)维的混合特征向量可以有效的区分出来相似的前景和背景区域。图4选取的是背景复杂的情况,从图4中可以看到现有算法在面对复杂性较高的图像上难以精确提取主体,本申请所述方法采用聚类获取聚簇计算像素点属于主体区域的像素概率,可以有效解决背景上不仅在颜色同时结构上复杂度很高的图像主体提取问题,大大提高检测精度。FIG. 3 and FIG. 4 are schematic diagrams of extracting a main body region by using an image region detecting method according to the present application, and FIG. 3 and FIG. 4 are respectively to-be-processed images, existing algorithm extraction results, and extraction by the present invention from left to right. result. As shown in FIG. 3, an image with a very similar color between the foreground and the background area is selected. It can be seen from FIG. 3 that the existing algorithm cannot detect the highlighted portion of the garment when processing such an image because The color here is very close to the white of the background. The mixed feature vector of (K+M) dimension of the present application can effectively distinguish similar foreground and background regions. FIG. 4 is a case where the background is complicated. It can be seen from FIG. 4 that the existing algorithm is difficult to accurately extract the subject on the image with high complexity. The method of the present application uses the cluster to acquire the cluster to calculate the pixel points. The pixel probability of the main body region can effectively solve the problem of image subject extraction on the background not only in color but also in structure, which greatly improves the detection accuracy.
基于本申请所述的一种图像区域检测方法,本申请还提供一种图像区域检测装置。图5是本申请所述一种图像区域检测装置的模块结构示意图,如图5所示,所述装置可以包括:Based on an image region detecting method described in the present application, the present application further provides an image region detecting device. FIG. 5 is a schematic structural diagram of a module of an image area detecting apparatus according to the present application. As shown in FIG. 5, the apparatus may include:
特征计算模块101,可以用于计算得出待处理图像像素点的颜色特征和梯度特征,并构建所述待处理图像的混合特征向量;The feature calculation module 101 is configured to calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
聚类模块102,可以用于对所述混合特征向量进行聚类,获取聚类后的聚簇; The clustering module 102 can be configured to cluster the mixed feature vectors to obtain clusters after clustering;
聚簇概率模块103,可以用于根据预定规则计算所述聚簇的聚簇概率;The clustering probability module 103 may be configured to calculate a clustering probability of the cluster according to a predetermined rule;
像素概率模块104,可以用于基于所述聚簇的概率计算所述聚簇中像素点的像素概率;a pixel probability module 104, configured to calculate a pixel probability of a pixel point in the cluster based on a probability of the clustering;
检测模块105,可以用于基于所述像素概率对所述待处理图像进行检测,获取目标区域。The detecting module 105 is configured to detect the image to be processed based on the pixel probability to acquire a target area.
在具体的实施过程中,所述的特征计算模块101可以分成多个子模块分别进行相应过程的处理。图6是本申请所述一种特征计算模块101一种实施例的模块结构示意图,如图6所示,所述特征计算模块101可以被设置成包括:In a specific implementation process, the feature calculation module 101 may be divided into multiple sub-modules to perform processing of respective processes. FIG. 6 is a schematic diagram of a module structure of an embodiment of a feature calculation module 101 according to the present application. As shown in FIG. 6, the feature calculation module 101 may be configured to include:
颜色特征模块1011,可以用于计算所述待处理图像像素点的颜色特征;The color feature module 1011 is configured to calculate a color feature of the pixel of the image to be processed;
梯度特征模块1012,可以用于计算所述待处理图像像素点的梯度特征;The gradient feature module 1012 can be configured to calculate a gradient feature of the pixel of the image to be processed;
混合特征模块1013,可以用于将所述颜色特征和梯度特征结合,形成待处理图像的混合特征向量。The blending feature module 1013 can be configured to combine the color features and the gradient features to form a mixed feature vector of the image to be processed.
图7是本申请所述一种特征计算模块1011一种实施例的模块结构示意图,如图7所示,所述颜色特征模块1011可以包括:FIG. 7 is a schematic diagram of a module structure of an embodiment of a feature calculation module 1011 according to the present application. As shown in FIG. 7, the color feature module 1011 may include:
Lab转化模块111,可以用于将所述待处理图像转化为Lab格式的数据;The Lab conversion module 111 can be configured to convert the image to be processed into data in a Lab format;
颜色特征向量模块112,可以用于以待处理像素为中心提取所述待处理图像中邻域窗口的像素点,将所述邻域窗口中像素点的L、a、b三个通道分别分为K个分组,形成3*K维的颜色特征向量;The color feature vector module 112 may be configured to extract pixel points of the neighborhood window in the image to be processed centering on the pixel to be processed, and divide the three channels L, a, and b of the pixel in the neighborhood window into two K groups, forming a color feature vector of 3*K dimensions;
特征计算模块113,可以用于将所述邻域窗口中每个像素点在所述L、a、b三个通道的颜色值累加到所述颜色特征向量所对应的维中,形成所述邻域窗口中待处理像素点的颜色特征。The feature calculation module 113 may be configured to add color values of each of the L, a, and b channels of each pixel in the neighborhood window to a dimension corresponding to the color feature vector to form the neighbor. The color characteristics of the pixels to be processed in the domain window.
经过上述模块处理,可以得到待处理图像的颜色特征。本申请为所述的装置提供一种K的取值范围,具体的所述K的取值可以为:6≤K≤16,在上述取值范围内可以保证本申请装置提取的颜色特征向量准确、有效、合适的表述待处理图像的颜色特征。Through the above module processing, the color characteristics of the image to be processed can be obtained. The present application provides a value range of K for the device, and the specific value of the K may be: 6 ≤ K ≤ 16, and the color feature vector extracted by the device of the present application can be ensured within the above range. Effectively and appropriately express the color characteristics of the image to be processed.
上述所述装置中聚簇概率模块103计算所述聚簇属于主体区域的概率,具体的可以包括:The clustering probability module 103 in the foregoing device calculates the probability that the cluster belongs to the body region, and may specifically include:
距离和计算模块,可以用于计算所述聚簇中每个聚簇与其他聚簇的距离和;a distance and calculation module, which can be used to calculate a distance between each cluster and other clusters in the cluster;
聚簇概率计算模块,可以用于根据所述聚簇和与所有聚簇的所述距离和的总和计算所述聚簇的聚簇概率。The clustering probability calculation module may be configured to calculate a clustering probability of the cluster according to the cluster and the sum of the distance sums of all clusters.
本申请所述一种图像区域检测装置的一种优选实施例中,所述距离计算模块计算所述聚簇中每个聚簇与其他聚簇的距离和具体的可以包括:In a preferred embodiment of the image area detecting apparatus of the present application, the distance calculating module calculates the distance and specificity of each cluster in the cluster from other clusters, and may include:
采用下式计算所述聚簇中每个聚簇与其他聚簇的距离和D(Ci): The distance and D(Ci) of each cluster in the cluster from other clusters are calculated using the following formula:
Figure PCTCN2016073274-appb-000004
Figure PCTCN2016073274-appb-000004
上式中,L为聚簇的个数,||ci,cj||为当前聚簇Ci的聚簇中心的混合特征向量与其他聚簇的聚簇中心混合特征向量的欧氏距离,Wj为根据当前聚簇Ci所包括的像素点设置的权重。In the above formula, L is the number of clusters, ||c i , c j || is the Euclidean distance of the mixed feature vector of the cluster center of the current cluster Ci and the clustered feature vector of the cluster cluster of other clusters, Wj is a weight set according to the pixel points included in the current cluster Ci.
图8是本申请所述一种像素概率模块104一种实施例的模块结构示意图,如图8所示,所述像素概率模块104可以包括下述中的至少一种:8 is a block diagram of an embodiment of a pixel probability module 104 of the present application. As shown in FIG. 8, the pixel probability module 104 may include at least one of the following:
第一概率模块1041,可以用于将像素点所属聚簇的聚簇概率作为该像素点的像素概率;The first probability module 1041 may be configured to use the clustering probability of the cluster to which the pixel belongs to be the pixel probability of the pixel;
第二概率模块1042,可以用于以待求像素点p为中心提取第一邻域窗口W(p)’的像素点,采用下式计算所述待求像素点p的像素概率Sal(p):The second probability module 1042 may be configured to extract a pixel point of the first neighborhood window W(p)′ centering on the pixel point p to be obtained, and calculate a pixel probability Sal(p) of the pixel point p to be obtained by using the following formula: :
Figure PCTCN2016073274-appb-000005
Figure PCTCN2016073274-appb-000005
上式中,P(q)为所述第一邻域窗W(p)’内的像素点q所属的聚簇属于主体区域的概率,t为待求像素点p所属的聚簇中像素点的个数,σ为设置的一个平滑参数。In the above formula, P(q) is the probability that the cluster to which the pixel point q in the first neighborhood window W(p)' belongs belongs to the body region, and t is the pixel point in the cluster to which the pixel point p to be sought belongs The number of σ is a smoothing parameter set.
所述提取模块105可以采取预先设置的不同的提取方式提取待处理图像的主体区域。具体的可以包括下述中的至少一种模块:The extraction module 105 may extract a body region of the image to be processed by using different extraction methods set in advance. Specifically, at least one of the following modules may be included:
第一提取模块,可以用于将所述待处理图像中像素点的像素概率值符合判断阈值PV要求的像素点作为所述待处理图像的目标区域;a first extraction module, configured to use, as a target area of the to-be-processed image, a pixel point in which a pixel probability value of a pixel in the image to be processed meets a determination threshold PV requirement;
第二提取模块,可以用于将所述待处理图像中像素点属于主体区域的概率值大于第一阈值的PF像素点作为种子像素点;还可以用于以所述种子像素点为中心计算与周围第二邻域窗口中像素点的欧式距离;还可以用于将所述欧式距离小于第二阈值的像素点作为新的种子像素点;还可以用于遍历所有所述种子像素点与周围所述第二邻域窗口中像素点的欧式距离并做出判断,将所述计算得出的种子像素点作为所述待处理图像的目标区域。The second extraction module may be configured to use, as a seed pixel point, a PF pixel point whose probability value of the pixel in the image to be processed belongs to the body region is greater than the first threshold; and may also be used to calculate and center the pixel pixel a Euclidean distance of a pixel in the surrounding second neighborhood window; and may also be used to use the pixel point whose Euclidean distance is less than the second threshold as a new seed pixel point; and may also be used to traverse all of the seed pixel points and surrounding areas Determining the Euclidean distance of the pixel in the second neighborhood window and making a judgment, and using the calculated seed pixel as the target area of the image to be processed.
上述所述的一种图像区域检测装置中,所述判断阈值PV的取值范围可以为:0.8≤PV≤0.95;In the image area detecting device described above, the value of the determination threshold PV may be: 0.8 ≤ PV ≤ 0.95;
和/或,and / or,
所述第一阈值PF的取值范围可以为:0.8≤PF≤0.95。The value of the first threshold PF may be: 0.8≤PF≤0.95.
本实施例提供的判断阈值PV或者第一阈值PF的取值范围,可以有效保证主体区域提取的正确、有效性,提高图像尤其所述复杂性较高的图像区域检测的准确性。 The determination threshold PV or the value range of the first threshold PF provided in this embodiment can effectively ensure the correctness and validity of the extraction of the main body region, and improve the accuracy of the image region detection with high image complexity.
利用本申请所述的一种图像区域检测装置,可以在平台型电商网站中用于分离复杂多变的商品图像中的主体区域和背景区域,能有效应对实际图像场景中各种复杂的情况,实现对复杂图像中主体区域进行准确、有效的分离,提高图像检测精确度。An image area detecting device according to the present application can be used in a platform type e-commerce website to separate a body area and a background area in a complex and varied product image, and can effectively cope with various complicated situations in an actual image scene. It can accurately and effectively separate the main body area in complex images and improve the accuracy of image detection.
本申请所述的一种图像区域检测装置可以使用于多种终端设备中,例如用户移动客户端的抠图应用,或者专门用于图像主体或者背景区域提取的客户端或者服务器。通常,所述图像检测装置在进行图像检测,获取目标区域后,可以将所述获取的目标区域的图像进行保存或者显示给用户进行进一步处理。本申请提供一种图像区域检测装置,可以适用于处理用户或者客户端的图像,进行图像检测,获取目标区域。具体的,所述装置可以被设置成,包括:An image area detecting apparatus described in the present application can be used in a variety of terminal devices, such as a mapping application of a user mobile client, or a client or server dedicated to image body or background area extraction. Generally, after performing image detection and acquiring a target area, the image detecting apparatus may save or display the image of the acquired target area to the user for further processing. The present application provides an image area detecting apparatus, which can be applied to process an image of a user or a client, perform image detection, and acquire a target area. Specifically, the device may be configured to include:
第一处理单元,可以用于获取用户/客户端的待处理图像,计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量;The first processing unit may be configured to acquire a to-be-processed image of the user/client, calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
第二处理单元,可以用于对所述混合特征向量进行聚类,获取聚类后的聚簇;还可以用于根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率;a second processing unit, configured to perform clustering on the hybrid feature vector to obtain clustered clusters; and may be further configured to calculate a clustering probability of the cluster according to a predetermined rule, and based on the clustering probability Calculating a pixel probability of a pixel point in the cluster;
输出单元,可以用于基于所述像素概率对所述待处理图像进行获取目标区域,并将所述获取的目标区域存储或者展示于指定位置。The output unit may be configured to acquire a target area of the image to be processed based on the pixel probability, and store or display the acquired target area at a specified location.
本实施例提供的图像去检测装置,可以在客户端或者服务器中有效、精确的提取待处理图片的目标区域,可以提高客户端图片处理用户体验或者客户端/服务器图像信息处理的精确度。The image de-detection device provided in this embodiment can effectively and accurately extract the target area of the to-be-processed picture in the client or the server, and can improve the accuracy of the client-side picture processing user experience or the client/server image information processing.
尽管本申请内容中提到不同图像格式转换、聚类方法、给定公式的计算等之类的描述,但是,本申请并不局限于必须是完全标准的格式转换、聚类方法或者本申请提供的固定公式的情况。本申请中各个实施例所涉及的上述描述仅是本申请中的一些实施例中的应用,在某些标准、方法的基础上略加修改后的处理方法也可以实行上述本申请各实施例的方案。当然,要符合本申请上述各实施例的中所述的处理方法步骤的其他无需创造性的变形,仍然可以实现相同的申请,在此不再赘述。Although the description of the different image format conversion, clustering method, calculation of a given formula, and the like is mentioned in the present application, the present application is not limited to a format conversion, a clustering method, or the present application, which must be completely standard. The case of the fixed formula. The above description of the various embodiments in the present application is only an application in some embodiments of the present application. The slightly modified processing method may also implement the foregoing embodiments of the present application. Program. Of course, the same application can still be implemented without any inventive variation of the processing method steps described in the above embodiments of the present application, and details are not described herein again.
上述实施例阐明的单元或模块,具体可以由计算机芯片或实体实现,或者由具有某种功能的产品来实现。为了描述的方便,描述以上装置时以功能分为各种模块分别描述。当然,在实施本申请时可以把各模块的功能在同一个或多个软件和/或硬件中实现,也可以将实现同一功能的模块由多个子模块或子单元的组合实现。The unit or module illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product having a certain function. For the convenience of description, the above devices are described as being separately divided into various modules by function. Of course, the functions of the modules may be implemented in the same software or software and/or hardware when implementing the present application, or the modules implementing the same functions may be implemented by multiple sub-modules or a combination of sub-units.
本邻域技术人员也知道,除了以纯计算机可读程序代码方式实现控制器以外,完全可以通过将方法步骤进行逻辑编程来使得控制器以逻辑门、开关、专用集成电路、可编程逻辑控 制器和嵌入微控制器等的形式来实现相同功能。因此这种控制器可以被认为是一种硬件部件,而对其内部包括的用于实现各种功能的装置也可以视为硬件部件内的结构。或者甚至,可以将用于实现各种功能的装置视为既可以是实现方法的软件模块又可以是硬件部件内的结构。The neighboring technical staff also knows that in addition to implementing the controller in a purely computer readable program code, the controller can be logically programmed by means of logic gates, switches, ASICs, programmable logic The form of the controller and embedded microcontroller, etc. to achieve the same function. Therefore, such a controller can be considered as a hardware component, and a device for internally implementing it for implementing various functions can also be regarded as a structure within a hardware component. Or even a device for implementing various functions can be considered as a software module that can be both a method of implementation and a structure within a hardware component.
本申请可以在由计算机执行的计算机可执行指令的一般上下文中描述,例如程序模块。一般地,程序模块包括执行特定任务或实现特定抽象数据类型的例程、程序、对象、组件、数据结构、类等等。也可以在分布式计算环境中实践本申请,在这些分布式计算环境中,由通过通信网络而被连接的远程处理设备来执行任务。在分布式计算环境中,程序模块可以位于包括存储设备在内的本地和远程计算机存储介质中。The application can be described in the general context of computer-executable instructions executed by a computer, such as a program module. Generally, program modules include routines, programs, objects, components, data structures, classes, and the like that perform particular tasks or implement particular abstract data types. The present application can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are connected through a communication network. In a distributed computing environment, program modules can be located in both local and remote computer storage media including storage devices.
通过以上的实施方式的描述可知,本邻域的技术人员可以清楚地了解到本申请可借助软件加必需的通用硬件平台的方式来实现。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,移动终端,服务器,或者网络设备等)执行本申请各个实施例或者实施例的某些部分所述的方法。It can be seen from the description of the above embodiments that the technicians in the neighborhood can clearly understand that the present application can be implemented by means of software plus a necessary general hardware platform. Based on such understanding, the technical solution of the present application may be embodied in the form of a software product in essence or in the form of a software product, which may be stored in a storage medium such as a ROM/RAM or a disk. , an optical disk, etc., includes instructions for causing a computer device (which may be a personal computer, mobile terminal, server, or network device, etc.) to perform the methods described in various embodiments of the present application or portions of the embodiments.
本说明书中的各个实施例采用递进的方式描述,各个实施例之间相同或相似的部分互相参见即可,每个实施例重点说明的都是与其他实施例的不同之处。本申请可用于众多通用或专用的计算机系统环境或配置中。例如:个人计算机、服务器计算机、手持设备或便携式设备、平板型设备、多处理器系统、基于微处理器的系统、可编程的电子设备、网络PC、小型计算机、大型计算机、包括以上任何系统或设备的分布式计算环境等等。The various embodiments in the specification are described in a progressive manner, and the same or similar parts between the various embodiments may be referred to each other, and each embodiment focuses on differences from other embodiments. This application can be used in a variety of general purpose or special purpose computer system environments or configurations. For example: personal computer, server computer, handheld or portable device, tablet device, multiprocessor system, microprocessor based system, programmable electronic device, network PC, small computer, mainframe computer, including any of the above systems or The distributed computing environment of the device, and so on.
虽然通过实施例描绘了本申请,本邻域普通技术人员知道,本申请有许多变形和变化而不脱离本申请的精神,希望所附的权利要求包括这些变形和变化而不脱离本申请的精神。 While the present invention has been described by the embodiments of the present invention, it will be understood by those skilled in the claims .

Claims (19)

  1. 一种图像区域检测方法,其特征在于,所述方法包括:An image area detecting method, the method comprising:
    计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量;Calculating a color feature and a gradient feature of the pixel of the image to be processed, and constructing a mixed feature vector of the image to be processed;
    对所述混合特征向量进行聚类,获取聚类后的聚簇;Clustering the mixed feature vectors to obtain clusters after clustering;
    根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率;Calculating a clustering probability of the cluster according to a predetermined rule, and calculating a pixel probability of the pixel point in the cluster based on the clustering probability;
    基于所述像素概率对所述待处理图像进行检测,获取目标区域。The image to be processed is detected based on the pixel probability to acquire a target area.
  2. 如权利要求1所述的一种图像区域检测方法,其特征在于,所述计算得出待处理图像像素点的颜色特征包括:The image area detecting method according to claim 1, wherein the calculating the color characteristics of the pixel points of the image to be processed comprises:
    如果所述待处理图像不为Lab格式的数据,将所述待处理图像的数据格式转化为Lab格式;If the image to be processed is not data in the Lab format, converting the data format of the image to be processed into a Lab format;
    以待处理像素为中心提取所述待处理图像中邻域窗口的像素点,将所述邻域窗口中像素点的L、a、b三个通道分别分为K个分组,形成3*K维的颜色特征向量;Extracting pixel points of the neighborhood window in the image to be processed centering on the pixel to be processed, and dividing the three channels L, a, and b of the pixel in the neighborhood window into K groups to form a 3*K dimension. Color feature vector;
    将所述邻域窗口中每个像素点在所述L、a、b三个通道的颜色值累加到所述颜色特征向量所对应的维中,形成所述邻域窗口中待处理像素点的颜色特征。Adding color values of the three channels of the L, a, and b in each of the neighboring windows to the dimension corresponding to the color feature vector to form a pixel to be processed in the neighborhood window. Color characteristics.
  3. 如权利要求2所述的一种图像区域检测方法,其特征在于,所述K的取值为:6≤K≤16。The image area detecting method according to claim 2, wherein the value of K is: 6 ≤ K ≤ 16.
  4. 如权利要求1所述的一种图像区域检测方法,其特征在于,所述按照预定规则计算所述聚簇的聚簇概率包括:The image region detecting method according to claim 1, wherein the calculating a clustering probability of the cluster according to a predetermined rule comprises:
    计算所述聚簇中每个聚簇与其他聚簇的距离和,以所述聚簇和与所有聚簇的所述距离和的总和的比值作为所述聚簇的聚簇概率。A distance sum of each cluster in the cluster to other clusters is calculated, and a ratio of the cluster and the sum of the distance sums of all clusters is used as a clustering probability of the cluster.
  5. 如权利要求4所述的一种图像区域检测方法,其特征在于,所述计算所述聚簇中每个聚簇与其他聚簇的距离和包括:The image region detecting method according to claim 4, wherein the calculating a distance between each cluster in the cluster and other clusters comprises:
    采用下式计算所述聚簇中每个聚簇与其他聚簇的距离和D(Ci):The distance and D(Ci) of each cluster in the cluster from other clusters are calculated using the following formula:
    Figure PCTCN2016073274-appb-100001
    Figure PCTCN2016073274-appb-100001
    上式中,L为聚簇的个数,||ci,cj||为当前聚簇Ci的聚簇中心的混合特征向量与其他聚簇的聚簇中心混合特征向量的欧氏距离,Wj为根据当前聚簇Ci所包括的像素点设置的权 重。In the above formula, L is the number of clusters, ||c i , c j || is the Euclidean distance of the mixed feature vector of the cluster center of the current cluster Ci and the clustered feature vector of the cluster cluster of other clusters, Wj is a weight set according to the pixel points included in the current cluster Ci.
  6. 如权利要求1所述的一种图像区域检测方法,其特征在于,所述基于所述聚簇概率计算所述聚簇中像素点的像素概率包括:The image region detecting method according to claim 1, wherein the calculating a pixel probability of a pixel point in the cluster based on the clustering probability comprises:
    所述聚簇中像素点的像素概率为该像素点所属聚簇的聚簇概率。The pixel probability of the pixel in the cluster is the clustering probability of the cluster to which the pixel belongs.
  7. 如权利要求1所述的一种图像区域检测方法,其特征在于,所述基于所述聚簇概率计算所述聚簇中像素点的像素概率包括:The image region detecting method according to claim 1, wherein the calculating a pixel probability of a pixel point in the cluster based on the clustering probability comprises:
    以待求像素点p为中心提取第一邻域窗口W(p)’的像素点,采用下式计算所述待求像素点p的像素概率Sal(p):The pixel point of the first neighborhood window W(p)' is extracted centering on the pixel point p to be obtained, and the pixel probability Sal(p) of the pixel point p to be obtained is calculated by the following formula:
    Figure PCTCN2016073274-appb-100002
    Figure PCTCN2016073274-appb-100002
    上式中,P(q)为所述第一邻域窗W(p)’内的像素点q所属聚簇的聚簇概率,t为待求像素点p所属的聚簇中像素点的个数,σ为设置的平滑参数。In the above formula, P(q) is the clustering probability of the cluster to which the pixel point q in the first neighborhood window W(p)' belongs, and t is the number of pixels in the cluster to which the pixel point p to be obtained belongs. The number, σ is the set smoothing parameter.
  8. 如权利要求1所述的一种图像区域检测方法,其特征在于,所述基于所述像素概率对所述待处理图像进行检测获取目标区域包括:The image region detecting method according to claim 1, wherein the detecting the image to be processed based on the pixel probability to acquire the target region comprises:
    将所述待处理图像中像素点的像素概率值符合判断阈值PV要求的像素点作为所述待处理图像的目标区域;Pixel points whose pixel probability values of pixel points in the image to be processed meet the judgment threshold PV requirement are used as the target area of the image to be processed;
    或者,or,
    将所述待处理图像中像素的概率值大于第一阈值PF的像素点作为种子像素点;Pixel points whose probability values of pixels in the image to be processed are greater than the first threshold PF are used as seed pixel points;
    以所述种子像素点为中心计算与周围第二邻域窗口中像素点的欧式距离;Calculating an Euclidean distance from a pixel in a surrounding second neighborhood window centering on the seed pixel point;
    将所述欧式距离小于第二阈值的像素点作为新的种子像素点;Pixel points whose Euclidean distance is less than the second threshold are used as new seed pixel points;
    遍历所有所述种子像素点与周围所述第二邻域窗口中像素点的欧式距离并做出判断,将所述计算得出的种子像素点作为所述待处理图像的目标区域。And traversing the Euclidean distance of all the seed pixel points and the pixel points in the surrounding second neighborhood window and making a determination, and using the calculated seed pixel point as the target area of the to-be-processed image.
  9. 如权利要求8所述的一种图像区域检测方法,其特征在于,所述判断阈值PV的取值范围为:0.8≤PV≤0.95;The method for detecting an image region according to claim 8, wherein the value of the determination threshold PV is: 0.8 ≤ PV ≤ 0.95;
    或者,or,
    所述第一阈值PF的取值范围为:0.8≤PF≤0.95。The value of the first threshold PF ranges from 0.8 ≤ PF ≤ 0.95.
  10. 一种图像区域检测装置,其特征在于,所述装置包括:An image area detecting device, characterized in that the device comprises:
    特征计算模块,用于计算得出待处理图像像素点的颜色特征和梯度特征,并构建所述待处理图像的混合特征向量;a feature calculation module, configured to calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
    聚类模块,用于对所述混合特征向量进行聚类,获取聚类后的聚簇; a clustering module, configured to cluster the mixed feature vectors to obtain clusters after clustering;
    聚簇概率模块,用于根据预定规则计算所述聚簇的聚簇概率;a clustering probability module, configured to calculate a clustering probability of the cluster according to a predetermined rule;
    像素概率模块,用于基于所述聚簇概率计算所述聚簇中像素点的像素概率;a pixel probability module, configured to calculate a pixel probability of a pixel point in the cluster based on the clustering probability;
    检测模块,用于基于所述像素概率对所述待处理图像进行检测,获取目标区域。And a detecting module, configured to detect the image to be processed based on the pixel probability, and acquire a target area.
  11. 如权利要求10所述的一种图像区域检测装置,其特征在于,所述特征计算模块包括:The image area detecting apparatus according to claim 10, wherein the feature calculating module comprises:
    颜色特征模块,用于计算所述待处理图像像素点的颜色特征;a color feature module, configured to calculate a color feature of the pixel of the image to be processed;
    梯度特征模块,用于计算所述待处理图像像素点的梯度特征;a gradient feature module, configured to calculate a gradient feature of the pixel of the image to be processed;
    混合特征模块,用于将所述颜色特征和梯度特征结合,形成待处理图像的混合特征向量。A hybrid feature module is configured to combine the color feature and the gradient feature to form a mixed feature vector of the image to be processed.
  12. 如权利要求11所述的一种图像区域检测装置,其特征在于,所述颜色特征模块包括:The image area detecting apparatus according to claim 11, wherein the color feature module comprises:
    Lab转化模块,用于将所述待处理图像转化为Lab格式的数据;a Lab conversion module, configured to convert the image to be processed into data in a Lab format;
    颜色特征向量模块,用于以待处理像素为中心提取所述待处理图像中邻域窗口的像素点,将所述邻域窗口中像素点的L、a、b三个通道分别分为K个分组,形成3*K维的颜色特征向量;a color feature vector module, configured to extract pixel points of the neighborhood window in the image to be processed centering on the pixel to be processed, and divide the three channels L, a, and b of the pixel in the neighborhood window into K Grouping to form a color feature vector of 3*K dimensions;
    特征计算模块,用于将所述邻域窗口中每个像素点在所述L、a、b三个通道的颜色值累加到所述颜色特征向量所对应的维中,形成所述邻域窗口中待处理像素点的颜色特征。a feature calculation module, configured to accumulate color values of each of the L, a, and b channels in each of the neighboring windows in a dimension corresponding to the color feature vector to form the neighborhood window The color characteristics of the pixel to be processed.
  13. 如权利要求12所述的一种图像区域检测装置,其特征在于,所述颜色特征向量模块中K的取值范围为:6≤K≤16。The image region detecting device according to claim 12, wherein the value range of K in the color feature vector module is: 6 ≤ K ≤ 16.
  14. 如权利要求10所述的一种图像区域检测装置,其特征在于,所述聚簇概率模块包括:The image region detecting apparatus according to claim 10, wherein the clustering probability module comprises:
    距离和计算模块,用于计算所述聚簇中每个聚簇与其他聚簇的距离和;a distance and calculation module for calculating a distance between each cluster and the other clusters in the cluster;
    聚簇概率计算模块,用于根据所述聚簇和与所有聚簇的所述距离和的总和计算所述聚簇的聚簇概率。a clustering probability calculation module, configured to calculate a clustering probability of the cluster according to the cluster and the sum of the distance sums of all clusters.
  15. 如权利要求14所述的一种图像区域检测装置,其特征在于,所述距离计算模块计算所述聚簇中每个聚簇与其他聚簇的距离和包括:The image region detecting apparatus according to claim 14, wherein the distance calculating module calculates a distance between each cluster in the cluster and other clusters and includes:
    采用下式计算所述聚簇中每个聚簇与其他聚簇的距离和D(Ci):The distance and D(Ci) of each cluster in the cluster from other clusters are calculated using the following formula:
    Figure PCTCN2016073274-appb-100003
    Figure PCTCN2016073274-appb-100003
    上式中,L为聚簇的个数,||ci,cj||为当前聚簇Ci的聚簇中心的混合特征向量与其他聚簇的聚簇中心混合特征向量的欧氏距离,Wj为根据当前聚簇Ci所包括的像素点设置的权 重。In the above formula, L is the number of clusters, ||c i , c j || is the Euclidean distance of the mixed feature vector of the cluster center of the current cluster Ci and the clustered feature vector of the cluster cluster of other clusters, Wj is a weight set according to the pixel points included in the current cluster Ci.
  16. 如权利要求10所述的一种图像区域检测装置,其特征在于,所述像素概率模块包括下述中的至少一种:An image region detecting apparatus according to claim 10, wherein said pixel probability module comprises at least one of the following:
    第一概率模块,用于将像素点所属聚簇的聚簇概率作为该像素点的像素概率;a first probability module, configured to use a clustering probability of a cluster to which the pixel belongs is a pixel probability of the pixel;
    第二概率模块,用于以待求像素点p为中心提取第一邻域窗口W(p)’的像素点,采用下式计算所述待求像素点p的像素概率Sal(p):a second probability module, configured to extract a pixel point of the first neighborhood window W(p)' centering on the pixel point p to be obtained, and calculate a pixel probability Sal(p) of the pixel point p to be obtained by using the following formula:
    Figure PCTCN2016073274-appb-100004
    Figure PCTCN2016073274-appb-100004
    上式中,P(q)为所述第一邻域窗W(p)’内的像素点q所属的聚簇属于主体区域的概率,t为待求像素点p所属的聚簇中像素点的个数,σ为设置的平滑参数。In the above formula, P(q) is the probability that the cluster to which the pixel point q in the first neighborhood window W(p)' belongs belongs to the body region, and t is the pixel point in the cluster to which the pixel point p to be sought belongs The number of σ is the set smoothing parameter.
  17. 如权利要求10所述的一种图像区域检测装置,其特征在于,所述提取模块包括下述中的至少一种模块:An image region detecting apparatus according to claim 10, wherein said extraction module comprises at least one of the following:
    第一提取模块,用于将所述待处理图像中像素点的像素概率值符合判断阈值PV要求的像素点作为所述待处理图像的目标区域;a first extraction module, configured to use, as a target area of the to-be-processed image, a pixel point whose pixel probability value of the pixel in the image to be processed meets the requirement of the determination threshold PV;
    第二提取模块,用于将所述待处理图像中像素点属于主体区域的概率值大于第一阈值PF的像素点作为种子像素点;还用于以所述种子像素点为中心计算与周围第二邻域窗口中像素点的欧式距离;还用于将所述欧式距离小于第二阈值的像素点作为新的种子像素点;还用于遍历所有所述种子像素点与周围所述第二邻域窗口中像素点的欧式距离并做出判断,将所述计算得出的种子像素点作为所述待处理图像的目标区域。a second extraction module, configured to use, as a seed pixel point, a pixel point whose probability value of the pixel in the image to be processed belongs to the body region is greater than the first threshold PF; The Euclidean distance of the pixel in the two neighboring window; the pixel point of the Euclidean distance less than the second threshold is also used as a new seed pixel; and is further used to traverse all the seed pixels and the surrounding second neighbor The Euclidean distance of the pixel in the domain window is determined and the calculated seed pixel is used as the target area of the image to be processed.
  18. 如权利要求17所述的一种图像区域检测装置,其特征在于,所述判断阈值PV的取值范围为:0.8≤PV≤0.95;The image area detecting apparatus according to claim 17, wherein the value of the determination threshold PV is: 0.8 ≤ PV ≤ 0.95;
    和/或,and / or,
    所述第一阈值PF的取值范围为:0.8≤PF≤0.95。The value of the first threshold PF ranges from 0.8 ≤ PF ≤ 0.95.
  19. 一种图像区域检测装置,其特征在于,所述装置被设置成,包括:An image area detecting apparatus, wherein the apparatus is configured to include:
    第一处理单元,用于获取用户/客户端的待处理图像,计算得出待处理图像像素点的颜色特征和梯度特征,构建所述待处理图像的混合特征向量;a first processing unit, configured to acquire a to-be-processed image of the user/client, calculate a color feature and a gradient feature of the pixel of the image to be processed, and construct a mixed feature vector of the image to be processed;
    第二处理单元,用于对所述混合特征向量进行聚类,获取聚类后的聚簇;还用于根据预定规则计算所述聚簇的聚簇概率,并基于所述聚簇概率计算所述聚簇中像素点的像素概率;a second processing unit, configured to cluster the mixed feature vector to obtain clustered clusters; and further configured to calculate a clustering probability of the cluster according to a predetermined rule, and calculate the clustering probability based on the clustering probability The pixel probability of a pixel in the cluster;
    输出单元,用于基于所述像素概率对所述待处理图像进行获取目标区域,并将所述获取的目标区域存储或者展示于指定位置。 And an output unit, configured to acquire a target area of the to-be-processed image based on the pixel probability, and store or display the acquired target area in a specified location.
PCT/CN2016/073274 2015-02-12 2016-02-03 Image area detection method and device WO2016127883A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510075465.8A CN105989594B (en) 2015-02-12 2015-02-12 A kind of image region detection method and device
CN201510075465.8 2015-02-12

Publications (1)

Publication Number Publication Date
WO2016127883A1 true WO2016127883A1 (en) 2016-08-18

Family

ID=56614213

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/073274 WO2016127883A1 (en) 2015-02-12 2016-02-03 Image area detection method and device

Country Status (2)

Country Link
CN (1) CN105989594B (en)
WO (1) WO2016127883A1 (en)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110334936A (en) * 2019-06-28 2019-10-15 阿里巴巴集团控股有限公司 A kind of construction method, device and the equipment of credit qualification Rating Model
CN110610438A (en) * 2019-09-16 2019-12-24 黑龙江八一农垦大学 Crop canopy petiole included angle calculation method and system
CN111178118A (en) * 2018-11-13 2020-05-19 浙江宇视科技有限公司 Image acquisition processing method and device and computer readable storage medium
CN111862244A (en) * 2020-07-16 2020-10-30 安徽慧视金瞳科技有限公司 Plastic sheet intelligent color sorting method based on image processing
CN113034454A (en) * 2021-03-16 2021-06-25 上海交通大学 Underwater image quality evaluation method based on human visual sense
CN113034509A (en) * 2021-02-26 2021-06-25 北京达佳互联信息技术有限公司 Image processing method and device
CN113052922A (en) * 2021-03-26 2021-06-29 重庆紫光华山智安科技有限公司 Bare soil identification method, system, device and medium
CN113298829A (en) * 2021-06-15 2021-08-24 Oppo广东移动通信有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN113920528A (en) * 2020-07-08 2022-01-11 顺丰科技有限公司 Subject object detection method and device, computer equipment and storage medium
CN114445483A (en) * 2022-01-28 2022-05-06 泗阳三江橡塑有限公司 Injection molding part quality analysis method based on image pyramid
CN115115893A (en) * 2022-07-26 2022-09-27 金乡县富通金属回收有限公司 Intelligent sorting method for waste metal recovery
CN115496760A (en) * 2022-11-17 2022-12-20 澳润(山东)药业有限公司 Donkey-hide gelatin quality identification method
CN115526890A (en) * 2022-11-25 2022-12-27 深圳市腾泰博科技有限公司 Method for identifying fault factors of record player head
CN115661669A (en) * 2022-12-13 2023-01-31 山东省土地发展集团有限公司 Method and system for monitoring illegal farmland occupancy based on video monitoring
CN115861301A (en) * 2023-02-16 2023-03-28 山东百成新材料科技股份有限公司 Multi-material uniformity visual evaluation method for modified asphalt production
CN116092013A (en) * 2023-03-06 2023-05-09 广东汇通信息科技股份有限公司 Dangerous road condition identification method for intelligent monitoring
CN116205923A (en) * 2023-05-06 2023-06-02 威海锐鑫丰金属科技有限公司 Nondestructive testing method for internal defects of automobile hub based on X-RAY
CN116309577A (en) * 2023-05-19 2023-06-23 山东晨光胶带有限公司 Intelligent detection method and system for high-strength conveyor belt materials
CN116385434A (en) * 2023-06-02 2023-07-04 同济检测(济宁)有限公司 Intelligent detection method for precast beam cracks
CN116503404A (en) * 2023-06-27 2023-07-28 梁山县创新工艺品股份有限公司 Plastic toy quality detection method and device, electronic equipment and storage medium
CN116645354A (en) * 2023-06-02 2023-08-25 江西绿萌科技控股有限公司 Multifunctional sorting system-oriented surface flaw image detection method
CN116740809A (en) * 2023-06-05 2023-09-12 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture
CN116824168A (en) * 2023-08-29 2023-09-29 青岛市中医医院(青岛市海慈医院、青岛市康复医学研究所) Ear CT feature extraction method based on image processing
CN116977329A (en) * 2023-09-21 2023-10-31 江苏惠汕新能源集团有限公司 Photovoltaic grid line detection method based on machine vision
CN117011303A (en) * 2023-10-08 2023-11-07 泰安金冠宏油脂工业有限公司 Oil production quality detection method based on machine vision
CN117173164A (en) * 2023-11-02 2023-12-05 江苏君杰新能源有限公司 Grid line detection method of solar cell panel
CN117237384A (en) * 2023-11-16 2023-12-15 潍坊科技学院 Visual detection method and system for intelligent agricultural planted crops
CN117333484A (en) * 2023-11-30 2024-01-02 山东罗斯夫新材料科技有限公司 Method for detecting acrylic emulsion production wastewater based on image processing
CN117115494B (en) * 2023-10-23 2024-02-06 卡松科技股份有限公司 Lubricating oil impurity pollution detection method and device based on artificial intelligence
CN117520102A (en) * 2024-01-04 2024-02-06 广州大一互联网络科技有限公司 Intelligent energy consumption monitoring method for IDC data center
CN117635507A (en) * 2024-01-26 2024-03-01 深圳市精森源科技有限公司 Plastic particle online visual detection method and system

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108961316B (en) * 2017-05-23 2022-05-31 华为技术有限公司 Image processing method and device and server
CN108389205B (en) * 2018-03-19 2022-10-25 北京航空航天大学 Rail foreign matter monitoring method and device based on air-based platform image
CN108520259B (en) * 2018-04-13 2021-05-25 国光电器股份有限公司 Foreground target extraction method, device, equipment and storage medium
CN109611916B (en) * 2018-12-06 2020-08-04 绍兴京越智能科技有限公司 Smoking gear automatic adjustment platform
CN109815762B (en) * 2018-12-29 2022-02-11 福建天泉教育科技有限公司 Method and storage medium for remotely identifying two-dimensional code
CN110689478B (en) * 2019-09-25 2023-12-01 北京字节跳动网络技术有限公司 Image stylization processing method and device, electronic equipment and readable medium
CN113627453A (en) * 2020-05-08 2021-11-09 珠海金山办公软件有限公司 Pure-color background image matting method and device and electronic equipment
CN111726476B (en) * 2020-07-06 2022-05-31 北京字节跳动网络技术有限公司 Image processing method, device, equipment and computer readable medium
CN111862084B (en) * 2020-07-31 2024-02-02 东软教育科技集团有限公司 Image quality evaluation method, device and storage medium based on complex network
CN113077308B (en) * 2021-03-30 2022-03-22 张雨钊 Intelligent visual commercial transaction platform
CN114663316B (en) * 2022-05-17 2022-11-04 深圳市普渡科技有限公司 Method for determining edgewise path, mobile device and computer storage medium
CN114827711B (en) * 2022-06-24 2022-09-20 如你所视(北京)科技有限公司 Image information display method and device
CN116485819B (en) * 2023-06-21 2023-09-01 青岛大学附属医院 Ear-nose-throat examination image segmentation method and system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101447076A (en) * 2008-12-02 2009-06-03 浙江大学 Method for partitioning interested areas in WEB image
CN102842135A (en) * 2012-07-17 2012-12-26 杭州淘淘搜科技有限公司 Method of detecting main body area of product image
CN102867313A (en) * 2012-08-29 2013-01-09 杭州电子科技大学 Visual saliency detection method with fusion of region color and HoG (histogram of oriented gradient) features
US20140241624A1 (en) * 2013-02-27 2014-08-28 Sony Corporation Method and system for image processing
CN104217438A (en) * 2014-09-19 2014-12-17 西安电子科技大学 Image significance detection method based on semi-supervision
CN104268595A (en) * 2014-09-24 2015-01-07 深圳市华尊科技有限公司 General object detecting method and system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7039239B2 (en) * 2002-02-07 2006-05-02 Eastman Kodak Company Method for image region classification using unsupervised and supervised learning
CN103824283B (en) * 2014-01-22 2016-08-17 西安电子科技大学 Large format map segmentation method based on random chance sampling with multi-level fusion

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101447076A (en) * 2008-12-02 2009-06-03 浙江大学 Method for partitioning interested areas in WEB image
CN102842135A (en) * 2012-07-17 2012-12-26 杭州淘淘搜科技有限公司 Method of detecting main body area of product image
CN102867313A (en) * 2012-08-29 2013-01-09 杭州电子科技大学 Visual saliency detection method with fusion of region color and HoG (histogram of oriented gradient) features
US20140241624A1 (en) * 2013-02-27 2014-08-28 Sony Corporation Method and system for image processing
CN104217438A (en) * 2014-09-19 2014-12-17 西安电子科技大学 Image significance detection method based on semi-supervision
CN104268595A (en) * 2014-09-24 2015-01-07 深圳市华尊科技有限公司 General object detecting method and system

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111178118A (en) * 2018-11-13 2020-05-19 浙江宇视科技有限公司 Image acquisition processing method and device and computer readable storage medium
CN111178118B (en) * 2018-11-13 2023-07-21 浙江宇视科技有限公司 Image acquisition processing method, device and computer readable storage medium
CN110334936B (en) * 2019-06-28 2023-09-29 创新先进技术有限公司 Method, device and equipment for constructing credit qualification scoring model
CN110334936A (en) * 2019-06-28 2019-10-15 阿里巴巴集团控股有限公司 A kind of construction method, device and the equipment of credit qualification Rating Model
CN110610438A (en) * 2019-09-16 2019-12-24 黑龙江八一农垦大学 Crop canopy petiole included angle calculation method and system
CN113920528A (en) * 2020-07-08 2022-01-11 顺丰科技有限公司 Subject object detection method and device, computer equipment and storage medium
CN111862244A (en) * 2020-07-16 2020-10-30 安徽慧视金瞳科技有限公司 Plastic sheet intelligent color sorting method based on image processing
CN113034509A (en) * 2021-02-26 2021-06-25 北京达佳互联信息技术有限公司 Image processing method and device
CN113034454A (en) * 2021-03-16 2021-06-25 上海交通大学 Underwater image quality evaluation method based on human visual sense
CN113034454B (en) * 2021-03-16 2023-11-24 上海交通大学 Underwater image quality evaluation method based on human visual sense
CN113052922A (en) * 2021-03-26 2021-06-29 重庆紫光华山智安科技有限公司 Bare soil identification method, system, device and medium
CN113052922B (en) * 2021-03-26 2023-03-14 重庆紫光华山智安科技有限公司 Bare soil identification method, system, device and medium
CN113298829A (en) * 2021-06-15 2021-08-24 Oppo广东移动通信有限公司 Image processing method, image processing device, electronic equipment and computer readable storage medium
CN113298829B (en) * 2021-06-15 2024-01-23 Oppo广东移动通信有限公司 Image processing method, device, electronic equipment and computer readable storage medium
CN114445483A (en) * 2022-01-28 2022-05-06 泗阳三江橡塑有限公司 Injection molding part quality analysis method based on image pyramid
CN115115893A (en) * 2022-07-26 2022-09-27 金乡县富通金属回收有限公司 Intelligent sorting method for waste metal recovery
CN115496760A (en) * 2022-11-17 2022-12-20 澳润(山东)药业有限公司 Donkey-hide gelatin quality identification method
CN115526890A (en) * 2022-11-25 2022-12-27 深圳市腾泰博科技有限公司 Method for identifying fault factors of record player head
CN115661669A (en) * 2022-12-13 2023-01-31 山东省土地发展集团有限公司 Method and system for monitoring illegal farmland occupancy based on video monitoring
CN115861301A (en) * 2023-02-16 2023-03-28 山东百成新材料科技股份有限公司 Multi-material uniformity visual evaluation method for modified asphalt production
CN116092013B (en) * 2023-03-06 2023-06-09 广东汇通信息科技股份有限公司 Dangerous road condition identification method for intelligent monitoring
CN116092013A (en) * 2023-03-06 2023-05-09 广东汇通信息科技股份有限公司 Dangerous road condition identification method for intelligent monitoring
CN116205923A (en) * 2023-05-06 2023-06-02 威海锐鑫丰金属科技有限公司 Nondestructive testing method for internal defects of automobile hub based on X-RAY
CN116309577A (en) * 2023-05-19 2023-06-23 山东晨光胶带有限公司 Intelligent detection method and system for high-strength conveyor belt materials
CN116309577B (en) * 2023-05-19 2023-08-04 山东晨光胶带有限公司 Intelligent detection method and system for high-strength conveyor belt materials
CN116385434B (en) * 2023-06-02 2023-08-08 同济检测(济宁)有限公司 Intelligent detection method for precast beam cracks
CN116385434A (en) * 2023-06-02 2023-07-04 同济检测(济宁)有限公司 Intelligent detection method for precast beam cracks
CN116645354A (en) * 2023-06-02 2023-08-25 江西绿萌科技控股有限公司 Multifunctional sorting system-oriented surface flaw image detection method
CN116740809A (en) * 2023-06-05 2023-09-12 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture
CN116740809B (en) * 2023-06-05 2024-03-29 嘉兴米兰映像家具有限公司 Intelligent sofa control method based on user gesture
CN116503404A (en) * 2023-06-27 2023-07-28 梁山县创新工艺品股份有限公司 Plastic toy quality detection method and device, electronic equipment and storage medium
CN116503404B (en) * 2023-06-27 2023-09-01 梁山县创新工艺品股份有限公司 Plastic toy quality detection method and device, electronic equipment and storage medium
CN116824168B (en) * 2023-08-29 2023-11-21 青岛市中医医院(青岛市海慈医院、青岛市康复医学研究所) Ear CT feature extraction method based on image processing
CN116824168A (en) * 2023-08-29 2023-09-29 青岛市中医医院(青岛市海慈医院、青岛市康复医学研究所) Ear CT feature extraction method based on image processing
CN116977329B (en) * 2023-09-21 2023-12-08 江苏惠汕新能源集团有限公司 Photovoltaic grid line detection method based on machine vision
CN116977329A (en) * 2023-09-21 2023-10-31 江苏惠汕新能源集团有限公司 Photovoltaic grid line detection method based on machine vision
CN117011303A (en) * 2023-10-08 2023-11-07 泰安金冠宏油脂工业有限公司 Oil production quality detection method based on machine vision
CN117011303B (en) * 2023-10-08 2024-01-09 泰安金冠宏油脂工业有限公司 Oil production quality detection method based on machine vision
CN117115494B (en) * 2023-10-23 2024-02-06 卡松科技股份有限公司 Lubricating oil impurity pollution detection method and device based on artificial intelligence
CN117173164A (en) * 2023-11-02 2023-12-05 江苏君杰新能源有限公司 Grid line detection method of solar cell panel
CN117173164B (en) * 2023-11-02 2023-12-26 江苏君杰新能源有限公司 Grid line detection method of solar cell panel
CN117237384A (en) * 2023-11-16 2023-12-15 潍坊科技学院 Visual detection method and system for intelligent agricultural planted crops
CN117237384B (en) * 2023-11-16 2024-02-02 潍坊科技学院 Visual detection method and system for intelligent agricultural planted crops
CN117333484B (en) * 2023-11-30 2024-02-20 山东罗斯夫新材料科技有限公司 Method for detecting acrylic emulsion production wastewater based on image processing
CN117333484A (en) * 2023-11-30 2024-01-02 山东罗斯夫新材料科技有限公司 Method for detecting acrylic emulsion production wastewater based on image processing
CN117520102A (en) * 2024-01-04 2024-02-06 广州大一互联网络科技有限公司 Intelligent energy consumption monitoring method for IDC data center
CN117520102B (en) * 2024-01-04 2024-04-26 广州大一互联网络科技有限公司 Intelligent energy consumption monitoring method for IDC data center
CN117635507A (en) * 2024-01-26 2024-03-01 深圳市精森源科技有限公司 Plastic particle online visual detection method and system
CN117635507B (en) * 2024-01-26 2024-04-09 深圳市精森源科技有限公司 Plastic particle online visual detection method and system

Also Published As

Publication number Publication date
CN105989594B (en) 2019-02-12
CN105989594A (en) 2016-10-05

Similar Documents

Publication Publication Date Title
WO2016127883A1 (en) Image area detection method and device
CN108304435B (en) Information recommendation method and device, computer equipment and storage medium
CN101763429B (en) Image retrieval method based on color and shape features
KR101662972B1 (en) Estimating depth from a single image
CN109360028B (en) Method and device for pushing information
US20170161591A1 (en) System and method for deep-learning based object tracking
CN110059728B (en) RGB-D image visual saliency detection method based on attention model
CN106575280B (en) System and method for analyzing user-associated images to produce non-user generated labels and utilizing the generated labels
CN104391879B (en) The method and device of hierarchical clustering
CN113301409B (en) Video synthesis method and device, electronic equipment and readable storage medium
CN107301643B (en) Well-marked target detection method based on robust rarefaction representation Yu Laplce&#39;s regular terms
EP3779775A1 (en) Media processing method and related apparatus
CN104123718A (en) Device and method for image processing, image processing control program, and recording medium
WO2014036813A1 (en) Method and device for extracting image features
CN108960041B (en) Image feature extraction method and device
CN114998320B (en) Method, system, electronic device and storage medium for visual saliency detection
WO2014000427A1 (en) Method and device for detecting flame
CN112052960A (en) Longitudinal federal modeling method, device, equipment and computer readable storage medium
TW201839665A (en) Object recognition method and object recognition system
CN113301408A (en) Video data processing method and device, electronic equipment and readable storage medium
CN105979283A (en) Video transcoding method and device
Jin et al. Adaptive propagation-based color-sampling for alpha matting
CN110570402A (en) Binocular salient object detection method based on boundary perception neural network
TWI711004B (en) Picture processing method and device
JP2013182330A (en) Image processor and image processing method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16748684

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16748684

Country of ref document: EP

Kind code of ref document: A1